Abstract
In this essay, I consider how queer methods can inform survey research on sex, gender, sexuality, and other dimensions of social difference and inequality. Critical attention to methodological assumptions, disciplinary norms, and inherited tools can afford sexualities scholars opportunities to reconsider established practices and help chart a more thoroughly transformative era in the sociology of sexualities and related fields. I review a number of recent methodological advances that can facilitate the integration of queer methods across registers of survey research—from study design and data collection to analysis and reporting. I conclude with several recommendations, including that scholars: (1) treat surveys as representational tools, not as neutral mechanisms of revelation; (2) resist statistical fetishism and the tacit notion that statistical complexity equals conceptual sophistication; and (3) recognize that data created by all kinds of surveys can be used as tools for social justice or domination, regardless of their intended consequences.
Keywords
I will admit to feeling a bit uneasy when I hear concerns among sexuality scholars and other critical methodologists about the need to refine survey methods to better capture human experiences of sex and sexuality. I’ve spent much of my career thinking about the relationships between survey tools and the ways we use them to produce knowledge about sex, gender, race, and sexuality (Grzanka 2016; Grzanka, Zeiders, and Miles 2016). There is certainly room for improvement, and survey researchers must be sensitive to shifts in the ways that, for example, individuals conceptualize and report their sexualities and gender. Indeed, transformations and cross-cultural differences in the terms and categories people use to define and label their sexual orientation and gender identities present important challenges to researchers seeking to do the things most survey methodologists do with data: describe observable patterns (see Silva [2025], this issue). And it is crucial that sexualities scholars stay vigilant about the kinds of questions social scientists and policymakers are increasingly asking to determine who’s queer and gender expansive and how many of us there are (Wang 2023).
I think the ongoing work to more accurately and justly identify and describe diverse populations and social problems is vital, but critiques of existing tools and practices in survey research (including those from the Left) often frame such concerns in epistemic terms that are quite consistent with traditional, masculinist, and positivist perspectives. These are the very frameworks that queer methods, feminisms, critical race theories, and other post-structural and anti-positivist movements have worked to deconstruct (Compton, Meadow, and Schilt 2018; Ghaziani and Brim 2019; Pfeffer 2018). Thus, we should ask ourselves: when we talk about improving survey methods so that we can better assess (e.g., count and account for) people
I think that questions about critiquing and improving methods must begin with a serious interrogation of our methodologies, or the ideas and values that undergird our use of specific research methods (Harding 1987). This work is not quick, but it is often dirty insomuch as it necessitates the messy feelings and daunting self-reflexivity that come from looking at our cherished, multigenerational, inherited ideas and wondering which of them must be changed, which should be kept, and which should be thrown out (McClelland 2017). One of the bad habits learned in disciplinary settings—hardly exclusive to sociology!—is the failure to even name the epistemic values and goals that motivate a given research project or area of study (Grzanka and Cole 2021). But this is a requisite first step in better understanding our typically tacit investments in what kinds of knowledge are possibly produced by our research. Put simply:
Assumptions/Attachments/Disinvestments
To illustrate this point, it’s helpful to consider the similarities and differences in how social scientists in various disciplinary contexts generally conduct survey research. In American sociology, the historical gold standard of survey research has been the nationally representative, probability sample (Luker 2008). One advantage of representative datasets that are collected annually is that this allows for the study of change over time; we can look to the General Social Survey, for example, for evidence of how abortion attitudes have shifted (or not) across several decades. These kinds of datasets are rarely produced by individual researchers or small labs. They are instead the product of massive research operations at major research centers, such as NORC or the U.S. Census Bureau. In this research context, individual researchers who may want to probe these datasets to test hypotheses and/or ask questions about sexualities (or gender, or race) are beholden to data they did not create and questions they did not ask. The sloppy and often crude ways gender identity and sexual orientation—not to mention race and other relevant identity variables—are asked on these national surveys artificially limit what can be learned about the population and arbitrarily constrain, in particular, how nonnormative sexual and gender identities are represented (Baumle 2018). Survey methodologists informed by queer and trans perspectives have made many important scholarly interventions in the past several decades (Westbrook and Saperstein 2015). But the impact of these developments in queer and trans survey research is limited by the extent to which these insights are incorporated into the datasets most commonly used by sociologists, demographers, and other social scientists who are concerned with data that represents demographic trends in the broader population.
In psychology, whose disciplinary norms are less organized around representative datasets, researchers are far more accustomed to creating their own survey research instruments even if they use existing and standardized measurement tools. 1 In other words, because psychologists often rely on convenience (i.e., nonprobability) samples of undergraduate students or samples collected using online surveys distributed via social media or collected via crowdsourcing networks (e.g., Amazon’s Mechanical Turk [MTurk], Prolific, etc.), they often exercise far more power over the type of data they collect. This does not mean their surveys are invariably or even typically higher quality, but it does mean that they can nimbly tailor questions to suit specific research aims. One advantage of this approach is that surveys can be adjusted to address contemporary and highly dynamic concerns, such as the emergence of a term (e.g., “pansexual”), or its decline in use or cultural salience (e.g., “heteroflexible”). Previously validated scales, such as measures of internalized heterosexism, can also be easily tweaked to function in different populations or replaced to help answer new questions. For example, an item such as “I feel discriminated against as a gay man” used in one study could become “I feel discriminated against as a bisexual man” in a follow-up study. A drawback, however, is that it is difficult to make meaningful comparisons across nonequivalent surveys, which hinders the capacity to confidently assess changes over time for items or questions that are inconsistently asked. Even slight differences in the format (e.g., “Check all that apply” versus “Please choose one”) or content (e.g., “Homosexual” versus “Gay”) in a demographic item can make comparisons spurious or impossible. Though convenience samples are not statistically representative of the population of interest, that may not be the goal of the study, and therefore not a fundamental limitation. Instead, the findings may be framed as logically or analytically representative (Luker 2008) insomuch as they illuminate the substance and contours of a dynamic or construct under investigation (akin to the goals and strengths of qualitative inquiry).
So, on the occasion of the inaugural volume of
Though statistics are regularly positioned as more objective tools than qualitative methods (Luker 2008), the angle at which one approaches quantitative data can obscure or illuminate what’s actually going on statistically. To illustrate this point, Mittleman (2023) quantitatively explored a theoretical axiom of queer studies that is under-investigated in mainstream population research: sexual fluidity, or the extent to which individuals’ sexual identities and attractions change over time (Diamond 2009). Using five waves of annual data collected since 2013 among a nonclinical population of U.S. adults (
Mittleman (2023) then tested whether accounting for sexual fluidity might change our conclusions about observed health outcomes for sexual minority adults; it did not. However, sexually fluid adults reported significantly worse outcomes than those who consistently reported that they were straight. Sexually fluid respondents were far more like stable sexual minorities than those who were stably straight. Moreover, the failure to account for sexual fluidity grossly underestimated the overall size of the population of sexual minorities in the data, which Mittleman notes is a prime goal of demographic research. So, while accounting for sexual fluidity did not drastically change the conclusions drawn from the PATH survey regarding health inequities, measuring sexual fluidity did unsettle assumptions about what demographic knowledge can be gleaned from cross-sectional (i.e., single timepoint) assessments of who counts as LGBQ+. Mittleman concludes that no single survey administration can accurately quantify how many people “are” queer. Instead, he argued demographers should dedicate attention to the dynamics of social complexity that are already well-established in the sexualities literature. So, the problem for Mittleman isn’t in how we ask about sexuality
Relatedly, West et al. (2024) found that national survey data may misdirect our conclusions about sexual minority health outcomes, particularly if measures arbitrarily constrain options for sexual orientation identity. West and colleagues used five consecutive years of the National Survey of Family Growth (2015–2019) to test whether differences in how sexual identity was asked would significantly affect tests of subgroup differences (straight vs. everyone else) on various health outcomes, including substance use and reproductive health. This national survey was randomly distributed with two different versions of the sexual identity question: one 3-question option (gay, straight, bisexual) and one with four choices, including “something else.” They found that, for both cisgender women and men, there were statistically different conclusions on 15% to 16% of the outcome measures between the three- and four-category versions of the survey questionnaire. When using the four-category version, they found significant differences in a variety of dependent variables (three each for women and men), including wanting a/another child, cigarette and marijuana use, and having ever contracted a sexually transmitted infection. Whereas the exclusive straight/gay/bisexual triad reflects heteronormative ideas that “bisexual” captures anyone who is not monosexual in their attraction or identity, the simple addition of a fourth “something else” category allowed West and colleagues to draw more nuanced and statistically meaningful conclusions from the data. Echoing Mittleman (2023), West et al. (2024) do not suggest the four-category sexual identity question is a fix-all solution for measuring sexual identities. They instead underscore the need for all survey researchers to consider “fluidity and complexity when it comes to the construct of sexual orientation identity” (p. 123).
Mittleman (2023) and West et al. (2024) point toward the wily issue of
In feminist psychology, Galupo et al. (2014) asked sexual minorities (
Similarly, critical methodologist McClelland (2017) demonstrated how survey tools can (mis)lead researchers into a false sense of measurement validity. Consistent with Cantril’s self-anchored ladder approach popularized in the 1960s in his nationwide survey of American well-being,
2
McClelland asked respondents to rate themselves on a 10-point scale of “overall sexual satisfaction”
Galupo and McClelland’s respective projects underscore the extent to which survey research tools can give us false impressions of both (a) the ideological distance between us, as researchers, and our participants, as well as (b) the homogeneity of perspectives among our participants. In other words, if a measurement tool is initially found to be reliable and valid, it can lead us to think we share an understanding of a term or construct—such as sexual pleasure—with most or even all our respondents. However, our respondents (i.e., sample of the population of interest) may actually define this construct in ways that are quite different from how we have conceptualized it. There also may be critical distinctions among our participants such that “sexual pleasure” to one subset or respondents is completely orthogonal to another subgroup’s understanding of the term. This is not the same as measuring actual variance in responses, such as different rates at which individuals report experiencing sexual pleasure, which is something we want to know if it exists. If we think we’re measuring “sexual pleasure” but we’re really measuring something else entirely, this is
Another way to mitigate these threats to construct validity is to develop survey instruments that are rooted in the lived experiences and actual language of stigmatized minority groups, who are experts in their own lives. In my own research on sexual orientation beliefs—or what everyday people believe sexual orientation is, where it comes from, and the extent to which it can change—my colleagues and I inverted the typical way that prejudice and discrimination measures are developed. Rather than base our Sexual Orientation Beliefs Scale (SOBS; Arseneau et al. 2013) on the worldviews of straight and cis people, we began developing and validating the scale using ideas from queer and feminist theory and the thoughts and perspectives of queer and trans people. This meant that the items we proposed for the scale incorporated social constructionist themes—such as the notion that sexual orientation is culturally and historically contingent—with more depth and frequency than they appear in heteronormative discourse about sexuality. We also used exclusively queer and trans respondents to conduct the initial validation of the instrument and then moved on to study how the measure might perform psychometrically with straight respondents. Every aspect of the SOBS’s development was designed to buck the commonplace, yet methodologically suspect, ways most psychometric work on prejudice is conducted. We wanted to start with the marginalized groups who are the targets of discrimination and anchor the instrument in their subjectivities, rather than those of the dominant group (Grzanka 2016). Ultimately, this allowed us to develop an instrument that moves beyond a reductionist “born this way”-or-not binary and to capture elements of essentialist and social constructionist beliefs that are not exclusively about how straight people perceive or discriminate against sexual minorities.
From Queer Methods to Queer Data?
Though none of the empirical projects I described in the previous section explicitly invoke the term “queer method,” they circulate in a time where increased scholarly attention has been paid to the potential for queer theory to catalyze consequential methodological innovations and contestations, as well as to destabilize the divide between “theory” and “method.” Four celebrated volumes by Browne and Nash (2010), Ghaziani and Brim (2016, 2019), and Compton, Meadow, and Schilt (2018) represent the gestalt of the queer methods movement, which has importantly been led by several key figures in the sociology of sexualities. These contributions have challenged social scientists and humanists to reconsider the role of empiricism in the study of
Queer data studies is an exciting complement to the queer methods movement, drawing on critical science and technology studies approaches to consider the travels and consequences of knowledge about queer lives. Similar to queer methods, Keilty (2023) notes in the introduction to
I began this brief meditation by naming my skepticism toward scholarly talk about refining survey tools and building better instruments, even though, admittedly, that is a lot of what occupies my time and thoughts when it comes to the sociology of sexualities. Queer methods and, now, queer data studies offer critical resources for avoiding the twin trapdoors of enlightened positivism or universal pessimism. In conclusion, rather than desire the perfect tool, analyze ourselves into statistical oblivion, or leverage our political futures with data we can never fully control, I think a methodological calculus informed by queer methods can reframe our investments thusly:
We should humbly and modestly consider how our survey tools and ancillary methods function as representational heuristics—ways of thinking about the empirical world that offer clues to patterns, anomalies, differences,
Statistics, like literally any other mode of analysis, is a language for claims-making and knowledge creation. Complex statistics are not fundamentally better than more straightforward approaches, and statistical abstraction is sometimes helpful and sometimes harmful. We should use statistics when appropriate to the research aims and pursue complexity when necessary, but never as a maxim and never without justification. Further, we should consider statistics to be productive of knowledge, rather than revelatory.
Political work in survey research, like the kind many sexualities scholars pursue in the interest of social justice for members of marginalized and oppressed social groups, has the potential to buttress efforts for collective, expansive liberation. But data are never in and of themselves liberatory: data are spoken for and about by political actors, and they exist within discourses that are impossible to control and extremely challenging even to influence (Clare, Grzanka, and Wuest 2023). The same data we think will minimize violence or promote reparation can be used to expand theaters of violence and resist efforts at restitution (see Shuster and Campos-Castillo [2025], this issue). We should question the hope(s) we invest in our data, not because we should be hopeless, but because data—including survey data—are among many tools that can be used to fight structural violence.
