Abstract
Sex differences in mental rotation performance are one of the largest in cognitive psychology. Men outperform women by up to 1 SD in psychometric mental rotation tests, but it is often neglected that there are no or only small sex differences for chronometric tests. As both tests are supposed to measure the same ability, we suspect some features of the tests themselves to affect sex differences in performance. Following a theoretical review of the test features, we evaluate the effects of the number of possible answer alternatives, whether they are presented as pairwise mirrored, and their interaction on sex differences in mental rotation performance. In an online experiment, 838 German-speaking participants, 421 women, 417 men, Mage = 42.58 (SD = 12.54) years, solved four blocks of mental rotation trials with two or eight alternatives, which were either pairwise mirrored or not. The results show that that the overall performance was lower for more alternatives and for mixed alternatives but not for their interaction. We could not determine explanations for sex differences as we did not observe meaningful sex differences at all. Possible reasons include the differences between men and women in age and education. This study suggests that the differences between tests affect performance. Sex differences, however, need more investigation, including possible effects and interactions of the test design, education, and age.
Mental rotation describes the cognitive ability to rotate objects or images in the mind. As a spatial ability, the performance in mental rotation tests shows strong links to STEM (Science, Technology, Engineering, and Mathematics) performance (e.g., Hausmann, 2014; Moè, 2016; Moè et al., 2018). Gender/sex differences in mental rotation performance are often described as one of the largest in cognitive psychology (Voyer et al., 1995). They are thus often interpreted as a possible reason for the gender gap in STEM fields. A better understanding of these sex/gender differences could in turn help in closing the gender gap. However, it is often neglected that these sex/gender differences are not observed for all mental rotation tests.
Two types of mental rotation tests are widely used: (1) chronometric tests (for brevity referred to as SM tests) based on the study of R. N. Shepard and Metzler (1971), where two objects are presented which are either “same” (rotated) or “different” (mirrored) and (2) psychometric tests (VK tests) based on the study of Vandenberg and Kuse (1978), where compared with one target, two out of four alternatives are rotated and the other two are mirrored or structurally different. Both tests are suggested to measure the same mental rotation ability and the results of different mental rotation tests do correlate with each other (Voyer et al., 2006). However, large performance differences favouring men are only detected in VK tests, whereas there are only small or no differences in SM tests between men and women (Peters & Battista, 2008). Recently, Jost and Jansen (2020) evaluated another test design (JJ tests), where one target is compared with two alternatives, which are mirrored to each other and thus one is rotated to the target and one is mirrored. Interestingly, sex differences were also not observed for this type of test (Jost & Jansen, 2022).
Research has focused on examining gender and/or sex differences in VK tests relating to biological factors such as hormones, menstrual cycle, and sexual orientation (Hausmann et al., 2000, 2009; Peters et al., 2007; Peters, Laeng, et al., 1995), social factors such as gender stereotypes and stimulus familiarity (Hausmann et al., 2009; Moè et al., 2021; Ruthsatz et al., 2014, 2015, 2017), cultural aspects (Jansen et al., 2016; Peters et al., 2006), education and academic background (Moè et al., 2021; Peters et al., 2006, 2007; Peters, Laeng, et al., 1995), emphasising the relevance of role models (Neuburger et al., 2013), or strategy selection (Hegarty, 2018; Heil & Jansen-Osmann, 2008; Toth & Campbell, 2019; Voyer et al., 2020). 1 However, there exists no conclusive theory for the occurrence of these sex differences. As sex differences also seem to increase with age during childhood and adolescence, research has investigated possible causes such as sex-stereotyped activities, parental language, or spatial toys, but again with no conclusive results (Newcombe, 2020). Another question regards the distinction between gender and sex. Studies investigating biological factors would indicate sex differences, whereas social factors would suggest differences by gender. However, most participants in studies are cis-gender persons and these might be confounded. 2
Many of the aforementioned studies indeed identified interactions with sex differences. However, it must be emphasised that all those studies only investigated the VK test, which typically produces sex differences. If these factors influenced the mental rotation process, which all tests are supposed to measure, the same performance difference between sexes should occur in chronometric tests as most of those factors are also applicable. Thus, the question remains, why one test design provokes these factors to manifest in performance differences by sex while the other design does not. Peters and Battista (2008) and Bors and Vigneau (2011) already note that these differences might not be related to differences in actual mental rotation ability, but, for example, to the switching between pairwise comparisons within one trial for VK tests or dwelling on details of the stimuli. However, research investigating the test design has been scarcer but has analysed effects of the existence of a time limit (e.g., Peters, 2005; Voyer, 2011), the answering format (two out of four choice vs. same/different, Titze et al., 2010), or the distractor configuration (Voyer & Hou, 2006). While a relaxation of time limits and some aspects of the distractor configuration have been found to reduce performance differences between sexes, neither can fully explain the differences between tests.
To understand when and how sex differences in mental rotation tests emerge and whether they are linked to mental rotation ability, it seems necessary to consider the differences between the tests. However, a systematic comparison of test features is still lacking and to date no theory or experimental investigation exists that can explain the varying occurrences of sex differences in performance between tests. Despite much research focusing on sex differences in mental rotation ability, it is still unclear whether they are actually attributed to mental rotation ability or spatial abilities in general. One reason for the missing explanation for sex differences in VK tests could be that they are not actually linked to mental rotation ability but other features of the test. We thus believe this to be of great importance moving forward, to systematically investigate effects of the test design on mental rotation performance and sex differences.
In a first step, we attempt to systematise features of mental rotation tests. Concurrently, we will revisit existing theories for sex differences in mental rotation performance due to test features and how they hold up for a comparison between tests. Subsequently, we will formulate new theories regarding sex differences in mental rotation test performance. Furthermore, we will experimentally investigate the basis of these theories.
Part 1: theories regarding sex differences in performance and a comparison between mental rotation tests
In this chapter, we aim to analyse the properties of mental rotation test designs and provide an overview on the current state of research on their influence on sex differences in performance. We review existing explanations for sex differences in mental rotation performance, which involve features of the tests. We will not review explanations that do not consider test features, as these cannot explain the observed pattern of differing sex differences between tests. A problem, again, is that no systematic test concept exists. Tests have often been varied in multiple ways, often unrelated to the specific research question (such as variations of time limits for investigations of stimulus material), which makes it difficult to isolate effects of specific features. This once again underlines the importance of a review of features of the tests themselves.
The differences (and similarities) between mental rotation tests can be grouped into two parts. First, there are differences between the individual trials of each test. Second, there are overarching differences of the test design and organisation mostly unrelated to individual trials.
Differences between individual trials of different mental rotation tests
While the recent JJ test was inspired by the SM tests, the trials show a similarity to trials of VK tests, as always half of the answers are correct. In that way, it can be seen as a computerised type of VK trials with only two alternatives. Moreover, the two alternatives are aligned across their mirroring plane such that they are easily identifiable as mirrored. In VK trials the four alternatives are not easily pairwise identifiable as mirrored to each other. By modifying the layout of JJ trials to a horizontal positioning of all three figures, the differences and similarities between the tests are shown in Figure 1.

Different mental rotation trials.
Thus, we propose three major differences between trials of the three mentioned mental rotation tests:
Compared with SM trials, participants are informed that always exactly half of the possible answers are correct in both JJ trials and VK trials. In most versions of the SM test, overall, about half of the trials are “same” trials and the other half “different” trials, but this is typically neither exact nor known to participants.
The trials use different number of alternatives (1, 2, or 4) and total number of items per trial (2, 3, or 5).
In JJ trials, the alternatives are pairwise identifiable as mirrored and aligned across the mirroring plane, whereas in VK trials, they are ordered randomly and not pairwise mirrored. For the SM trials, this distinction is not applicable.
In addition, there is a possible difference in the upright orientation of the figures. For abstract figures, such as the widely used cube figures and polygons, this is somewhat arbitrary as there is no clear upright orientation. While we can for the cube figures describe the upright orientation as an alignment with the natural axes (which are most of the time also the rotational axes), this is often not reported and might need further investigation. Another feature of the tests is the number of targets, but the used tests do not differ in this feature, as all use one target per trial.
Although the differences in the layout were necessary to highlight the differences between tests, they do not seem to play a major role in test performance. For SM trials, Xue et al. (2017) found no difference whether the left or the right item was rotated. For a version of VK trials with a vertically separated target, J. K. Krüger and Suchan (2016) found large sex differences in their control group, which was a sample of the general population. A comparison of the horizontal distances between stimuli mentioned by Battista and Peters (2010) also revealed no differences. Layout variations have been used without further inspection of the layout itself by Wohlschläger and Wohlschläger (1998), Wohlschläger (2001), and M. Krüger and Krist (2009) and different experiments regularly use varying computer screens and image sizes as well as non-standardised distances to the screen or test material and resulting optical distances between stimuli. Sex differences in VK tests and non-differences in SM tests have emerged under various of these non-systematically varied conditions.
Out of the suspected major differences, only the number of items per trial has yet been manipulated. Both Titze et al. (2008) and Voyer et al. (2020) restricted the visibility of the answer alternatives in VK tests. In the study of Titze et al., three out of four alternatives were hidden by a template and participants only saw the target and one alternative at a time. In two experiments, Voyer et al. employed a gaze-contingent display and participants could only see one item (either the target or one of the alternatives) at one time. However, these manipulations did not aim to change the number of alternatives and items within each trial, but only the visibility. To actually manipulate the number of items per trial, Titze et al. (2010) conducted a test that resembles a paper–pencil version of an SM test. Each trial of a VK test was split into four pairwise same/different judgements, which were then presented in random order. However, none of the manipulations could eliminate sex differences in performance. All three studies reported significantly better performance for men compared with women in all experiments, which are comparable with the performance differences of other VK tests. Titze et al. (2008) also found no reduction in sex differences compared with the regular test (the sex differences increased non-significantly from d = 0.40 to d = 0.45 by using the template).
Differences between mental rotation tests
Next to the differences between individual trials of the tests, there are some overall differences between the tests. These were already partially manipulated by researchers, although not always as part of an analysis of sex differences. In the following, we describe some of these manipulations and their influence on sex differences. These studies do, however, often vary in more than the parameter in question (e.g., different time limits for digital tests compared with paper and pencil tests) and use different numbers of trials (we expect smaller standardised effect sizes of sex differences with less trials due to the increased relative variance of guessing).
Ending condition/time limits
In their most often used versions, the VK test is limited by both time and the number of trials (two times 3 min for 12 trials, Peters, Laeng, et al., 1995), the SM test is limited by the number of trials, and the JJ test is limited only by time. By relaxing or removing time limits in VK tests, sex differences in performance were reduced, but still easily detectable at Cohen’s d ≈ 0.5 and much larger than in SM tests (Peters, 2005; Voyer, 2011). As sex differences are neither commonly observed for SM tests limited by the number of stimuli and JJ tests limited by time, the ending conditions of tests cannot be the sole reason for sex differences.
Stimulus material
Whereas all tests most commonly use abstract cube figures as stimuli, there are small differences between the used figures. However, the original stimuli of Vandenberg and Kuse (1978) were based on the stimuli developed by R. N. Shepard and Metzler (1971) and redrawn versions of the stimuli have been used interchangeably between tests. For example J. K. Krüger and Suchan (2016) used the figures of Peters and Battista (2008), which are widely used in SM tests. Thus, other aspects of the stimuli and different stimuli have been investigated as possible reasons for sex differences in VK tests. The possible reasons include the embodiment of figures, gender stereotypes about the items, and the realism of figures. Moreover, the occlusions in the two-dimensional representations of three-dimensional objects have been analysed.
For human figures and a separation of partially occluded and nonoccluded figures, sex differences are still reported in VK tests, although often reduced for tests with overall higher average scores (e.g., Alexander & Evardone, 2008; Doyle & Voyer, 2013; Jansen & Lehmann, 2013; Voyer et al., 2020). As the widely used cube figures are seen as more male stereotyped, Ruthsatz et al. (2014, 2015, 2017) investigated female stereotyped stimuli and partially found reduced and even reversed sex differences in the VK test performance of children. These results have, however, only been partially transferred to adults with reduced or negated sex differences but no female advantage (Jost & Jansen, 2021; Rahe & Quaiser-Pohl, 2020; Rahe et al., 2020). Regarding the realism of stimuli, both Fisher et al. (2018) and Robert and Chevrier (2003) found no significant sex differences for real three-dimensional objects resembling cube figures. However, men were much faster than women in the study of Robert and Chevrier and scores approached the maximally attainable scores in the study of Fisher et al., indicating possible ceiling effects. While it is possible that more realistic figures reduce sex differences, no clear conclusions can be drawn as neither study was sufficiently powered to reliably detect smaller than large effects with only about 20 men and women per condition.
For SM tests, Voyer and Jansen (2016) did find better performance for men with human figures as stimuli. 3 However, this is one of the few studies also reporting better performance for men for cube figures and the performance differences between stimuli were comparable for all stimulus types or even larger for the human figures (although quantified by differences in reaction time and accuracy on both same and different trials). Using hands as stimuli, Voyer, Jansen, et al. (2017) found better performance for men, whereas Campbell et al. (2018) detected possibly better performance for women. By comparing multiple stimulus types in SM tests, Jansen-Osmann and Heil (2007) found sex differences only for two-dimensional polygons. Heil and Jansen-Osmann (2008) replicated these sex differences for polygons and found larger differences for more complex polygons. Except for the polygons, all reported standardised effect sizes for sex differences in SM tests are small. As the cube figures show the largest reaction times and the lowest accuracies in comparisons of different stimuli and thus are the most complex in the study of Jansen-Osmann and Heil (2007), there does not seem to be a monotonous relationship between complexity and sex differences in SM tests. However, it is not clear whether the same effect occurs for VK tests as polygons have not yet been used as stimuli. Overall, it is likely that the used stimulus material and its complexity interacts with sex differences in mental rotation tests but that they are not a major reason for differences between different tests using cube figures.
The rotational axis
For many versions of VK tests only rotations in depth are used, whereas SM tests often use rotations in the picture plane or combine items rotated both in depth and in the picture plane in one test. In a study of a VK test with children, Ruthsatz et al. (2014) found an interaction of gender and axis: Larger sex differences occurred for cube figures rotated in depth compared with rotations in the picture plane. This effect was, however, not found for the second stimulus type, pellet figures. For adults, Battista and Peters (2010) did find better performance for rotations around a vertical axis in depth compared with a horizontal axis but no significant interaction with sex differences.
For SM tests using only rotations in depth, no or only small sex differences are found (Jordan et al., 2002; Peters, 2005; Rahe, Ruthsatz, Schürmann, & Quaiser-Pohl, 2019). The aforementioned sex differences for polygons are also only found for rotations in the picture plane. Many tests use multiple rotational axes as the widely used stimulus library of Peters and Battista (2008) provides stimuli rotated around all three canonical axes. The original study of R. N. Shepard and Metzler (1971) already suggested a strong similarity between rotational axes and these are often not further distinguished. For realistic stimuli, there is some evidence for better performance for rotations around a vertical axis compared with a horizontal rotation in picture plane (Foulkes & Hollifield, 1989). A possible reason could be the relevance of vertical rotations and gravity in real-life, but as Shiffrar and Shepard (1991) point out, these differences could be related to how the rotational axes are aligned with the features of the stimuli even for abstract figures. Despite the possibility for much further research into the effect of rotational axes, it seems unlikely that rotations in depth are a major reason for the discrepancy between sex differences found in different tests.
Type of distractors
Both SM tests and JJ tests use mirrored items as distractors, whereas VK tests also use structurally different items. By separating the answers for trials with mirrored and structural distractors in VK tests, mostly either non-significant effects of the distractor type or reduced sex differences and overall better performance for structural distractors are found (Boone & Hegarty, 2017; Bors & Vigneau, 2011; Doyle & Voyer, 2013; Monahan et al., 2008; Voyer & Hou, 2006). Thus, it seems that structural distractors in VK tests are likely not the reason for sex differences. However, Boone and Hegarty (2017) also analysed the incorporation of structural distractors with or without mirrored distractors in SM tests and found large sex differences, indicating that the combination with structural distractors is a possible reason for sex differences. But as they also found large sex differences on a VK test using only mirrored distractors in line with other uses of only mirrored distractors in VK tests (Rahe et al., 2020; Rahe & Quaiser-Pohl, 2020; Rahe, Ruthsatz, Jansen, & Quaiser-Pohl, 2019), their results only open this possibility to increase sex differences in SM tests but not how to reduce sex differences in VK tests.
Test administration
The VK tests were originally developed and are still often administered as a paper and pencil test, whereas SM and JJ tests are computerised to measure reaction times of individual trials. With the rising prevalence of computers since the original conception, many computerised and online versions of VK tests have been administered in more recent studies. These tests regularly report better performance for men, although sometimes with reduced effect sizes compared with paper and pencil tests (e.g., Debarnot et al., 2013; J. K. Krüger & Suchan, 2016; Monahan et al., 2008; Peters et al., 2007; Voyer et al., 2020). As mentioned before, in one study of a paper and pencil SM test, Titze et al. (2010) did find large sex differences in performance. While the test administration cannot be the sole reason for varying sex differences between mental rotation tests, the results of Titze et al. need to be investigated further.
Participant organisation
As with the test administration, the VK test was developed to quickly test large groups and multiple participants often perform the test in the same room at the same time. The group composition and stereotypes associated with genders have been hypothesised to influence performance differences (Moè, 2018). However, the comparison by Peters et al. (2006) of individual and group testing revealed no differences. In addition, but lacking systematic comparisons, associated with the form of the test administration, computerised and online versions of VK tests are often conducted individually or in the same manner as SM tests and still produce sex differences. Moreover, also individual testing using paper and pencil VK tests still produces sex differences (e.g., Titze et al., 2008).
Scoring system
To reduce the impact of guessing and based on the original work of Vandenberg and Kuse (1978), the most widely used scoring system for VK tests awards one point to a trial if and only if both correct alternatives are selected. These scores are not directly comparable to SM tests, for which the reaction time and accuracy is analysed for each “same” item and not analysed for “different” items. For JJ tests, there is no need to remove items, but reaction time and accuracy are still evaluated on a by-item basis. While participants are mostly instructed about the scoring system in VK tests, the task in SM tests is typically to solve the items as quickly and accurately as possible.
Due to the combination of multiple items into one trial, the reaction time in VK tests cannot be attributed to single items. For a comparison of scoring systems on VK tests, Voyer et al. (1995) found smaller, but still large sex differences for test scores on single items. For a comparison of both scoring systems within one test, Titze et al. (2010) found a high correlation and large sex differences for both. While VK tests bundle rotated items and distractors into one score, most SM tests only report scores for rotated items. In the few studies analysing both performance for mirrored stimuli and sex differences in SM tests, Peters (2005) observed no sex differences and Voyer and Jansen (2016) found small and comparable sex differences on both mirrored and non-mirrored trials. Kerkman et al. (2000) did find sex differences only for the mirrored trials, indicating a possible reason for the worse performance of women in VK tests. Their design, however, was different as very different stimuli were used and participants were also required to identify the direction of rotation. Moreover, their results indicate the tendency of women to guess that trials are non-mirrored (leading to higher accuracy for non-mirrored trials and lower accuracy on mirrored trials). While this led to sex differences only for the mirrored trials, this could also be interpreted as an overall worse performance. While the differences in the scoring systems prevent most results from being directly compared between tests, they may be a possible reason for reduced performance differences between sexes and could use some further investigation.
Practice trials and feedback
The mental rotation tests also differ in the number of practice trials before the results are recorded. While VK tests most commonly use 3 or 4 practice trials (composed of 12–16 alternatives), there is no standard for SM or JJ tests. The number of practice trials ranges from 0 to >100 with many studies using 10–20 practice trials comparable with VK tests, but there does not seem to be a systematic investigation of sex differences, depending on the amount of practice. Moreover, in repeated VK tests, sex differences are still found in later tests (Peters, 2005; Peters, Laeng, et al., 1995) despite the increased practice, although sometimes reduced (Peters, Chisholm, & Laeng, 1995). Another aspect is the inclusion of feedback for every trial during SM tests in some procedures. Rahe, Ruthsatz, Schürmann, et al. (2019) identified the missing feedback as a possible reason for sex differences. However, for example, both Jansen-Osmann and Heil (2007) and Voyer and Jansen (2016) used feedback for all trials and partially found sex differences.
Theories on influences on sex differences in mental rotation tests
To conclude the comparison of mental rotation tests, there is no conclusive evidence or feature explaining the reasons for sex differences in mental rotation tests. Thus, new theories regarding the occurrence of the said sex differences are necessary.
A link between test difficulty and sex differences
One theory is that sex differences only emerge on tests with sufficient difficulty. A reoccurring pattern in the previous chapter is that sex differences in VK tests are typically reduced in test versions with higher scores (tests using simpler stimuli, easier identifiable distractors, increased available time, or digitalized tests). Still, sex differences are observed in VK tests but not or with smaller effect sizes in SM tests independent of these parameters. Moreover, even a combination of multiple factors reducing sex differences does not eliminate them. For example, Doyle and Voyer (2013, 2018) and the free-viewing experiment of Voyer et al. (2020) used computerised VK tests without time limits and human figures as stimuli and found better performance in men. Nevertheless, a more comprehensive comparison of mental rotation test scores seems necessary. Unfortunately, there is no direct way to compare scores between tests due to the different scoring systems and the lack of such a comparison once again points out the scarcity of comparisons between different mental rotation tests. We will thus attempt to provide some approximation of error rates for individual items in VK tests and a comparison with SM tests in the following.
Due to the information that always half of the alternatives are correct, the individual items are not independent. However, the exact dependency depends on the employed strategy and is impossible to estimate in general. In the following, we present two simple approximations of overall accuracy of single item comparisons in VK tests. These simple methods only estimate overall accuracies and do not account for differences between trials (e.g., lower accuracy with larger angular disparity). While a typical correction for multiple-choice tests concerns the number of answer alternatives (i.e., six possibilities to choose two out of four alternatives in VK tests), this is not necessary for a comparison between SM and VK tests. As we try to compute the probability that one single item in VK tests was solved correctly from the overall score, this includes the probability of guessing correctly in half of the cases. As a result, the corrected scores are directly comparable between tests as the probability to correctly identify rotated pairs, although they do not represent the true accuracy.
One way to estimate the probability
where
Another way to estimate single item accuracies is assuming the same probability for each item. This, however, is heavily dependent on the employed strategy. Assuming that a trial is solved correctly if three single item comparisons are solved correctly,
5
the probability to solve a trial correctly is
We have compared both methods on all studies cited above which report the full values for accuracy of adult participants in VK tests in their standard form, which produce the most known and largest sex differences, and compared them with accuracies in SM tests. The full comparison is available in the supplementary material. Using either method, the overall accuracy for single task comparisons is estimated as .51 to .95. Assuming three to five single item comparisons are performed per trial (independent of the assumption for
Based on these estimations, there do not seem to be sufficiently large differences in single item difficulty between tests to explain the large differences in sex differences. While reducing the overall difficulty of VK tests often reduces sex differences in performance, the non-differences in SM tests can likely not be explained by the difficulty of single items. Moreover, varying stimulus material in SM tests has rarely produced sex differences and these are not systematically related to task difficulty. Accuracy also drops and reaction time increases with larger angular disparity and sex differences have not consistently been observed with this increasing difficulty (see also Bors & Vigneau, 2011).
To conclude, it is unlikely that sex differences in VK tests are solely due to the increased difficulty of this test. It is, however, possible that increased difficulty amplifies sex differences or the detection of them due to the underlying binomial distribution of answers.
A link between trial design and sex differences
A second theory is that sex differences emerge due to features of the trial design, which differ between tests and have not been sufficiently investigated. These features are the number of alternatives per trial, the presentation of alternatives as pairwise mirrored, and, to an extent, the number of simultaneously presented trials and overall visible stimuli. These correspond to necessary and possible pairwise comparisons of stimuli to solve trials. In general, all mental rotation tests can be solved by repeated pairwise comparisons but differ in the number of necessary and possible pairwise comparisons to solve each trial. The overall number of possible pairwise comparisons between two figures are one for SM trials, three for JJ trials, and 10 for VK trials. Out of these, one, two, and four involve the target. Due to the information that exactly half of the alternatives are correct only one comparison for JJ trials and two or three comparisons for VK trials are necessary to complete the trial as the other items can then be solved by exclusion. Between computerised and paper–pencil tests, the tests also differ in the number of visible stimuli unrelated to the trial at hand as multiple trials are presented on one page. These offer further possible comparisons, which are not necessary and not helpful to find the solutions of individual trials. Either the number of necessary comparisons per trial, the number of possible comparisons per trial, the overall number of possible comparisons within a test, or a combination of all of them could be related to sex differences as all offer additional comparisons, which are not related to test performance.
However, neither of these effects nor their interaction has been conclusively investigated. Related to the comparisons within one trial, both the template task of Titze et al. (2008) and the restricted viewing experiments of Voyer et al. (2020) still required the same number of same/different judgements for one trial as the classical VK trials. However, in the trials of Titze et al., all pairwise comparisons involved the target. Regarding possible comparisons between trials, the paper–pencil SM test of Titze et al. (2010) had multiple and possibly related trials (same target) being visible at once.
The layout and visibility of stimuli could require spatial resources due to strategies of optimally disentangling the linked same/different judgements within one trial and the unlinked comparisons between different trials. The link to performance could thus be due to the spatial demands of the layout. The mediation study of Kaufman (2007) provides evidence that a large part of sex differences in VK tests is due to spatial working memory. In contrast, for a letter variation of the SM test, the dual-task study of Hyun and Luck (2007) indicates that what they called object working memory rather than spatial working memory is required to solve the task. However, due to the difference in designs and the rather small number of participants in the case of the study of Hyun and Luck, more research is needed to establish definitive links between working memory components and test versions. Next to possible spatial demands of the trial layout, both the number of comparisons necessary and the number of visible stimuli could be a measure of “perceived complexity.” Sex differences could then emerge due to different strategies between men and women for dealing with challenges due to, for example, speculated differences in upbringing and education (Määttä & Uusiautti, 2020).
Predictions by the theory
Thus, there are three possible predictions by the theory. The first possibility is that only the layout of the trials produces sex differences in tests. This may be due to the spatial demands of the layout and sex differences in spatial working memory. A consequence of this would be sex differences in favour of men in any test of other cognitive abilities using similar layout variations. The second and third possibilities suggest an interaction between the layout and the task. The second possible prediction is that the layout itself requires resources of a working memory subcomponent, likely the object or spatial working memory. If the task at hand requires resources of an overlapping working memory subcomponent and either of these working memory subcomponents has sex differences, then sex differences should be enhanced in the task performance. This would predict sex differences at least in tests of spatial abilities if a similar layout to VK tests is applied. To align this with the results of Hyun and Luck (2007), this would mean that either the layout requires resources of the object working memory, for which, however, a female advantage is observed (Voyer et al., 2007). Or the mental rotation task does require spatial working memory, just not to an extent that was exceeded in the dual task condition of Hyun and Luck and non-differences between sexes in SM tests are explained by superior object working memory of women similar to the observations of L. J. Levy et al. (2005). The third possibility is that the increased perceived complexity of the task due to the layout leads to sex differences if the added complexity of the tasks exceeds what women are confident to deal with more than for men. This would predict sex differences due to the trial layout for tests, which are already perceived as rather challenging.
Part 2: experimental investigation of within-trial factors and sex differences in mental rotation performance
In any of the cases predicted by the theories discussed beforehand, sex differences should be enhanced in mental rotation tests due to the trial layout. Next to the possible underlying theories, the gathered evidence necessarily suggests that some difference between tests is a reason for varying sex differences. To better formulate, understand, and differentiate theories, it seems important to determine whether and which specific aspects of the trial layout produces or enhances sex differences. Due to the widely used different mental rotation tests, they thus are a prime candidate to identify these features.
Here, we aim to investigate the influence of the previously identified differences in the trial design on sex differences in test performance. The two most promising parameters seem to be the number of alternatives to each target and whether the alternatives are pairwise easily identifiable as mirrored to one another.
We will use a combination of paired or mixed alternatives with varying numbers of alternatives and the following hypotheses are investigated:
We expect larger performance differences favoring men for the test using mixed alternatives.
We expect larger performance differences favoring men for tests using more alternatives.
Moreover, we expect the effect of the number of alternatives to be larger for the mixed alternatives.
As secondary hypotheses and due to the results of the pilot study, we predicted learning during the test and better performance in later blocks (Heil et al., 1998; Peters, 2005) and lower accuracies for larger angular disparities of items in line with common observations in SM tests. Moreover, we predicted lower overall performance for the mixed alternatives, for more alternatives, and for the interaction of both. 7 The performance in later trials within a block will also be investigated, but the reduced performance in later trials observed in the pilot study contrasts the improvements observed during chronometric tests (Jost & Jansen, 2020).
Pilot study
A preregistered first experiment was conducted using similar methods and investigating the same main hypotheses (the full method and results are also available at the preregistration and at Jost, 2022). However, due to an error in the a priori power analysis, the experiment was not sufficiently powered to detect all effect sizes of interest for the main hypotheses concerning effects between subjects. It should rather be treated as a pilot study. Nevertheless, there are some conclusions to improve the method for a second experiment and these will be presented briefly.
For the primary hypotheses, the number of alternatives were two, four, or eight. As secondary hypotheses, we also investigated the effects of additional covariates. Between subjects, STEM affiliation and previous experience with mental rotation were investigated. Within subjects, the angular disparity of trials and the learning within and between blocks were investigated.
The results provide some evidence that sex differences in mental rotation tests could be due to the trial layout, especially due to the mixed presentation of alternatives instead of a pairwise mirroring. The sex differences due to the trial layout interacted with both STEM education and previous experience. In line with the preregistered recruitment, the group of participants of main interest were those without previous experience with mental rotation. These, and especially the subgroup of non-STEM students, are the most tested sample in other studies, which provided the evidence for large sex differences. For these participants, the results show that sex differences are larger for mixed alternatives and smaller and not significant for paired alternatives in line with the hypothesis and the small or non-differences found for SM and JJ tests.
For the within-subjects effects, overall performance was lower for the mixed alternatives and for more alternatives. These effects also interacted for an even lower performance for more mixed alternatives. In line with the predictions, performance increased between blocks and decreased for larger angular disparities. Contrary to our prediction, performance decreased within blocks even when excluding unanswered trials due to the time limit.
Regarding the adaptation of the method for this experiment, the pilot study provides two insights. First, the overall effect of the number of alternatives was monotonic and roughly linear. This means employing more alternatives can enhance effects and this allows us to use larger differences in the number of alternatives to increase effect sizes and thus the power to detect differences. Second, the overall achieved scores are on average very similar to the assumed values in the power analysis and estimated in Table S1 (the average achieved scores ranging from .70 to .86 compared with the assumed values of .66–.81). The absolute differences between the scores of men and women were close to the expected values. However, the variance between participants was larger than expected and the expected large relative sex differences for VK tests (Cohen’s d > 0.7) for non-STEM participants without previous experience were only achieved for eight mixed alternatives and not for four mixed alternatives (Cohen’s d = 0.26). Possible reasons could be the larger variance in the sample due to recruitment from multiple universities or effects of the less controlled testing environment (one participant provided feedback that he would have preferred a mouse over a touchpad). Another reason for the lower relative sex differences could be the improvements between blocks and the usage of more blocks in this experiment compared with traditional VK tests, but also that only 24 items were used per condition instead of 48, which increases the relative variance of the binomial distribution. Whereas there are studies reporting also large sex differences for VK tests with only 12 trials (24 rotated items), these effects could be overestimated. While the estimated effects should not be overinterpreted (Brysbaert, 2019), the possibility of lower than previously estimated sex differences in performance and a lower correlation shall be considered in the following power analysis.
As we are also interested in explaining sex differences in VK tests found in the literature, we will focus on non-STEM affiliated people without extensive experience with mental rotation as these are the most studied population of interest. While the differential effects of STEM education and experience are still of interest, they exceed the scope of this work.
Method
Power analysis
As a general guideline, Brysbaert (2019) suggests at least 860 participants to conclude that an effect is rather negligible at an effect size smaller than d = 0.2 for the comparison of two groups, which is a plausible lower bound here. As we are interested in finding all influencing factors on sex differences and explaining both sex differences and non-differences for the separate manipulations, this approach seems reasonable. An additional simulation was conducted to reach more specific estimates for the planned analyses. Based on the linearity of the effect of the number of alternatives observed in the pilot study and the theoretical considerations of differences between designs as well as the specifics of the planned tests using only 24 items per block, we estimated an overall difference of interest between two paired and eight mixed alternatives of the order of d = 0.8. Smaller overall sex differences would lead to harder to detect influencing factors, but would also point towards the influence of the trial design being smaller than expected. Moreover, for the conditions in between, we estimated the minimal values of interest of Cohen’s f for these interactions as 0.1, 0.2, and 0.4. The simulations suggested appropriate power of .8 at around 800 participants (see Figure 2). Thus, we opted for the suggestion of 860 participants at an equal distribution of men and women.

Simulated power for different numbers of participants using loess smoothing in ggplot2 (Wickham, 2016).
Participants
Participants were recruited through the online panels provided by Bilendi (www.meinungsplatz.de) and Respondi (mingle.respondi.de). The experiment targeted 430 adult men and 430 women without specific STEM affiliation and previous experience with mental rotation. Basic outlier detection was performed online such that these were not counted towards the final sample. The experiment was closed once 868 participations were logged for the final sample. Overall, 1,251 participations were logged at that time.
Participants were rewarded for their participation according to the guidelines of the panel providers. Before starting the study, all participants gave informed consent using a digital checkbox. Participants consented to anonymous sharing of their data and use of their data for scientific purposes. The experimental procedure was approved by the universities’ Ethics committee (Ethikkommission bei der Universität Regensburg, No. 20-2096_1-101).
Of the 838 participants included in the final sample, 421 were female and 417 were male. The mean age was 42.58 (SD = 12.54) years with men (46.85 ± 11.88) being older than women (38.36 ± 11.73); 5 women and 15 men had previous experience with mental rotation but neither acute nor chronic. Descriptively, the women were also better educated than men, with more women with university degrees (except for PhDs) and more men with lower school degrees. The distributions of age and education are presented in Figure 3 as they were used for further exploratory analyses.

Distribution of age (left) and education (right) in the final sample.
Procedure
Participants were informed about the goals and procedure of the study before starting with a demographic questionnaire for prescreening purposes. Then, they completed the mental rotation tests succeeded by further questions about their experience with similar tests. Participants were asked to plan sufficient time (15–20 min) and complete the test alone. For purposes of monitoring participant numbers, there were two identical experiments. One was advertised to persons/households identified as female/women in the panel providers’ database and one to male/men.
Measures
Mental rotation
The mental rotation test was implemented using OSWeb (version 1.3.11) as part of OpenSesame (version 3.3.6; Mathôt et al., 2012) and made available online by JATOS (Lange et al., 2015). The test consisted of four blocks with a time limit of 3 min for each block. Each block consisted of trials of only one type of trials for a total of 48 alternatives (i.e., 24 trials with two alternatives or six trials with eight alternatives) in line with 12 trials for each block of VK tests. Trials were varied by the number of alternatives (two or eight) and by the pairing of alternatives (paired or mixed). Every participant completed one block for each combination. In the paired condition, the alternatives were pairwise horizontally mirrored to one another and aligned with the canonical axes while the target was rotated compared with the canonical axes. In the mixed condition, the target was aligned with the canonical axes and the alternatives were ordered randomly. For all numbers of alternatives, the target was on the left side of the screen and the alternatives were ordered from left to right with an additional space of 50px (at a resolution of 1,920 × 1,080) between the target and the first alternative. In the case of eight alternatives, there were two rows of four alternatives each. Examples of trials are shown in Figure 4.

Examples of mental rotation trials with two mixed alternatives (top), two paired alternatives (centre), and eight mixed alternatives (bottom).
The order of the blocks was randomised for each participant. Within each block for each participant, the used cube models and rotation angles for the target were randomised such that each combination of parameters occurred again only after all other choices had been presented at least once. The orientation (of the target and alternatives) and angles of the alternatives were chosen randomly such that no two figures were the same and all alternatives differed from the target by angle. All randomizations were performed using inline JavaScript in OpenSesame using the modern version of the Fisher–Yates algorithm.
The stimuli were generated using the library of Jost and Jansen (2020) using the parameters given in Table 1. Stimuli were rotated around the z-axis (vertical) first and the x-axis (horizontal) afterwards and the different orientations used different base angles such that the paired alternatives were actual mirror images of one another (see Figure 4). For the structure of the cube figures, Models 2–8 and 12–16 of the library of Peters and Battista (2008) were used as the other models can be transformed into these models by mirroring and/or rotating.
Parameters for stimuli generation.
Participants could select and deselect the alternatives by clicking them with the mouse. Selected alternatives were marked by a quadratic white border. In the bottom-right corner of the screen, there was a button to continue with the next trial. The button could only be clicked once exactly half of the alternatives were selected. Otherwise, the button text asked the participants to select exactly half of the alternatives and could not be clicked. There was no option to return to previous trials.
The test was preceded by four practice trials in random order, one for each combination of two or eight alternatives and mixed or paired alternatives. For the practice trials, the participants had sufficient time (at most 15 min) and received feedback in the form of a red or green border around the selected alternatives.
Before each block, participants were instructed about the number of trials, the number of alternatives, and the time limit, but not about the pairing of alternatives. Between blocks, they were allowed a self-paced break. As in VK tests, participants were instructed that they would get one point if and only if they selected all correct alternatives. During the trials, participants could see the number of the current trial and the number of overall trials for the block in the top-left corner and the time left (since the start of the trial) in the top-right corner of the screen.
Demographics
A digital questionnaire was used to collect demographic information. Before the mental rotation test, participants were asked about their sex (male, female, or diverse), their age in years, their education (in eight levels ranging from “No high school diploma” to “PhD”), and their main field of work and education (from 18 possible categories). Participants were excluded at this stage if they selected one of the categories belonging to STEM fields (“Computer science,” “Engineering & Technology,” “Natural sciences & mathematics”).
After the mental rotation test, participants were asked about their previous experience with mental rotation (participants had to indicate if they had or had not participated in other mental rotation experiments with similar tasks before). Participants were excluded from analyses if they had experience with mental rotation tests in the past year (acute experience) or with more than three such tests (chronic experience).
Statistical analysis
Traditionally, mental rotation scores have been analysed using ANOVAs assuming a normal distribution despite the theoretically more correct binomial distribution for accuracy data. However, accuracy data for angular disparities in chronometric mental rotation tests as the only numerical predictor have regularly shown linear or less than linear decreases. If the underlying effect is assumed to be linear, these results contrast the predictions of the logit link function of the binomial distribution. For comparability with older work on sex differences in VK tests and due to the unclear linearity of effects, we thus used the normal distribution as the main analysis and provide a secondary analysis using the binomial distribution. 8 For all analyses, we employed mixed effects models due to the advantages over traditional ANOVAs. For example, linear mixed models allow the simultaneous analysis of by-participant and by-item variances, thus eliminating the need to average over participants or items, while also facilitating the analysis of unbalanced data and achieving higher statistical power (Baayen et al., 2008; Barr et al., 2013; Hilbert et al., 2019).
The proportion of correctly solved items in each block was used as dependent variable and the number and pairing of alternatives, the sex, and their interaction were used as independent variables. For the secondary hypotheses, the position of the block was included as a predictor. For the binomial distribution, the angular disparity of items and the position of the trial within the block were used as additional predictors. In line with common scoring procedures, unattempted trials were treated as wrongly answered. For trials which were not finished due to the time limit, the answers were evaluated if at most half of the figures were selected. If more than half were selected, all answers to that trial were treated as unattempted.
Statistical analysis was performed with linear mixed models using MixedModels package (version 4.0.0; Bates et al., 2021) in Julia (version 1.6.2; Bezanson et al., 2017). Additional secondary analyses were performed using generalised mixed models with a binomial distribution. Model fit was calculated by using likelihood ratio tests to compare models with and without the effect of interest. Participants were used as random effects. Random slopes were selected stepwise starting with a maximal model and removing random slopes by dropping variance components using an LRT backwards heuristic at α = .2 (Matuschek et al., 2017). Non-significant fixed effects were further stepwise removed from the model, such that effects that least decreased model fits were removed first and a model containing only significant fixed effects remained. Non-significant effects were then tested for an improvement of model fit by inclusion in the resulting model, while significant effects were tested for worsening of model fit by exclusion of the effect. The resulting p-values were compared with a significance level of .05. The analysis of main effects contained in significant interactions was performed according to R. Levy (2014) and we thus used normalised sum contrasts for all categorical variables and centred all numerical variables and normalised them to a range of 1.
While there are several advantages to mixed models, the internal optimization procedures are not exact, which produces large imprecisions in both estimated effect sizes and p-values when the random effects are overparametrized. In addition to the procedures of Matuschek et al. (2017), we planned to reduce the random effects structures until these uncertainties were of magnitudes of .001 for the p-values and point estimates in our checked samples.
Four procedures were implemented to detect outliers. Participants were excluded if overall performance was below chance level (.5) on their attempted trials. This is the most often used outlier detection in mental rotation tests. However, performance of guessers should be symmetrically distributed around chance level and this procedure would thus only exclude about half of guessers. We thus implemented further outlier detection. These procedures were also designed to exclude participants who did not attempt to achieve the maximum score as instructed, as participants were given credit by the platform for simply completing the test, which may have incentivized participants to complete the tests quickly rather than actually trying to solve the tasks. Participants were thus excluded if the sum of their relative time used and their accuracy on their attempted trials was <1, which indicates a too strong focus on finishing the test quickly instead of accurately. These two outlier detection procedures were implemented online (albeit the online calculation of relative time did not account for the time limit and was thus less restrictive than the final outlier detection), such that these participants were not counted towards the desired sample size. Participants were further inspected as possible outliers if their overall accuracy was <.66 and they selected the first answer (leftmost alternative) in >90% of all trials or <10% of all trials indicating answering in a pattern instead of correctly. Participants who attempted less than half of all trials were also inspected for possible exclusion if their overall accuracy was <.66.
Results
Outliers
Overall, 390 outliers had to be removed for performance reasons (216 men, 173 women, 1 diverse), 232 participants because of performance below chance level, an additional 156 participants for a too large focus on speed over accuracy, and two further participants with too few overall attempts (see Figure 5). Furthermore, 9 women and 14 men were excluded because of either acute or chronic experience with mental rotation tests.

Speed–accuracy trade-offs and distribution of outliers.
Mental rotation data
The overall proportion of correctly solved items per condition is shown in Figure 6. Overall Cohen’s d for the four conditions was 0.04 (8 paired), 0.11 (8 mixed), 0.13 (2 mixed), and 0.16 (2 paired alternatives).

Proportion of correctly solved rotated items, separated by conditions and sex. Error bars represent standard error.
The random effects structure of the mixed model included random slopes for the interaction and main effects of the number and type of alternatives as well as main effects for blocks and random intercepts. The main effects of the number of alternatives, the type of alternatives, and blocks were significant but neither interaction of interest was significant. Performance was better for less alternatives, for paired alternatives, and in later blocks. Men performed non-significantly better than women. The secondary alternative analysis using a binomial distribution yielded the same significance of the predictors. In addition, the angular disparity and the number of the trial within the block showed significant main effects. Performance was lower for larger angular disparities and for trials later within a block (see Table 2).
Statistical analysis of proportion of correct items.
SE: Standard error.
Exploratory analyses
Because of the unexpectedly small sex differences in all conditions compared with both the first experiment and the literature, the following variables were analysed in addition to the preplanned analyses. The descriptive data for these analyses is shown in Figure 7 and the statistical values in Table 3.

Descriptive data for the exploratory analyses using more restrictive outlier detection (top-left), the number of attempted items (top-right), the proportion of correctly solved attempted items (bottom-left), and the fully correctly solved trials (bottom-right). Error bars represent standard error.
Exploratory statistical analyses.
SE: Standard error.
First, we looked at the effect of age and education because the women were younger and had higher degrees than the men and mental rotation performance was found to be higher for younger (for adults) and better educated persons (Peters et al., 2007). We included linear effects of age and education in ascending order. Both produced effects in the expected direction but the effect of education was only significant (p = .001) when not accounting for age. Controlling for these covariates also produced sex differences in overall performance, albeit the largest effect found when controlling for age was still smaller than expected (compared with the values in table S1) with men solving only about 6% more items correctly than women.
Second, because of the large number of outliers, we explored a more restrictive outlier detection by excluding all participants performing at or below 60% correct answers on their attempted trials (the mean between chance level and lowest expected accuracy scores of 70%). This procedure should include almost all participants who guessed but performed above chance level due to luck. An additional 136 participants (63 men, 71 women) were excluded. However, there was no visible or significant change in the patterns of the main hypotheses except for the small main effect of sex now being significant.
We furthermore looked at sex differences in the number of attempted items, the performance on only the attempted items, and for the trials using eight alternatives also the number of fully correctly solved trials in line with the traditional scoring system on VK tests. The number of attempted items showed some variation compared with the overall accuracy. The triple interaction of interest reached significance, but only after excluding non-significant effects. Moreover, participants attempted more of the paired trials compared with mixed trials, but the main effect of the number of alternatives was not significant. For the accuracy on only the attempted items, there was a significant sex difference, which also interacted with the type of alternatives. Despite these variations, sex differences for the accuracy on attempted items and for the fully correctly solved trials were of similar magnitude as for the overall correctly solved items.
Discussion
Because sex differences are often assumed to be in mental rotation ability, the manipulation of VK tests, which produce the largest sex differences, is suspected as a means to uncover the reasons. However, the fact that sex differences do not occur or are much smaller on other tests of the same ability is often neglected. Thus, we have theoretically reviewed differences and similarities both overall and between individual trials of different mental rotation tests. Based on this comparison, it seems likely that the trial design is an important factor in the search for sex differences and we have experimentally investigated the number of alternatives and pairwise mirroring of alternatives.
The experiment provides evidence that the trial layout indeed influences the performance. The overall performance was lower for more alternatives and for mixed alternatives. There was, however, no evidence for the main hypotheses concerning varying sex differences between trial layouts. Given that we unexpectedly did not observe meaningful sex differences in overall performance, it is rather unsurprising that the search for explanations for such sex differences was fruitless. A major point of discussion must thus be whether sex differences in mental rotation tests exist at all or whether the non-detection of sex differences is related to our test or our sample. In addition, we also want to discuss the observed overall effects of the trial design and further results of the exploratory analyses as well as findings and future directions following from the theoretical review and comparison of tests.
Sex differences in mental rotation and the relevance of the studied sample
To interpret the results as supporting or not supporting sex differences, we first need to establish the validity of the results. An issue of concern regarding our sample is the overall low performance and the large number of outliers. As Jost (2021) lined out, this could seriously reduce the interpretability of results. If participants were mostly guessing, the result could not be meaningfully related to performance and non-differences between men and women would not be surprising. In our case, almost one-third of the recorded participants were deemed outliers. With 230 of these performing below chance level as an indication of guessing, this would suggest that there are equally many guessers performing above chance level (due to the symmetry of the binomial distribution). The symmetry in the graphical inspection of the outliers suggests that many of the additional outliers were guessers performing above chance level, leaving around 100 guessers in the final dataset. This is still much more than observed in many other mental rotation studies, which may be due to the online nature, the recruitment, the incentive for participation, or the overall still rather low performance leading to more participants performing below chance level despite not guessing due to a binomial distribution of answers. However, even with this large number of outliers, the majority of participants should have performed the tasks correctly and the results should be interpretable. This is also supported by the additional more restrictive outlier detection, which should have excluded almost all guessers (at the cost of excluding more non-guessers) and produced the same pattern of results. Both the originally and more restrictively detected outliers also were roughly equally men and women, suggesting that regularly detected male performance advantages are not due to more (undetected) female guessers. Nevertheless, it should be noted that more guessers included in the sample should reduce overall effect sizes (because the effect size for guessers is 0).
Despite the low performance, it is thus now valid to discuss that we could not find the expected sex differences in performance. In most studies of mental rotation, participants are psychology students (or very similar), whereas our study targeted people from the general population without STEM affiliation. The final sample was less educated, with almost half of the participants not having the prerequisite school degrees for academic studies, and older than most university students. However, sex differences in mental rotation performance have not only been identified for students. Even online tests with large samples of the general population have identified a large male advantage contrary to the results presented here (J. K. Krüger & Suchan, 2016; Peters et al., 2007). One difference in the samples could be the exclusion of STEM affiliated persons in our study. As STEM affiliation is linked to better mental rotation performance (Hausmann, 2014; Moè et al., 2018; Peters, Laeng, et al., 1995), the larger proportion of men in STEM fields could be one cause of larger sex differences. This explanation, however, cannot explain the sex differences for university students and is thus unlikely to fully explain non-differences in performance. A further reason for the lower performance of women observed by J. K. Krüger and Suchan (2016) might be the fact that their control group excluded women with university degrees. Peters et al. (2007) did find that education was an important factor for mental rotation test performance even when accounting for age, which we could only partially replicate here. It must be noted that we only used a linear effect of education which surely cannot cover the differences between the recorded education levels and a possible nonlinear effect of overall education on performance. We also used different education levels compared with Peters et al. (2007) and their results suggest a nonlinear relationship. Regarding the age of participants, Peters et al. (2007) found a monotonous decrease in test performance for age groups >20 years, which was also supported by our data.
As we did observe sex differences in performance when accounting for age and education as well as when more restrictively excluding outliers, 9 the sample characteristics in this study might be one key reason for us not finding overall performance differences and thus the difficulty in finding explanations for these. Our results can thus be interpreted as at least not contradictory to sex differences in mental rotation test performance in the general population. However, as the influence of the trial design on test performance could interact similarly for age, education, and sex, our results cannot provide evidence for or against the suggested explanations due to the trial design.
To conclude the discussion about our sample, it is necessary to search for explanations for sex differences on samples that actually produce sex differences. In the case of mental rotation, this means controlling especially for age, but possibly also for education, STEM affiliation, and previous experience with mental rotation. Moreover, the difficulty of the test should likely be adapted to the target sample to avoid flooring effects and the mixture of low performers and guessers as well as ceiling effects. This is already often done for children, but similar adaptations could be done for elder and less educated participants.
The influence of the trial design
Despite our failure to gather meaningful evidence for the main hypotheses, we can draw some insights about the trial design from the results. While there is conflicting evidence for the interaction, both the pilot study and the final experiment provide evidence that performance decreases for mixed and for more alternatives. Thus, the complexity of the test is influenced not only by the task and the items but by the layout itself. In theory, all mental rotation tests should test the same mental rotation ability. The identified systematic variation of test performance, however, suggests that it is not that simple. Different trial layouts could tap differently into other areas such as working memory. One speculated difference could be a stronger involvement of spatial working memory in trials that require or allow more comparisons. Spatial working memory has shown involvement in VK tests (Kaufman, 2007) and a male advantage (Voyer, Voyer, & Saint-Aubin, 2017), which could thus explain male advantages in VK tests but not in SM or JJ tests.
The results also support common results of SM and JJ tests that accuracy decreases with angular disparity, which is interpreted as the involvement of mental rotation ability. The observed improvements between blocks also align with the same observation by Peters (2005), with improvements between blocks of SM tests (Heil et al., 1998), and between multiple VK tests (Meneghetti et al., 2017; Peters, Laeng, et al., 1995). For learning within blocks, however, our results are somewhat surprising. Despite controlling for unattempted trials at the end of blocks, we observed a decreasing performance. As we would have expected better performance due to learning between trials similar to the observed learning between blocks and within JJ tests (Jost & Jansen, 2020, 2022), this at least points to non-trivial effects of speed–accuracy trade-offs.
The exploratory analyses revealed some additional insights for the current dataset. Interestingly, significantly fewer items were attempted for the mixed alternatives compared with the paired alternatives but not for the trials with more alternatives compared with fewer alternatives. Despite both being harder as evidenced by the lower accuracy scores, participants took more time to reach their solutions only for the mixed alternatives. A possible reason could be the larger number of trials for fewer alternatives, including more breaks between trials and requiring more additional button clicks to proceed between trials. In another exploratory result, the effect of block was not significant for the accuracy of the attempted items. The increase in overall accuracy throughout the blocks was thus mainly due to more attempted items. These indicate, again, that effects of practice and the trial layout could influence speed–accuracy trade-offs and the perceived difficulty of items differently.
The detected small sex differences were also due to lower accuracy on attempted items and not fewer attempted items, which is contrary to the observation of significantly fewer attempted trials by women in the study of Peters (2005). On the contrary, these could explain the sex differences in psychometric tests without time limits. Given the small magnitude of sex differences observed here and the exploratory nature of these analyses, the transfer to the general population needs further confirmatory hypothesis testing.
The influence of test difficulty
In the first part, we also discussed a possible effect of test difficulty. Although it seemed unlikely to be a sole explanation for sex differences, test difficulty may have had effects on both observed (non-) effects of sex differences and effects of trial design. First, sex differences are often enhanced at increased difficulty, and this may be both due to mathematical reasons (ceiling effects or the variance of the binomial distribution at different probabilities) or theoretical reasons (sex differences interact with test difficulty). In our case, it is unlikely that the test was too easy as average performance was even lower than expected and the number of outliers was higher than expected. Conversely, it is possible that the test was too difficult. In turn, it is possible that flooring effects rather than ceiling effects conceal the effects of interest. This is supported by the mathematical consideration that it should be increasingly difficult to lower performance the closer the performance gets to chance level, similar to increasing performance the closer the performance gets to full scores. This, however, is statistically not reflected in neither the binomial distribution nor the normal distribution used for analysis and the observation of effects might be further hindered by the larger relative variance of the binomial distribution the closer the probability gets to .5. These could have resulted in both the lower-than-expected sex differences as well as the conflicting results between this study and the pilot study regarding the effects of the interaction of the number and pairing of alternatives. As a side note, test difficulty is typically equated with inverse average performance, but these may not necessarily be the same.
Theoretical review
Despite not being able to identify or refute reasons for sex differences in some mental rotation tests, the combination of the review of existing theories and the experimental results provides some insights as well as future directions. Most importantly, the experimental results suggest that performance is dependent on features of the trials. This means that widely used mental rotation tests measure more than just mental rotation ability. Because sex differences vary between tests, it is likely that the reason for sex differences lies in the aspects exceeding mental rotation ability. The proposed theories provide some possible explanations, but to further research it is necessary to first identify the exact test features and relatedly measured abilities and then, second, distinguish between possible theories. It must be noted that this may integrate all existing evidence explaining sex differences in VK tests independent of the consideration of test features. This evidence should, however, be related to the test and this may interact with mental rotation ability, but it should not be related to mental rotation ability in isolation.
Whereas much research has focused on reducing sex differences in VK tests, another interesting question is how to increase sex differences in tests. This could allow easier detection and thus more easily obtainable evidence to identify features or distinguish between theories. We have observed a monotonous effect of the number of alternatives on performance in general in the pilot study and this could be further employed in the search for sex differences, but there are further possibilities. Boone and Hegarty (2017) identified the incorporation of structural distractors to achieve sex differences in SM tests, but the removal of structural distractors in VK tests still produced large sex differences. Using polygons as stimuli has produced large sex differences even in SM test performance (Heil & Jansen-Osmann, 2008; Jansen-Osmann & Heil, 2007), but they have not yet been investigated for VK tests. It seems possible that polygons as stimuli would produce even larger sex differences in VK tests. Whereas polygons have been varied in complexity, similar approaches have not been employed for cube figures. They have only been compared with other item types despite the multiple possible modifications due to the abstract nature. While not directly contributing to reducing sex differences, exploring the effects of the trial design and the possibility to also increase sex differences could help us pinpoint the exact reasons and thus offer further understanding of the occurrence of sex differences in performance.
Conclusion
In summary, we have reviewed and experimentally investigated the effect of the trial design on sex differences in mental rotation performance. The fact that sex differences vary strongly between different test versions has often been neglected in the search for explanations. Comparing the test versions has uncovered parameters of the trial designs, which were then evaluated experimentally. The results show that the overall performance depends on these parameters of the trial design. We, however, could not evaluate their effect on sex differences as we did not observe meaningful sex differences likely due to demographic differences in our sample.
Outlook
Mental rotation is one special case of an ability where two tests are widely used, which both are supposed to measure the same ability, yet differ by quite a bit. Applying features, which have been found to reduce or produce sex differences in one test, to the other test could enhance our understanding of the occurrence of sex differences. Moreover, aspects identified to reduce sex differences in VK tests should be investigated regarding whether they simply reduce sex differences or instead increase female performance. Better understanding of the effects of the test design could also transfer to other cognitive tests.
Regarding the test design itself, we only investigated the features we deemed to be the most promising reasons for sex differences. Further potential trial and test manipulations include the upright orientation of figures, more complex rotations, the inclusion of multiple targets, or the presentation of multiple trials at the same time. On the contrary, we might have missed some features of VK tests in the tests employed in this study, which could potentially also be influential for sex differences.
Supplemental Material
sj-docx-1-qjp-10.1177_17470218231200127 – Supplemental material for The influence of the design of mental rotation trials on performance and possible differences between sexes: A theoretical review and experimental investigation
Supplemental material, sj-docx-1-qjp-10.1177_17470218231200127 for The influence of the design of mental rotation trials on performance and possible differences between sexes: A theoretical review and experimental investigation by Leonardo Jost and Petra Jansen in Quarterly Journal of Experimental Psychology
Footnotes
Author contributions
L.J.: conceptualisation, data curation, formal analysis, funding acquisition, methodology, project administration, software, visualisation, writing—original draft; P.J.: conceptualisation, supervision, writing—review & editing.
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Ethical standards
The experimental procedure was approved by the universities’ Ethics committee (Ethikkommission bei der Universität Regensburg, No. 20-2096_1-101).
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: Data collection of the present study was funded by PsychLab, a service of the Leibniz Institute for Psychology (ZPID).
Informed consent
Informed consent was obtained from all individual participants included in the study.
Data accessibility statement
This study’s design, hypotheses, and material were preregistered at https://doi.org/10.23668/psycharchives.5394. Data for this study are publicly available at https://doi.org/10.23668/psycharchives.8110. All material, data, and code for this study are also available at the associated GitHub repository https://github.com/LeonardoJost/MCMR and can be accessed at
.
Supplemental material
The supplementary material is available at qjep.sagepub.com
Notes
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
