Abstract
Pediatric intelligence tests, such as the Wechsler Intelligence Scale for Children, are commonly used diagnostic tools used in the process of diagnosing learning and behavior disabilities. Decisions concerning treatment are made based on the results of these tests and they are frequently used in educational and vocational contexts for important decisions that impact persons’ academic or professional lives. Research has however shown that important errors may occur despite the application of validation processes and adherence to quality criteria for psychometric tests. At the same time this evidence seems not to be pervasively acknowledged in psychological practice and research. In this article, I will showcase research that places attention on sources of measurement error in pediatric intelligence testing, discuss a process-performance approach to measurement in intelligence testing, and propose the “pretest methods,” methods stemming from the field of survey methodology commonly used in questionnaire construction, as a method to help address the problem of sources of measurement error in pediatric intelligence testing and improve the development of these intelligence tests.
When diagnosing and treating learning and behavior disabilities in children, intelligence tests are commonly used diagnostic tools. Based on the results of these tests, decisions are made concerning educational future and treatment.
Two tests are the most commonly used intelligence tests for children, namely the Wechsler Intelligence Scale for Children, and the Woodcock-Johnson Tests of Cognitive abilities (Freeman and Chen, 2019). Prior to the WISC-III, the WISC-R was the most popular individual test of intelligence for children (Oakland and Hu, 1992). All over the world, numerous children were administered the Wechsler Intelligence Scale for Children–Revised (WISC-R; Wechsler, 1974), the Wechsler Intelligence Scale for Children–Third Edition (WISC-III; Wechsler, 1991), and The Wechsler Intelligence Scale for Children–Fourth Edition (WISC-IV; Wechsler, 2003) according to Watkins et al. (2006). It is used, among others, in the US, to determine entitlement for special education services (Kamphaus et al., 2000). Currently, the Wechsler tests including the latest version, the WISC-V (Wechsler, 2014), are still the most frequently used measures in clinical practice and research context (Freeman and Chen, 2019; McCrimmon et al., 2015).
In this respect, Stout (2002) talks about the issue of test fairness and relates it to cultural and ethnic diversity. In addition, neurodiversity and individual personality diversity in the form of person related confounding variables may also be aspects to be considered. Plante and Sykora, for example, (1994: 749) proposed that “(. . .) these decisions may be based upon test scores, with little consideration for emotional variables that might impact negatively a child’s ability to perform up to his or her potential on standardized tests.” Thus, the issue of measurement error and test fairness might not only relate to groups from different cultural contexts but also to test performance being influenced by motivational and emotional attributes (Steinmayr and Spinath, 2019: 392).
It is possible that high- and low-IQ scoring individuals differ in other ways than for example “neural efficiency” (Eysenck, 1973; Vernon, 1987), just to mention one of the many views on what (psychometric) intelligence is. For example, According to Neisser et al. (1996) variables that might be implicated and impact speeded performance include motivation, response criteria (emphasis on speed vs accuracy), perceptual strategies, attentional strategies, and even differential familiarity with the material itself.
Could, in a similar way, certain variables also be of influence on other types of psychometric testing other than the speeded performance test mentioned by Neisser et al. (1996)?
This would be problematic as in general the morale is that intelligence tests should measure the construct they are intended to measure and should not be biased. Better said, there should be no measurement error—neither systematic nor too large unsystematic measurement error. That is, we would not want test scores to be influenced by construct-irrelevant sources of variances (McNamara and Roever, 2006). However, as we will discuss here, there is evidence of such measurement error occurring in pediatric intelligence testing while at the same time this evidence seems not to be pervasively acknowledged in psychological practice and research. Despite numerous theories of intelligence, in general one could say IQ tests measure general cognitive ability (Neisser et al., 1996). Considering that extraneous factors can play a role in either the outcomes of an IQ test, or cognitive ability in general (i.e. the construct) it seems of utmost importance to gain insights into these factors, and in ways to consider these factors in WISC IQ test-measurements.
In this article, I showcase research that places attention on sources of measurement error in pediatric intelligence testing, I will discuss a process-performance approach to measurement in intelligence testing, and propose the pretesting methods, stemming from the field of survey methodology and commonly used in questionnaire construction, as a method to help improve the development of pediatric intelligence tests and help address the problem of sources of measurement error in pediatric intelligence testing.
A few words on the Wechsler intelligence scale for children
As mentioned, the Wechsler tests are among the most frequently used measures in clinical practice and research context. This despite the fact there have been doubts considering the quality of the WISC tests (see e.g. Baron, 2005; Schmukle and Schulze, 2016).
In a test review of the WISC-IV, Burns and O’Leary (2004: 234) report that “In an effort to increase developmental awareness, instructions to the child were reworded to simplify more difficult concepts. The WISC-IV has increased the use of teaching items, queries and prompts to increase the child’s understanding of the task.” Though this might not have been sufficient as, for example “. . . some clinicians have reported the use of the word “advantages” within the Comprehension subtest is poorly understood by some children” (Burns and O’Leary, 2004: 234). This example shows that a careful and detailed scrutiny of the test items is not in vain, and specifically looking at the response process and understanding of the questions could be very fruitful. In questionnaire construction, understanding of the question or task at hand by the respondent, is seen as a very important aspect when preventing measurement error (see e.g. De Leeuw et al., 2008). Comprehension by itself (understanding of the survey questions), is quite an extensive concern in survey methodology (e.g. Sudman et al., 1996; Tourangeau et al., 2000).
The WISC-V (age range 6–16 years) continues the evolution of the WISC series and represents a substantial update over the previous version. However, independent analysis of the WISC-V has resulted in several substantial critiques of the test and information presented in the manuals (for a review, see Goldstein et al., 2019). Specifically relevant for present discussion is that the WISC-V manual presents traditional metrics of reliability that are biased when multidimensionality is present (Raykov, 1997). Importantly, while an understanding of the abilities being tapped by the index and subtest scores are fundamental to using this test, these scores must also be placed in the larger context of each child and their “world.” There are a huge number of endogenous factors such as genetics and a host of external factors such as culture and education that impact not only the growth and expression of intelligence, but also performance on intelligence tests (Weiss et al., 2016).
Sources of measurement error in pediatric intelligence testing
According to Weiler et al. (2019), several sources of error may obfuscate the validity of pediatric neuropsychological evaluations, including issues of incremental validity, demographic characteristics of children, ecological validity, potential for malingering, and errors of clinical decision making. Test of intelligence are generally an essential part of such a neuro-psychological evaluation. Notwithstanding, an up-to-date comprehensive overview of sources of measurement error specifically for pediatric intelligence testing is missing (but see e.g. Hanna et al., 1981; Styck and Walsh, 2016). Hanna et al. (1981) estimated the magnitude of four sources of measurement error, namely content sampling, time sampling, scoring and administration in the WISC-R. Some other sources of measurement error in the Wechsler scales may be factors related to the interaction between the examiner and the examinee and situational factors (Styck and Walsh, 2016). Hanna et al.'s (1981) composite was found to produce a far larger standard error of measurement than reported by the manual (Wechsler, 1974). Current estimates of measurement error for most IQ tests rely solely on internal-consistency reliability and therefore the standard error of measurement published in test manuals may not adequately capture error (Styck and Walsh, 2016).
In the following paragraphs I will discuss a few specific sources of measurement error in pediatric intelligence testing and where possible specifically the Wechsler scales found in empirical research.
Emotional and motivational variables and interaction with pediatric test characteristics
Already in the seventies, it was shown that emotional stress can result in decreases in cognitive performance in children (see e.g. Fuerst et al., 1990; Mandler, 1975, 1984; Meyer et al., 1989). Problem-solving and memory processes involved in aptitude testing had as well been shown to be affected by stress levels in children (Kaufman, 1979).
For example, the Arithmetic, Coding and Digit Span subtests of the WISC-R had been found in numerous studies to be more susceptible to anxiety, depression, and distractibility than the other WISC-R subtests (e.g. Kaufman, 1975, 1979; Ogdon, 1982; Stewart and Moely, 1983). In two other studies, Plante et al. (1993) and Plante and Sykora (1994) demonstrated that stress and coping variables were associated with aptitude and achievement testing where the WISC-R and WJ-R were used as measures, and with WISC-III performance respectively. However, the studies indicated the necessity of further research to examine the role of specific childhood stressors on testing performance. The instigator of these studies was an experiment on the effects of Thorazine on Wechsler scores of adult catatonic schizophrenics. Gilgash (1957) found that the intellectual functioning of the experimental group was decidedly improved after the introduction of Thorazine and 30days of Thorazine medication. Upon these results, it is hypothesized that, as schizophrenia is usually regarded as an affective rather that an intellectual disorder, the improvement in intellectual functioning was due to an improvement in the patient’s affective responsiveness, which indicates that emotional variables have an impact on the results of an IQ test.
In Steinmayr et al.’s (2014) study, SEM analyses revealed that all in their study included motivational variables played a role in explaining the relation between gender and numerical intelligence in a population of 305 German pupils (average age 17.54 years, SD = 1.08). Specifically, self-estimated numerical intelligence, intrinsic value of math, and worry significantly predicted numerical intelligence. Gender however, still explained a statistically significant amount of variance in numerical intelligence. This could be due on the other hand to other aspects, such as stereotype threat, socialization, and characteristics of the test such as time restriction (Steinmayr et al., 2014). Carr and Davis (2001) proposed that boys compared to girls display better scores if there is a time restriction regarding the tasks on standardized tests, and Wilhelm and Engle (2005) show that imposing a time limit on an intelligence test might threaten its validity. Differences on certain intelligence tests between culturally different groups are explained to a large degree by time limits that are too strict on such intelligence tests, bias which in turn is explained by the different socialization of the groups (Kersting, 1996; Knapp, 1960). In the case of numerical intelligence tests (Steinmayr and Spinath, 2019), girls are disadvantaged when this test is presented with a time limit too strict, which is largely explained by emotional and motivational factors. When looking at a specific emotional variable, test anxiety, it has been demonstrated that anxious students in particular suffer from time constraints and that their performance increases when time constraints are relaxed (see Onwuegbuzie and Seaman, 1995).
Fear of failure, among other motivational variables, might function as a significant mediator of the relationship between performance on the administered intelligence test and gender in adolescents (Steinmayr and Spinath, 2008). Motivational variables have been often identified as mediators of the relationship between gender and achievement in standardized mathematic achievement tests (Steinmayr et al., 2014). These motivational variables then interact with traits of the test, such as the mentioned time restrictions. Looking at another aspect of test design, namely test instructions, results of experimental studies (e.g. Johnson et al., 2012) have shown that specific test instructions can be used to increase girl’s performance on different mathematical achievement tests
In sum, a characteristic of the test, in this case time restriction, or test instruction, interacts with personal characteristics of the subject, favoring certain people with certain characteristics over others, incurring in measurement error and therefore a loss in validity. This would apply to differences between groups, but also to characteristics that may or may not be attributed differently to different groups and thus not easily identifiable and operationalizable. In the case of mentioned gender-research, we have an easily identifiable and measurable variable–gender–that directs the attention to certain differences, where only an in-depth scrutiny points to other variables as the causal attribution. At the same time, other factors that might influence the results of an IQ test might have not been identified and escaped our attention.
Several motivational and emotional variables have been found to be significant predictors of performance in standardized mathematical performance tests in schools in the target population of the WISC, such as expectations for success (e.g. Dickhäuser and Reinhard, 2009); interest and mathematical ability self-concept (e.g. Steinmayr and Spinath, 2009); and test anxiety (Hembree, 1990). It awaits further research to investigate whether they could also be predictors of performance in numerical and general intelligence tests.
Examinee interest and cooperation throughout the session (Glutting et al., 1994) and familiarity with the examiner (Fuchs and Fuchs, 1989; Szarko et al., 2013) have also been shown to be related to variance in FSIQ scores in children.
Segal (2006) asserts that low-stakes test scores, measured in surveys, may be partially determined by test-taking motivation, which is associated with personality traits and not with cognitive ability in a mixed population of mostly minors. Moreover, correlations found in survey data between high test scores during adolescence and economic success later in life seem partially caused by favorable personality traits. The relevance of motivation concerning performance on pediatric intelligence tests has been investigated also by Duckworth et al. (2011) in a meta-analysis. The authors found an effect of material incentives on performance in the intelligence test in a large study population (n = 2008) of which 1912minors (<18) and partly using the WISC. Moreover, observer ratings of test motivation were associated with both WISC-R scores and important life outcomes in a population of adolescent boys (n = 251, average age 12.5 years). Children who tried harder earned higher WISC-R scores. Thus, performance in WISC-R varies depending on the child’s motivation in the testing situation and furthermore test motivation “can act as a third-variable confound that inflates estimates of the predictive validity of intelligence for life outcomes” (Duckworth et al., 2011: 7716). Importantly, Extrinsic motivation thus seems to be important for performance on pediatric intelligence tests.
Measurement error in pediatric tests related to the scorer
Examiner error or scorer error is a source of error in the use of intelligence testing that has been amply researched, also specifically for Wechsler pediatric scales (e.g. McDermott et al., 2014; Styck and Walsh, 2016). Examiner or scorer error may have an unsystematic component, but may also be influenced by a more systematic component, such as a halo effect, a generosity error, race and sex stereotypes or other sources of consistent error (Hanna et al., 1981). Just as in a questionnaire, also in a WISC-test, to an unknown extent “responses given are a function of the manner in which the examiner conducts the examination” (Kaspar et al., 1968: 475).
In an early observational study on examiner coding errors on individual test taking with WISC and Binet protocols, Warren and Brown (1972) use a checklist of errors frequently encountered that they developed over years of experience. A total of 1939 errors were found in the 1873 subtests examined, 350 of which were consequential for the IQ reported. Some of these types of errors were: failure to record response, failure to follow procedure given in manual, scoring and tabulating. Some of these different types of examiner errors have been investigated later on in different studies, namely clerical errors, scoring errors, administration errors, failure to record examinee responses as an error, or “correctly” scoring open-ended items (for an overview see Styck and Walsh, 2016), but not altogether.
However, it is unclear how these findings have been translated into practice or improvement of IQ-tests. Practicing psychologists do not always seem to consider the possibility of measurement error in adult or pediatric intelligence testing or specifically measurement error related to themselves—the assessor. In a recent meta-analysis, Styck and Walsh (2016) conclude that assessor errors occur frequently and impact index and FSIQ scores on the WISC. Similarly, McDermott et al. (2014) found that nearly all WISC-IV scores conveyed significant amounts of variation that could not be accounted to children’s actual differences and that the FSIQ and VCI (Verbal Comprehension Index) scores evidenced substantial assessor bias. Consequently, current estimates for the standard error of measurement of the WISC may not adequately capture the variance due to the examiner.
A process-performance approach to intelligence test results: Stable trait or performance and interaction with the test environment?
Generally, the results of an intelligence test are considered an indicator of a stable characteristic (Eckert et al., 2006), also by practicing (school) psychologists and in schools. However, some authors indicate that it should be also taken as a kind of performance indicator (Eckert et al., 2006; Steinmayr et al., 2014). In intelligence testing it is commonly ignored that individuals differ among each other concerning motivational factors such as academic self-concept, which endangers the validity of intelligence testing, as research shows that test results under certain conditions are influenced by motivational factors (Eckert et al., 2006).
There have been studies about personality and performance in multiple-choice testing and other achievement tests. Kubinger and Wolfsbauer (2010: 303) contend that from the point of view of personality psychology, examinees in a multiple-choice test are likely to differ in the way they deal with multiple choice items, which might impact their results. Dochy et al. (2001) on their side presume that, if an examinee reaches a particular answer for an item in a multiple-choice test, which, however, is not offered as an option, then, the higher the examinee’s general assertiveness is, the more he/she uses “none of the other options is right.” Ávila and Torrubia (2004) established that extraverted, impulsive, and low-anxiety examinees give more incorrect responses in multiple-choice tasks but make fewer omission errors and Alker et al. (1969) indicate that examinees with a “nonconformist” personality score higher on achievement tests than others because of fewer skipped items and not at all using the “I don’t know the solution” option. Lastly, Stoeber and Kersting (2007) showed that “perfectionist” examinees attain higher scores in achievement tests.
To the best of my knowledge, similar researches have not been carried out in the investigation of personality, emotional and motivational variables, and response process and performance in pediatric intelligence testing. Different personality types might nevertheless also react differently to an intelligence test, because of which the design of the test considering response process and the interaction with the responder deserves attention, such as commonly is done in questionnaire construction procedures (e.g. De Leeuw et al., 2008).
Thus, it is standard to interpret test scores on intelligence tests as an ability and a rather stable trait. However, intelligence test results are preceded by a response process in which persons work on intelligence test tasks and give a performance that—to a better or worse extent—may reflect their intelligence. From this point of view, a score on an intelligence test can be interpreted as an indicator of performance on this intelligence test (Steinmayr et al., 2014: 141), reflecting both the response process and the potential. This response process, then, is a result of the interaction of aspects such as the test characteristics, personality, and motivational and emotional variables.
Validity: Multidimensionality
From the psychometric point of view, “(. . .) it has long been known (Stout, 1987) that items, for example in a test, an interview or behavior observation, do not only capture the intended trait but also other aspects, for example, other traits or specifics of the test person (. . .)” (Ziegler and Hagemann, 2015: 232). That is, assuring the unidimensionality of an item is an extremely difficult if not impossible task (Ziegler and Hagemann, 2015: 232). Though when constructing items researchers at least try to ensure the items are not loaded with other traits, unfortunately addressing this aspect is not part of general procedures on test construction (Ziegler, 2014). Following the ABC of test construction according to Ziegler (2014), psychometric criteria such as reliability and validity are of large importance, to which Ziegler and Hagemann (2015) added unidimensionality, which can however be seen as generally treated as a sub aspect. Concerning reliability, reliable findings not necessarily translate into valid findings.
Validation syntheses have shown that internal structure and relations to other variables have been the dominant sources of evidences in the literature (e.g. Greiff and Scherer, 2018; Zumbo and Chan, 2014). The first issue here is that even these procedures could not suffice. EFA (Exploratory Factor Analysis) is one of the classical approaches to use in test construction. EFA results obtained during test construction are often used to select items (Ziegler and Hagemann, 2015: 234). With regards to unidimensionality, this item selection is deemed to ensure that no other relevant variance sources are part of the items. Ziegler and Hagemann (2015) show that this can be misleading if all items contain the same (two) sources of variance (see Ziegler and Hagemann, 2015, figures 1a–f: 232–233). Thus, all items loading on one factor does not necessarily imply unidimensionality of the items.
When considering (rather) stable person traits, such as traits that can affect the response process such as perfectionism, fear of failure or insecurity, that is, motivational and emotional variables, or non-stable person traits such as mood, but that may remain stable over the course of the whole response process, these could be such a source of variance contained in all items. This would be overlooked by current common test construction procedures, and would remain unidentified even with repeated measurements over the years in the case the related person traits would not change in this period. There is thus room for a multidimensionality which can be caused by person traits or person related variables.
Nonetheless a step forward compared to EFA, similar considerations hold for CFA (Confirmatory Factor Analysis), and IRT (Item Response Analysis). In response to this, Ziegler and Hagemann (2015) propose certain solutions, using MTMM and other methods. These solutions might however also have their limitations.
This issue can be additionally looked at from the viewpoint of test fairness (Stout, 2002). A procedure to increase test fairness is to do differential item functioning analyses (DIF), and related, differential bundle functioning (DBF) and differential test functioning (DTF) analyses. “. . . by definition, DIF occurs when examinees, matched on the latent variable θ that the test is intended to measure, perform differentially depending on their group membership.” (Stout, 2002: 501). This, however, would become a problem when group membership is not clear. When test fairness is focused at groups formed based on variables such as gender, social and cultural background, or race, then these are straightforward identifiable and measurable variables. The picture changes when it concerns groups that are not directly observable or even measurable, or are measurable but have not yet been (extensively) identified and measured. As would be the case with groups based on motivational or emotional variables, or personality. A test fairness analysis requires a prerequisite validity analysis, that is, any DIF/DBF/DTF analysis is not effective in the case of a serious contamination of the subtest caused by secondary dimensions, moreover, any DIF/DBF/DTF procedure, “is only as effective as the matching subtest is in matching examinees on the construct intended to be measured.” (Stout, 2002: 501).
Multidimensionality and the response process
Unidentified secondary dimensions or confounding variables thus pose a limitation toward differential functioning analyses. “. . .In the prototypical large standardized test setting where one wants to assess pretest items for possible DIF, matching examinees on the score on operational items, which have previously undergone careful test design and psychometric scrutiny, seems reasonable” (Stout, 2002: 501). However, if also in these prior developed tests or test items secondary dimensions remain unidentified, then this is of no use. In survey research, this phenomenon is referred to as response bias and is an extensively researched phenomenon, also with qualitative procedures. Problematic response bias occurs when the variables related to the response bias are not known and thus not measured.
Ziegler (2015: 156) also gave attention to the matter of response bias where it concerns psychological assessment in the form of for example personality tests used for the selection of job applicants, stating that “. . . all research dealing with such phenomena [response bias] should take the answer process into account.” His focus was thus on response bias related to the test takers in what Cattell called Q-data (Cattell, 1958), and not yet on psychometric intelligence testing. Cattell (1958: 286) defined Q-data as: “Observations of personality which come to the psychologist in terms of introspective, verbal, self-record and self-evaluation. . .”
Nevertheless, it seems that response processes and test consequences have been largely neglected in validation practice in the area of psychology (Hubley, 2018; Zumbo and Chan, 2014), though there has been some research into sources of model misfit that point to test characteristics and the response process, namely item wording (DiStefano and Motl, 2006; Rauch et al., 2007; Vautier et al., 2003) and position effect (Hartig et al., 2007; Schweizer et al., 2009). These aspects indicate correlated errors (which in turn might indicate undiscovered secondary dimensions) which are however normally not taken into account in test construction (Schweizer, 2010). On the other hand, item wording and position and their effect on outcomes are extensively researched in survey methodology (see e.g. Schwarz et al., 1991).
Survey research and practices and research in questionnaire construction have particularly addressed the role of psychology in the response process and its relationship with measurement outcomes. Specifically, survey methodologists have placed attention on understanding the cognitive and communicative processes underlying survey responses, increasingly turning the “art of asking questions” (Payne, 1951) into an applied science that is grounded in basic psychological research.
Thus, with the unsolved problem of multidimensionality and empirical research that shows the influence of extraneous variables on measurement in pediatric psychometric tests in mind, it seems fruitful to start looking more closely and in complementary alternative ways at the validity of psychometric tests like the WISC. One way to do this would be to look at the response process such as generally done in the practice of questionnaire construction and survey methodology (see e.g. De Leeuw et al., 2008), by considering results on intelligence tests as indicators of performance in an intelligence test, and considering that variables like motivation are involved in the process that forms that performance. These are then considered confounding variables that impact the target measure. Of course, as Ziegler and Hagemann (2015) propose, these variables can be investigated as independent variables in an experiment, with the target measure as a dependent variable. “An experimental effect of the independent variables on the dependent variables would reveal multidimensionality of the target variable in terms of contamination with undesirable constructs.” (Ziegler and Hagemann, 2015: 236). To be able to do this however, these variables have to first be identified and operationalized. Besides, many of the variables that seem relevant from previous research are not experimentally manipulable.
In addition to an exploration with the goal to identify possible confounding variables or dimensions, the mechanisms behind relationships (e.g. gender differences in intelligence test results) can be explored by means of looking in detail at the response process. Insights stemming from this exploration can then be used to improve pediatric psychometric tests, for example by developing more extensive practices of pretesting in the development of these psychometric tests.
Procedures commonly used in the pretesting phase of questionnaire construction are for example Respondent debriefing, Behavior coding, Cognitive Interviewing (consisting of the use of Probes and Think-aloud protocols), Focus groups, but also the coding of interview transcripts. These procedures are known as “cognitive laboratory methods” or “pretest methods” (Campanelli, 2008; Schwarz et al., 2008). They can give insights on processes taking place during intelligence testing that influence validity, bias and test fairness. One can think of for example the interaction between test taker and examiner and the interaction between personality traits and test characteristics. To fall back on some of the sources of measurement error discussed above, tasks with a time restriction for example seem interesting from this viewpoint, just like the interaction between the examiner and a child with anxiety. A qualitative explorative approach, looking at the response process in detail, might thus give insights in these issues.
In conclusion
Research indicates that motivational and emotional variables, but also several other variables, have a contribution to performance in pediatric intelligence testing with the WISC.
These sources of measurement error have moreover not received a lot of attention in psychological practice, and, therefore have not seeped through to this practice, where results of a pediatric IQ test such as the WISC-V are generally considered to be error-free. Neither seem have findings to have received a lot of attention in the development–and research into the development–of psychometric intelligence tests. If they have, it has been from a quantitative psychometric viewpoint, which has not been able to solve the issue satisfactorily.
Psychometric analyses use mathematical models to investigate a complex real-world phenomenon. An additional complementary viewpoint exists that can help address the problem of measurement error in pediatric psychometric testing. Measurement error due to secondary dimensions can be considered and analyzed as a continuous and quantitative variable resulting from the interaction between test taker personality and situational demands, like Ziegler et al. (2015) propose when doing research on faking in personality tests. It can also be quantitatively analyzed based on different response patterns or response styles or characteristics of the test. Yet another complementary way of researching this phenomenon is focusing on and exploring the answer process and cognitive processes taking place when completing an IQ test from a qualitative viewpoint as with questionnaire pretesting methods.
Using procedures from the field of questionnaire construction and survey research as substantive analyses of test bias, has the potential of achieving one of the imperatives for conducting test fairness research and practice according to Stout (2002: 503), namely integrating substantive analyses with psychometric analyses. It would be interesting to investigate whether standardized test procedures in IQ tests like the WISC-V, including standardized test instructions or time restriction in certain subtests, incur in bias, favoring or disadvantaging children with certain characteristics, including emotional and motivational traits. Complementary to and hopefully in synergy with psychometric procedures, this can be researched focusing on the response process and cognitive processes taking place when completing such an intelligence test.
Footnotes
Author’s note
The author is employed at the Section for Research Methods of the Department of Special Education and Rehabilitation of the Faculty of Human Sciences, University of Cologne.
Declaration of conflicting interests
The author declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author received no financial support for the research, authorship, and/or publication of this article.
