Abstract
Many decisions rest on people’s ability to make estimates of unknown quantities. In these judgments, the aggregate estimate of a crowd of individuals is often more accurate than most individual estimates. Remarkably, similar principles apply when multiple estimates from the same person are aggregated, and a key challenge is to identify strategies that improve the accuracy of people’s aggregate estimates. Here, we present the following strategy: Combine people’s first estimate with their second estimate, made from the perspective of someone they often disagree with. In five preregistered experiments (
Keywords
People often make estimates of some unknown quantities or events. In these types of judgments, a well-known phenomenon is that the average estimate of a crowd of individuals is often more accurate than most individual estimates (Galton, 1907; Surowiecki, 2005), and crowds have been used to improve judgments in areas such as economic forecasts (Clemen, 1989), medical diagnoses (Kurvers et al., 2016), weather forecasting (Baars & Mass, 2005), and scientific research (Altmejd et al., 2019; Gordon et al., 2021).
The “wisdom of crowds” arises from the (mathematical) principle whereby aggregating multiple imperfect, yet diverse, estimates diminishes the role of errors (Herzog & Hertwig, 2009; Stroop, 1932). That is, when multiple estimates are sufficiently diverse and independent, averaging increases accuracy by canceling out errors across individuals (Vul & Pashler, 2008). However, although this approach is beneficial, it is often not feasible for a single person to collect the estimates of multiple individuals. Remarkably, research suggests that the same principles underlying the wisdom of crowds also apply when multiple estimates from the same person are aggregated—a phenomenon known as the “wisdom of the inner crowd” (Herzog & Hertwig, 2009; Van Dolder & Van den Assem, 2018; Vul & Pashler, 2008).
It is not clear from the outset why aggregating multiple estimates from the same person would be beneficial. If a person’s first estimate represents their best guess, then any other estimate would simply add noise (Hourihan & Benjamin, 2010; Vul & Pashler, 2008). An alternative account based on probabilistic representations, however, posits that averaging estimates from the same person cancels out the errors that permeate people’s judgments. According to this account, people’s initial estimates represent samples drawn from an internal distribution of possible estimates, where second estimates are resampled guesses from that same distribution (Vul & Pashler, 2008; Wallsten et al., 1997). When second, resampled estimates are sufficiently diverse, averaging increases accuracy by canceling out errors across estimates (Ariely et al., 2000; Herzog & Hertwig, 2009; Keck & Tang, 2020; Litvinova et al., 2020).
With such a powerful tool available to individuals, a key challenge is to identify strategies that can help improve the accuracy of people’s aggregate estimates (Herzog & Hertwig, 2014a). Research so far agrees that the inner crowd falters when people anchor too heavily on their first estimate when generating a second guess, thereby reducing diversity and independence (Herzog & Hertwig, 2014a; Vul & Pashler, 2008). At least two methods have been applied to negate this. The first relies on the passage of time. For example, the benefits of aggregation tend to be higher with the introduction of time delays between both estimates (Steegen et al., 2014; Van Dolder & Van den Assem, 2018; Vul & Pashler, 2008). In these cases, the passage of time effectively deanchors people from their first estimate (presumably because they forget their initial estimate), thereby improving the diversity and independence of both estimates. A second method to increase diversity and independence is to rely on the mind’s ability to construct alternative, opposing realities (Herzog & Hertwig, 2009, 2014a, 2014b). A demonstrated way to do this has been through “dialectical bootstrapping,” in which people are prompted to base their second estimate on different assumptions and considerations (Herzog & Hertwig, 2009). These dialectical estimates ideally result in errors with different signs relative to first estimates, and there are different techniques to elicit such dialectical estimates. One technique, based on the “consider-the-opposite” strategy (Lord et al., 1984), instructs people to actively question the accuracy of their first estimate when generating a second guess. This technique has been shown to increase the accuracy of people’s aggregate estimate by getting the same person to generate first and second estimates that are more diverse and independent (Herzog & Hertwig, 2009, 2014b). In the present research, we similarly relied on the mind’s ability to construct opposing realities by prompting people to complement their initial estimate with a second estimate made from the perspective of someone they often disagree with.
Statement of Relevance
In today’s polarized society, disagreement is associated with conflict and division, but are there also benefits to disagreement? By utilizing people’s ability to take the perspective of others, we propose that disagreement is a powerful tool for producing accurate estimates. In five experiments, people made estimates of unknown quantities from various perspectives. Following principles of within-person aggregation, we found that aggregating people’s first estimate with their second estimate, made from the perspective of someone they often disagree with, produced accurate estimates. In explaining this accuracy, we found that taking a disagreeing perspective prompts people to consider estimates they normally would not consider to be viable options, resulting in first and second estimates that are more diverse and independent (and by extension more accurate when aggregated). Together, these results underscore the importance of perspective taking and disagreement as strategies to improve the accuracy of people’s quantitative estimates.
Disagreement is often decried as an undesirable component of people’s interactions with others. In today’s polarized society, disagreement has been associated with conflict, division, and misinformation (Kennedy & Pronin, 2008; Reeder et al., 2005; Sunstein, 2002). However, although disagreement is generally undesirable, research in group decision-making indicates that it may actually be beneficial when groups address complex problems, such as making estimates of unknown quantities or events (de Oliveira & Nisbett, 2018; Hong & Page, 2004; Mutz, 2006; Page, 2008). These effects occur because of the notion that disagreeing individuals tend to produce more diverse estimates, and by extension errors, which are canceled out across group members when averaged (Page, 2008). It is precisely this aspect of disagreement that we rely on in our pursuit to foster more diverse estimates from the same individual. More specifically, we surmise that just as disagreement between different individuals is beneficial for the wisdom of crowds, so too, through perspective taking, will this be beneficial for the wisdom of the inner crowd.
To understand the benefits of disagreement, we tested the hypothesis (in Experiment 3) that taking a disagreeing perspective leads to two distinct observations. First, from a disagreeing perspective, people are more likely to consider estimates that are strikingly different from their own guesses, thereby opening the sampling space of possible second estimates. And second, people are more likely to adopt these different estimates as their second estimates when viewing problems from a disagreeing perspective, leading to first and second estimates that are more diverse and independent. These conjectures follow from earlier work on anchoring showing that people typically avoid making second estimates that are strikingly different from prior estimates or anchors (Epley & Gilovich, 2006; Lewis et al., 2019; Tversky & Kahneman, 1992). Making an estimate from a disagreeing perspective is expected to attenuate this tendency, given that disagreeing others (almost by default) consider and adopt entirely different estimates as one’s own estimate. However, although taking a disagreeing perspective is generally beneficial, the final experiment (Experiment 4) identified a situation in which taking a disagreeing perspective backfired, undermining the benefit of averaging (i.e., in situations in which second estimates were likely to be made in the wrong direction).
The Present Research
For easier reading, we first present general methodological information that concerns all five experiments. Ethical approval for all experiments was obtained from the ethical review board at Eindhoven University of Technology (Reference No. ERB2020IEIS29). For all experiments, we report how we determined the sample size, all data exclusions (if any), all manipulations, and all measures. The questions used in all experiments can be found in the Supplemental Material available online. Data, code, and materials are publicly available on OSF at https://osf.io/qsxp8/. All experiments’ hypotheses, designs, and main analyses were preregistered 2 (see the Open Practices section for links).
Sample-size estimation for all experiments was based on a priori power analyses using G*Power (Version 3.1.9.4; Faul et al., 2007). For Experiments 1a, 1b, 2, and 3, analyses determined that 410 participants per condition would be necessary to achieve a Cohen’s
For analysis, we used mixed-effect models, which allowed us to make more generalizable claims across a wide range of participants and questions by employing random intercepts for participants and questions (Judd et al., 2012). We fitted the models using
Experiments 1a and 1b
Method
In Experiment 1a, participants made two weight estimates of 10 objects shown in pictures (see Table S1 in the Supplemental Material). In Experiment 1b, we used a different estimation task: Participants made two estimates of six questions on a scale ranging from 0% to 100%. The questions’ true answers were obtained from various online sources (e.g., Wikipedia for Experiment 1a and the
Participants were instructed not to look up the true answers during the study. They were randomly presented with the questions in two estimation stages (histograms for the distribution of participants’ answers on both estimates for all five experiments can be accessed at https://osf.io/q3tfh/). Participants were not told at the beginning of the experiment that they would be asked to make an additional, second estimate. In the first stage, participants simply provided their own estimates to the questions. The instructions for the second estimation stage were different depending on the condition. For the We will now ask you to provide a second guess at the answer to each of the [ten/six] questions you were asked in the first session. These answers should not be the same as your previous answers: these should reflect your ‘second guess’.
For the Now picture a friend whose views and opinions are very different from yours. To illustrate, when discussing politics, you often find yourself disagreeing on various issues. How would he or she answer these [ten/six] questions? Please answer these questions again, but now as this friend.
After responding to the questions, participants were asked to provide their age and gender. In addition, they were presented with a manipulation-check question instructing them to choose a particular option in a multiple-choice array and a question asking them whether they looked up any of the answers to the questions.
Results
Correlations
Comparing the two experimental conditions, we found that our instructions led to lower correlations when second estimates were made from a disagreeing perspective (Experiment 1a: mean

Signed errors in Experiment 1a (a), Experiment 1b (b), Experiment 2 (c), Experiment 3 (d), and Experiment 4 (e). Each dot represents the correlation between the error of a participant’s first estimate and the error of that participant’s second estimate in response to a particular question, separately for the disagreeing-perspective condition (in all experiments), the self-perspective condition (in all experiments), and the agreeing-perspective condition (in Experiments 2–4). Solids lines are lines of best fit.
Inner-crowd effects
For the inner crowd to be more accurate, the aggregate of both estimates should have a lower error than a person’s first estimate alone. Taking into account both conditions (overall) and looking at the self- and disagreeing-perspective conditions separately, we found an inner-crowd effect in Experiments 1a and 1b (see Table 1 for summary statistics). The average of both estimates had a lower mean square error than the first and second estimates alone, respectively (for descriptive statistics, see Tables S2a and S2b in the Supplemental Material).
Inner-Crowd Effect for Experiments 1a and 1b: Comparisons of the Average of Two Estimates With the First and Second Estimate
Benefit of averaging
Would participants in the disagreeing-perspective condition benefit more from averaging their estimates than participants in the self-perspective condition? To test this, we calculated the benefit of averaging by subtracting the square error of average estimates from the square error of first estimates (similar procedures have been used before; Herzog & Hertwig, 2009; Steegen et al., 2014; Vul & Pashler, 2008). The higher this number, the larger the benefit of averaging (i.e., the more accurate the inner crowd). Overall, the results indicated that in both Experiments 1a and 1b
4
participants in the disagreeing-perspective condition indeed benefited more from averaging their estimates than participants in the self-perspective condition (Experiment 1a:
Bracketing
To more concretely test whether people in the disagreeing-perspective condition benefited more from averaging, we looked at bracketing rates across conditions. Bracketing is a key component underpinning the benefit of aggregating multiple estimates (Larrick & Soll, 2006). It refers to the observation that if two estimates are on the opposite sides of the true answer, thus bracketing it (i.e., one overestimating the true answer and the other underestimating it), aggregating them will typically result in a more accurate average estimate (Larrick & Soll, 2006; Soll & Larrick, 2009). Consequently, for each question, we verified whether the question’s true answer was bracketed by the two estimates. As expected, the bracketing rate was much higher in the disagreeing-perspective condition at 29% (Experiment 1a) and 38% (Experiment 1b), compared with the self-perspective condition, in which 19% (Experiment 1a) and 20% (Experiment 1b) of people’s estimates bracketed the questions’ true answers (Experiment 1a:
Experiment 2
People who made a second estimate from the perspective of a person they often disagree with benefited more from averaging than people who simply made a second guess. Experiment 2 provided an important extension. Specifically, we included a third experimental condition in which participants were instructed to take the perspective of a person they often agree with. We included this condition to underscore the need to take a disagreeing perspective to improve the accuracy of people’s inner crowds.
Method
The procedure of this experiment was similar to that of Experiment 1b. However, we added an additional condition (i.e., the Now picture a friend whose views and opinions are very similar to yours. To illustrate, when discussing politics, you often find yourself agreeing on various issues. How would he or she answer these six questions? Please answer these questions again, but now as this friend.
We recruited 1,425 participants using MTurk. After excluding those who failed the instructional check and those who said they looked up the answers, we obtained a final sample of 1,389 U.S. participants (age:
Results
Correlations
The estimates’ errors were highly correlated in the self-perspective and the agreeing-perspective conditions (mean
Inner-crowd effects
There was again an inner-crowd effect overall (i.e., across all three conditions) and in the three conditions separately (see Table 2; for descriptive statistics, see Table S3 in the Supplemental Material).
Inner-Crowd Effects for Experiments 2 and 3: Comparisons of the Average of Two Estimates With the First and Second Estimate
Benefit of averaging
There was no difference between the self- and agreeing-perspective conditions with regard to benefit of averaging (
Bracketing
With 21% and 20% of people’s estimates bracketing the questions’ true answers, there was no difference in bracketing rates between the self- and agreeing-perspective conditions (
Experiment 3
In Experiment 3, we tested the proposed mechanism explaining our observation of more diversity and independence when second estimates are made from a disagreeing perspective. Earlier work on inner crowds suggests that people typically anchor too heavily on first estimates when generating a second guess, thereby not producing diverse enough estimates and errors (Herzog & Hertwig, 2009; Vul & Pashler, 2008). Making an estimate from a disagreeing perspective was expected to attenuate this tendency, given that disagreeing others (almost by default) consider and adopt entirely different estimates as one’s own estimate.
Method
The design and procedure was similar to that of Experiment 2. However, before making their second estimate, participants in the self-perspective condition were asked, “What is the most extreme estimate (either extremely high or extremely low) that you would consider as second guess to this question?” In the agreeing- and disagreeing-perspective conditions, participants were asked, “What is the most extreme estimate (either extremely high or extremely low) that your friend would consider as answer to this question?” We recruited 1,500 participants using MTurk. After excluding those who failed the instructional check and those who said they looked up the answers, we obtained a final sample of 1,426 U.S. participants (age:
Results
Correlations
The estimates’ errors were again highly correlated in the self- and agreeing-perspective conditions (mean
Inner-crowd effects
There was an inner-crowd effect overall (i.e., across all three conditions) and in the three conditions separately (see Table 2; for descriptive statistics, see Table S4 in the Supplemental Material).
Benefit of averaging
Participants in the agreeing-perspective condition benefitted slightly more from averaging than did those in the self-perspective condition (
Bracketing
With 20% and 21% of people’s estimates bracketing the questions’ true answers, there was no difference in bracketing rates between the self- and agreeing-perspective conditions (
Extreme-estimate analysis
To test the proposition that taking a disagreeing perspective prompts people to consider more extreme estimates as possible answers to a question, we computed the (absolute) difference score between each participant’s first estimate and the most extreme estimate that they (or their friend) would consider as an answer. As expected, there was no difference between participants in the agreeing- and self-perspective conditions (
As expected, there was no difference between participants in the agreeing- and self-perspective conditions (
Experiment 4
The final experiment identified a situation in which taking a disagreeing perspective backfired. Specifically, this was expected in situations where a question’s true answer lies close to the lower or upper end of a scale (e.g., if the true answer is 2% or 98% on a scale from 0% to 100%) and when a person’s initial estimate is close to this answer. For example, imagine being asked the following question: “What percentage of China’s population identifies as Christian?” The true answer to this question is 5.1%, and if you are like most people, your first estimate probably leaned toward the lower end of the scale (say your first estimate was 10%). Given the position of the question’s true answer and your first estimate, your second estimate is likely (in general) to move away from the answer toward the opposite side of the scale (Juslin et al., 2000), effectively hurting the accuracy of your average estimate. Importantly, such a movement is expected to be especially harmful when second estimates are made from a disagreeing perspective because, given people’s propensity to adopt more extreme estimates from such a perspective (see Experiment 3), these estimates move away from the true answer to a much greater extent (resulting in an average estimate that is far worse than the initial estimate).
Method
We gathered data in two waves. We preregistered our hypotheses and analysis plan for the second wave. Because the procedures in the two waves were identical, we decided to combine them (analyzing the data separately yielded similar results, which can be accessed at https://osf.io/ewpyq/). The procedure of this experiment was similar to that of Experiment 2 in all but two respects. First, we added an additional six questions to make 12 questions in total. Second, we categorized the questions according to where the true answer fell—that is, whether the true answer was in the middle of the scale or the end of the scale (0%–10% or 90%–100%). Participants thus made two estimates about a set of 12 questions, all of which had a true answer that was in the 0% to 100% range. Crucially, half of the questions’ true answers were close to the lower or upper end of the scale, from 0% to 10% and 90% to 100%. For the other half of the questions, true answers were relatively far from the end of the scale (e.g., 58%). Combining the two data-wave collections, we recruited 1,889 participants using MTurk. As in the prior experiments, we excluded those who failed the instructional check and those who said they looked up the answers, leaving a final sample of 1,836 U.S. participants (age:
Results
Correlations
Correlations between the estimates’ errors in the self- and agreeing-perspective conditions were again high (mean
Inner-crowd effects
Taking into account both types of questions (mid-scale and end-scale questions) and all three conditions, we did not find an inner-crowd effect (see Table 3). Importantly, and in line with our proposal, results showed that the perspective-taking instructions had a markedly different impact when the mid-scale and end-scale questions were considered separately. For the mid-scale questions, there was an inner-crowd effect similar to those in the previous experiments. However, when looking at the end-scale questions for the disagreeing-perspective condition, we found that the average of both estimates had a much higher error than the first estimate alone (for descriptive statistics, see Table S6 in the Supplemental Material).
Inner-Crowd Effect for Experiment 4: Comparisons of the Average of Two Estimates With the First and Second Estimate
Benefit of averaging
For the mid-scale questions, we found no difference between the self- and agreeing-perspective conditions (
Bracketing
For the mid-scale questions, with 23% and 21% of the estimates bracketing the questions’ true answers, there was slightly more bracketing in the self-perspective than the agreeing-perspective condition (
Focusing on the end-scale questions, we generally saw lower rates of bracketing. With 13% and 11% of the estimates bracketing the questions’ true answers, there was slightly more bracketing in the self-perspective than the agreeing-perspective condition (
Understanding averaging and bracketing—when is it beneficial?
Prior research suggests that bracketing is a key component in understanding why averaging estimates renders an improvement (Larrick & Soll, 2006; Soll & Larrick, 2009). However, as demonstrated by our results on the end-scale questions, this may not always be the case. Specifically, although we observed higher rates of bracketing in the disagreeing-perspective condition for end-scale questions, averaging nonetheless led to a greater overall disadvantage in this condition. To better understand why this occurred, we took a closer look at the underlying components that determine whether averaging first and second estimates is beneficial or not. We formalize each component in Equation 1.
where
To clarify, the left-hand side of the equation represents the benefit of averaging, and the right-hand side represents the unique components that make up this benefit of averaging. The first component (
The Benefit or Disbenefit of Averaging Per Individual Component for All Three Experimental Conditions, Experiment 4 (End-Scale Questions)
Note: Results are shown separately for the disagreeing-perspective (disagree), self-perspective (self), and agreeing-perspective (agree) conditions. See the text for an explanation of the equations.
What about the observed higher bracketing rate in the disagreeing-perspective condition for end-scale questions? There are two types of brackets (following Equation 1). There are brackets—which we refer to as
General Discussion
Many decisions depend on people’s ability to make accurate estimates of unknown quantities, and a demonstrated way to improve the accuracy of estimates is to aggregate multiple estimates made by the same person. The potential contained in such an intervention is enormous, and a key challenge is to identify strategies that can help improve the accuracy of people’s aggregate estimates (Herzog & Hertwig, 2014a). In this article, we introduced the following strategy: Combine people’s first estimate with their second estimate made from the perspective of someone they often disagree with. Across five experiments, we found evidence that such a strategy produces accurate estimates. These results underscore the importance of perspective taking and disagreement as strategies to improve the accuracy of people’s quantitative estimates.
The presented findings indicate the benefits of disagreement, a component of people’s social interactions that is usually presented as undesirable (Kennedy & Pronin, 2008; Reeder et al., 2005; Sunstein, 2002). What is particularly interesting is that people obtained more accurate estimates by changing their perspective. It remains to be seen whether taking the perspective of any other people—say, experts in a particular field—would lead to similar benefits. This might be an important future research direction, as our findings demonstrate that taking the perspective of other people (e.g., an agreeing perspective) might not always render an increase in accuracy compared with simply making a second guess.
Although the inner crowd offered a gain in accuracy, we also identified a situation in which it backfired, leading to no improvement or even worse performance. We found this to be the case when a question’s answer was close to the scale’s end. Importantly, for participants who employed the disagreeing-perspective strategy, the accuracy of their average estimate was much worse than their first estimate for these types of questions. What is particularly interesting is that the propensity of people to move away from the answer when making second estimates is introduced through a feature of the situation rather than some innate bias (Gaertig & Simmons, 2021; Herzog et al., 2019; Müller-Trede, 2011).
Our research also has several limitations. First, the presented evidence is restricted to populations from the United States and United Kingdom, and future work needs to confirm whether these findings hold true in other parts of the world. Second, although combining initial estimates with second estimates made from a disagreeing perspective is beneficial, the presented research remains mute as to whether people would be willing to aggregate both estimates when given the opportunity (Fraundorf & Benjamin, 2014; Herzog & Hertwig, 2014b; Müller-Trede, 2011). Although prior work indicates that people are more likely to combine their estimates when they actively opposed themselves through dialectical bootstrapping (Herzog & Hertwig, 2014b), it remains to be seen whether this holds true when the opposition comes from someone with whom they often disagree. People typically view others holding opposing views and opinions less favorably (Iyengar & Westwood, 2015; Kennedy & Pronin, 2008; Reeder et al., 2005), potentially undermining their willingness to include the viewpoints of disagreeing others into their own judgments. Future research could address this issue in more detail by testing under what conditions people are willing to combine their estimates with the estimates of disagreeing others to obtain more accurate estimates.
On a final note, whereas previous studies often relied on natural processes such as forgetting or the passage of time to improve the accuracy of inner crowds, the present findings report a strategy that is more convenient and time efficient. Similar to other, more active interventions (Herzog & Hertwig, 2009; Litvinova et al., 2020; Winkler & Clemen, 2004), taking a disagreeing perspective can likewise be used as a potent strategy when people cannot benefit from the wisdom of an actual crowd. Overall, combining one’s first estimate with a second estimate made from the perspective of disagreeing others proves to be a convenient and effective judgment tool.
Supplemental Material
sj-docx-1-pss-10.1177_09567976211061321 – Supplemental material for Taking a Disagreeing Perspective Improves the Accuracy of People’s Quantitative Estimates
Supplemental material, sj-docx-1-pss-10.1177_09567976211061321 for Taking a Disagreeing Perspective Improves the Accuracy of People’s Quantitative Estimates by Philippe P. F. M. Van de Calseyde and Emir Efendić in Psychological Science
Footnotes
Transparency
Both authors contributed equally to the work presented in this article, wrote the manuscript, and approved the final manuscript for submission.
Notes
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
