Evaluating Scientific Research in the Context of Prior Belief: Hindsight ...

12 downloads 847 Views 195KB Size Report
research finding to be obvious involves more than a simple hindsight bias. Keywords: confirmation bias; hindsight bias; evidence evaluation; reasoning; ...
EMPIRICAL REPORT

Evaluating Scientific Research in the Context of Prior Belief: Hindsight Bias or Confirmation Bias? Amy M. Masnick, PhD Hofstra University Corinne Zimmerman, PhD Illinois State University When there is a mismatch between new evidence and prior beliefs, do people reject the conclusions from this evidence because of confirmation bias or do they support them because of hindsight bias? Ninety-four participants expressed a belief about a study’s outcome before reading a research report. When belief was confirmed, the study’s methodology was subsequently rated more positively and findings (whether presented with or without an explanation) were rated as more obvious, important, and interesting than when beliefs were disconfirmed. However, the presence of an explanation for the reported findings affected ratings of obviousness and interestingness but not of methodology. These results indicate that judging a research finding to be obvious involves more than a simple hindsight bias.

Keywords: confirmation bias; hindsight bias; evidence evaluation; reasoning; obviousness

M

ost people have conceptions about how the natural and social world works. These beliefs may be held with varying degrees of conviction. At the same time, people encounter new information every day that may call into question some of these pre-existing beliefs. When new knowledge and prior beliefs conflict, how do individuals evaluate and reconcile the two? Reconciling existing beliefs with new evidence is one of the key components of scientific reasoning (across the lifespan), but it is also necessary for reasoning about the scientific findings individuals are exposed to regularly through a variety of media sources. Therefore, the answer to how individuals reconcile conflicting beliefs and evidence is important not only for understanding the cognitive processes involved in scientific reasoning, but it also has practical implications for understanding scientific literacy and reasoning about science.

Personal, professional, and public policy decisions may be informed by evaluating research findings (Miller, 2004; Shapin, 1992), and these evaluations may include assessing empirical findings, theoretical explanations, or both. However, past research in cognitive science and social cognition leads to different conclusions about how people evaluate new information. On one hand, there is evidence that people are reluctant to consider evidence that disagrees with their beliefs and have a bias to look for information that confirms prior beliefs. (In this article, we consider confirmation bias as a bias toward one’s initial beliefs rather than as a hypothesis-testing strategy [e.g., Klayman & Ha, 1987].) On the other hand, there is evidence that people are susceptible to hindsight bias; that is, they agree with newly learned conclusions and claim to have held these beliefs all along.

Journal of Psychology of Science and Technology, Volume 2, Number 1, 2009 © Springer Publishing Company DOI: 10.1891/1939-7054.2.1.29

29

30

Numerous studies of scientific reasoning have shown that it is difficult to integrate evidence disconfirming prior belief (see Zimmerman, 2000, 2007, for reviews). In particular, when new evidence challenges an existing causal belief, that belief is more resistant to change than when a noncausal belief is challenged (e.g., Kanari & Millar, 2004). Additionally, people are particularly likely to disbelieve new evidence when a theoretical reason or plausible causal mechanism to maintain the current belief is present, and are more likely to take seriously an implausible covariation between two factors when there is a plausible causal explanation (Ahn, Kalish, Medin, & Gelman, 1995; Koslowski, 1996). Similar types of studies on attitude change have also looked at how individuals evaluate evidence in the context of prior belief. Participants in these studies evaluate evidence about controversial topics for which they often have strong prior beliefs (e.g., gun control, capital punishment). For example, Munro, Leary, and Lasane (2004) found that participants whose beliefs were disconfirmed by a fictitious study about the relationship between homosexuality and mental illness were more critical of the study’s methodology than those whose beliefs were confirmed. In general, biased assimilation and attitude polarization are common when evaluating evidence (e.g., Lord, Ross, & Lepper, 1979). That is, research is evaluated more favorably if it supports initial attitudes (biased assimilation), and rather than becoming more moderate in the face of disconfirming evidence, attitudes often become more extreme (attitude polarization; see MacCoun, 1998, for a review). In contrast to research showing that prior beliefs are resistant to change, other research evaluation studies have focused on the perception that, independent of prior belief, social science and educational research findings are considered “obvious” or “common sense” (e.g., Gage, 1991; Yates, 2005). In these studies, participants are shown either true findings or their foils (i.e., false findings). When presented with research findings from the areas of personality (Barnett, 1986), developmental psychology (Barnett, Knust, McMillan, Kaufman, & Sinisi, 1988), and social psychology (Richard, Bond, & Stokes-Zoota, 2001), accuracy in distinguishing true findings from foils ranged from 66%–75% (chance is 50%). Using a similar methodology, Wong (1995) found that

MASNICK AND ZIMMERMAN

when presented with both a true finding and a foil about educational research, participants were equally likely to select either version as the actual finding. Moreover, ratings of obviousness were equivalent for findings and nonfindings. Results presented with an explanation for the finding, however, were rated as significantly more obvious than those without an explanation, for both actual findings and foils. These findings are consistent with research on hindsight bias (e.g., Guilbault, Bryant, Brockway, & Posavac, 2004). That is, when individuals have knowledge of an outcome, they tend to overestimate their knowledge in advance of knowing that outcome (i.e., the “I knew it all along” phenomenon). Descriptions of hindsight typically include two components: (a) overestimation of one’s ability to predict an outcome after the fact, and (b) the belief that one is not in fact influenced by the knowledge of the outcome (Hawkins & Hastie, 1990). More recent discussions, however, distinguish several subtypes of hindsight bias (Blank, Nestler, von Collani, & Fischer, 2008). Here we concentrate on a hindsight bias, focusing on one’s ability to assess the likelihood of an outcome after knowing the outcome, as measured by ratings of how obvious research findings are after one has read them. Wong’s findings suggest that this type of hindsight bias may be even stronger in the presence of a plausible explanation for the reported finding. Most previous research examining the obviousness of findings has not specifically examined the effects of prior expectations about the findings. Past research on scientific reasoning and social cognition typically uses topics for which individuals have prior expectations, due either to ingrained or politicized beliefs or to fairly robust science misconceptions. The current study explored whether findings from newly reported research would be subject to a “knew-itall-along” hindsight bias when a prior belief is expressed. We also explored whether the presence of an explanation for a finding influences evaluations of obviousness of a study after having one’s belief either confirmed or disconfirmed. We created a task that includes characteristics used by researchers studying reasoning, attitude change, and perceptions of the obviousness, importance, and interestingness of research findings. We explored the process of evaluating educational research findings because individuals have been shown

EVALUATING RESEARCH

to be less accurate in discriminating true findings and foils in this domain (Wong, 1995). If evaluating research after expressing a belief about the outcome leads to the confirmation bias found when evaluating evidence that challenges or confirms political attitudes (e.g., Lord et al., 1979; Munro et al., 2004), then evaluations should differ as a function of prior belief. Specifically, evaluations about the methodology used to conduct the study should be more positive after learning that the result confirms prior belief (and negative when results disconfirm prior belief ). Overall evaluations of the research would be more positive by those whose beliefs are confirmed by a study’s findings because participants would be allowing prior beliefs to bias interpretation of new information. Moreover, the presence of an explanation may lead to different cognitive processing based on prior belief. Explanations for confirmed beliefs may result in positive evaluations, whereas explanations for disconfirmed beliefs may result in negative evaluations, as predicted by confirmation bias. Alternatively, if simply reading about a research result can make a finding seem obvious in hindsight, then participants’ evaluations about study quality and the obviousness of results should be similar regardless of prior expectations about the findings (e.g., Richard et al., 2001; Wong, 1995). In addition, the presence of an explanation for the findings may affect evaluations even when a belief is challenged. Investigations of scientific reasoning have shown that the ability to reconcile beliefs and evidence is often mediated by the presence of (or ability to generate) plausible causal mechanisms for the pattern of evidence/data (e.g., Ahn et al., 1995; Koslowski, 1996; Wong, 1995). When reasoning about scientific research, it is reasonable to assume that individuals are also considering theoretical explanations or causal mechanisms (e.g., Chinn & Brewer, 2001). That is, hindsight bias may be more evident when explanations are present, such that an explained finding may seem even more obvious regardless of prior belief. In sum, the present study explored two factors that could influence the evaluation of a research study: (a) having one’s belief about an educational research question confirmed or disconfirmed by evidence, and (b) the presence or absence of a causal explanation for the finding.

31

METHOD

Participants Ninety-four undergraduate students (82 women, 12 men) participated for course credit. The average age was 22.1 (SD = 4.1). They had an average of 3.8 years of college-level instruction (SD = .74), with an average of 1.8 (SD = 1.1) methods/statistics courses. The racial composition of the sample was 73% White, 12% African American, and 7% each Hispanic and Asian. Most participants indicated they were majoring in psychology (37%), education (30%), or psychology and education (13%).

Materials and Procedure A questionnaire was administered in small groups. The first page briefly described an issue in early science education. Participants were told that researchers have been studying the relative efficacy of direct instruction (being taught information explicitly) as compared to discovery learning (being given a goal and some materials and told to explore on one’s own). Participants were asked to indicate which method they believed was the most effective (i.e., a forcedchoice question). The procedure of requiring a forced choice between the two sides of this issue allows participants to be later classified as having their belief confirmed or disconfirmed by the result of a study (Munro et al., 2004).1 All participants then read the same one-page description of an experimental study, including a brief introduction (background and research questions) and method section. Participants then evaluated the appropriateness of the methods, design, participants, and measures on a 7-point Likert scale (from strongly disagree to strongly agree that each was appropriate). After evaluating the study’s methodology, participants were randomly assigned to read one of four versions reporting the study’s findings. In two versions, direct instruction was described as the more effective intervention. One version asserted this finding without an explanation. The other version included the explanation that direct instruction was more effective because learning is facilitated by the teacher’s organized presentation and focus on relevant concepts and procedures. In the other two versions, discovery learning was described as the more effective

32

intervention. One version included the explanation that discovery learning is more effective because of students’ active involvement in the learning process, which makes the information more meaningful and likely to be remembered. After reading the findings (with or without an explanation) participants rated characteristics of the conclusions of the study using a 7-point Likert scale. Participants rated how strongly they disagreed or agreed with statements about the conclusions with respect to whether they were (a) obvious, (b) important, and (c) interesting. Participants then completed a second evaluation of the methods, design, participants, and measures. Participants were debriefed at the end of the study about the study’s purpose.

RESULTS Based on the initial forced-choice question, one-third of the students believed that direct instruction was more effective and two-thirds believed that discovery learning was more effective. Because of random assignment to which results were seen, approximately equivalent groups had initial belief confirmed (n = 46) and disconfirmed (n = 48). To assess the equivalence of versions, ratings were initially analyzed disregarding prior belief. Ratings of the obviousness of the reported findings were similar for those who read the discovery-learning conclusion (M = 4.9, SD = 1.4; n = 46) and the direct-instruction conclusion (M = 4.6, SD = 1.4; n = 48). In addition, no differences were found for the presence of an explanation or in initial or postfinding evaluations of methodology based on which version participants read (all Fs ≈ 1, ps > 0.10). Thus, the content of the conclusion participants read (discovery learning or direct instruction) did not influence participants’ reasoning about the study. We therefore focused on whether participants’ beliefs were confirmed or disconfirmed, disregarding prior belief. The seven ratings participants gave (method, design, participants, measures, obviousness, importance, and interestingness) were entered into a multivariate 2 (belief: confirmed vs. disconfirmed) × 2 (explanation: present vs. absent) analysis of variance (ANOVA).

MASNICK AND ZIMMERMAN

Judgments About Methodology No differences in initial ratings of appropriateness of the method, design, participants, or measures were found as a function of either belief or explanation (Fs ≈ 1, ps > 0.10) so analyses were conducted on change scores (second minus first assessment). Ratings of the appropriateness of the methods increased for those whose belief was confirmed (M = .50; SD = 1.1) but decreased for those whose belief was disconfirmed (M = −.56; SD = 1.6), F(1, 90) = 13.3, p < .001, partial η2 = .13 (Figure 1). No differences were found for the presence of an explanation, and there was no interaction between explanation and belief confirmation (Fs ≈ 1, ps > 0.10). Participants’ ratings of the design used by researchers increased when beliefs were confirmed (M = .74, SD = 1.1) but decreased (M = −.35, SD = 1.1) when beliefs were disconfirmed, F(1,90) = 13.9, p < .001, partial η2 = .13. No differences were found for the presence of an explanation, and there was no interaction (Fs ≈ 1, ps > 0.10). The pattern was similar to that illustrated in Figure 1. There were small positive changes in ratings about the appropriateness of the participants and measures but no differences among groups (all Fs ≈ 1, ps > 0.10).

Judgments About the Findings and Conclusions Participants’ ratings of the obviousness of the conclusion were greater when initial belief was confirmed (M = 5.1, SD = 1.4) than when initial belief was disconfirmed (M = 4.4, SD = 1.4), F(1,90) = 5.78, p = .018, partial η2 = .06. Although there was no main effect of explanation, there was an interaction between belief confirmation and explanation, F(1,90) = 4.26, p = .042, partial η2 = .05 (Figure 2). When there was no explanation for the finding, ratings were similar for those whose beliefs were confirmed or disconfirmed. Ratings of obviousness were much higher for those whose belief was confirmed compared to those whose belief was disconfirmed, F(1,90) = 9.70, p = .002, partial η2 = .10. That is, when findings were presented with an explanation, there was a large difference in perceived obviousness. Participants’ ratings of the importance of the research were also greater when beliefs were confirmed

EVALUATING RESEARCH

FIGURE 1. Mean change in appropriateness of methods rating (before and after reading the results of the study) as a function of whether one’s belief was confirmed or disconfirmed by the findings, and whether or not an explanation for the findings was presented.

FIGURE 2. Mean ratings of the obviousness of the findings/conclusions as a function of whether one’s belief was confirmed or disconfirmed by the findings, and whether or not an explanation for the findings was presented.

33

34

(M = 6.1, SD = 1.3) than when beliefs were disconfirmed (M = 5.4, SD = 1.4), F(1,90) = 7.94, p = .006, partial η2 = .08. The presence of an explanation resulted in higher ratings of the findings as important, F(1,90) = 4.06, p = .047, partial η2 = .04, but the presence of an explanation did not interact with belief. The same pattern was found for ratings of interestingness, with higher ratings from those whose belief was confirmed, F(1,90) = 12.09, p = .001, partial η2 = .19, and for those who read an explanation for the finding, F(1,90) = 7.80, p = .006, partial η2 = .08, with no interaction.

DISCUSSION The results demonstrate that when evaluating reports of social science research, the tendency toward hindsight bias is trumped by beliefs expressed before evaluating the findings. Evidence that contradicts initial belief is assessed differently from evidence that supports it. Consistent with predictions from research on attitude change and scientific reasoning, there was a strong effect of confirmation on ratings of methodology, design, and judgments of several qualities of the research, such as its obviousness. This finding indicates that when stating an expectation about the finding prior to reading the outcome, participants are more likely to indicate a confirmation bias than a hindsight bias. Both the methods and the findings were rated more positively after a belief was confirmed and more negatively after a belief was disconfirmed. Recall that both findings were judged to be equivalent, independent of prior belief (i.e., the discovery learning and direct instruction versions were rated equally obvious), and methodology ratings were equivalent prior to reading the findings. Thus, it was not the specific content that affected evaluation but rather the confirmation or disconfirmation of prior beliefs. Hindsight bias was not in evidence: people remained tethered to their initial beliefs more than the newly presented conclusions, as indicated by the effect of confirmation on the change in ratings of the appropriateness of the methodology. This finding suggests that working to change someone’s view simply based on content may be unsuccessful without considering initial views on the topic. Evaluating the methodological soundness

MASNICK AND ZIMMERMAN

of a study may also be influenced by pre-existing beliefs. One difference between the current study and many other studies of hindsight bias (e.g., Hawkins & Hastie, 1990) is that research findings can be assessed both on the nature of the conclusions and on methods by which the conclusions were drawn (unlike, for example, the assessment of a historical event). This distinction may be one reason we found little evidence of hindsight bias. Assessments of the appropriateness of participants (i.e., generalizabililty to other populations) and measures (i.e., construct validity) were not influenced by prior belief. Participants may have been able to evaluate these features of the study more objectively, or they may not have believed these characteristics were as important in evaluating a study’s validity. Having one’s belief confirmed also resulted in ratings of the reported findings as more obvious, interesting, and important. In contrast to previous research showing a type of hindsight bias in the evaluation of research findings, when prior belief was taken into consideration, ratings of obviousness varied as a function of that prior belief. If hindsight bias were at work, we would expect equivalent (and positive) obviousness ratings regardless of whether a prior belief was confirmed or challenged. Although lower ratings of obviousness do not necessarily indicate participants are discounting or discrediting the study, the differences in ratings are still changing based on confirmation. Two possible explanations seem plausible for this effect. First, although we presume participants’ initial opinions on the topic (i.e., educational interventions) were formed quickly and on the spot, most participants were psychology or education majors (80% of the sample), and so they may have had occasion to consider such topics prior to participation. If so, then challenging or confirming those beliefs would be expected to result in behavior similar to that observed when other strongly held attitudes are challenged (e.g., Munro et al., 2004). Second, it is also possible that being forced to state an opinion early, even if chosen somewhat arbitrarily, leads to a commitment to the position and consistency within the task. To explore this possibility further, it will be necessary to replicate this finding with a group that does not express an opinion prior to reading the methods and findings.

EVALUATING RESEARCH

The presence of an explanation for a finding led to higher ratings of how important and interesting that finding was. These results are consistent with past work indicating that explanations are a critical part of reasoning about science (e.g., Ahn et al., 1995; Koslowski, 1996; Wong, 1995). Individuals are more likely to believe an empirical finding if there is a theory or explanation for that finding. Thus, it is unsurprising that the presence of explanatory information would increase perceptions of how important and interesting a topic is. Yet the explanation effect was not completely straightforward. An unexpected finding was the interaction of explanation and belief confirmation on ratings of obviousness. Wong (1995) found that the presence of an explanation increased feelings of obviousness for both true findings and foils. We found similar results, but only under certain conditions. In the absence of an explanation, findings were rated as equally obvious regardless of prior belief. However, when an explanation was provided, a confirmed belief was rated as much more obvious than a disconfirmed belief. Thus, the absence of an explanation actually eliminated the effect of prior belief. It is possible that participants generate their own explanations for the results and therefore rate the findings as equally obvious. In addition, with disconfirming evidence, participants may become more critical and discredit both the finding and the explanation for that finding to maintain their belief, leading to a lower rating of obviousness (Klaczynski & Narasimham, 1998). If a belief has just been disconfirmed, then the accompanying explanation is presumably one that has either not been considered earlier or has been considered and dismissed. In summary, the methodology developed in the current research can be productive for examining the process of evidence evaluation and for making links among several research literatures. The findings add to the literature on scientific reasoning by showing that the presence of an explanation does not always affect evaluation of scientific reports in a uniform way. Consistent with previous research on confirmation bias, we found that prior belief exerts a powerful influence, but our results suggest this effect is not diminished by the presence of a plausible explanation. The current findings also contribute to the hindsight bias literature by demonstrating that although partici-

35

pants often consider research findings to be obvious, there are some boundary conditions for such judgments (e.g., belief confirmation or disconfirmation in conjunction with the presence of an explanation). The judgment of a research finding to be obvious clearly involves more than a simple hindsight bias. Additional research is needed to further explore the complexities of evaluating the kinds of research evidence commonly reported in the media that may be used to inform personal and policy decisions.

NOTE 1. Munro et al. (2004) allowed individuals to indicate ambivalence by selecting a midpoint on a Likert scale when expressing views about homosexuality. This procedure resulted in removing only a small minority of individuals from their analyses (4 of 92).

REFERENCES Ahn, W., Kalish, C. W., Medin, D. L., & Gelman, S. A. (1995). The role of covariation versus mechanism information in causal attribution. Cognition, 54, 299–352. Barnett, M. A. (1986). Commonsense and research findings in personality. Teaching of Psychology, 13, 62–64. Barnett, M. A., Knust, J., McMillan, T., Kaufman, J., & Sinisi, C. (1988). Research findings in developmental psychology: Common sense revisited. Teaching of Psychology, 15, 195–197. Blank, H., Nestler, S., von Collani, G., & Fischer, V. (2008). How many hindsight biases are there? Cognition, 106, 1408–1440. Gage, N. L. (1991). The obviousness of social and educational research results. Educational Researcher, 20, 10–16. Guilbault, R. L., Bryant, F. B., Brockway, J. H., & Posavac, E. J. (2004). A meta-analysis of research on hindsight bias. Basic and Applied Social Psychology, 26, 103–117. Hawkins, S. A., & Hastie, R. (1990). Hindsight: Biased judgments of past events after the outcomes are known. Psychological Bulletin, 107, 311–332. Kanari, Z., & Millar, R. (2004). Reasoning from data: How students collect and interpret data in science investigations. Journal of Research in Science Teaching, 41, 748–769. Klaczynski, P. A., & Narasimham, G. (1998). Development of scientific reasoning biases: Cognitive versus egoprotective explanations. Developmental Psychology, 34, 175–187. Klayman, J., & Ha, Y. (1987). Confirmation, disconfirmation, and information in hypothesis-testing. Psychological Review, 94, 211–228.

36 Koslowski, B. (1996). Theory and evidence: The development of scientific reasoning. Cambridge, MA: MIT Press. Lord, C. G., Ross, L., & Lepper, M. R. (1979). Biased assimilation and attitude polarization: The effects of prior theories on subsequently considered evidence. Journal of Personality and Social Psychology, 37, 2098–2109. MacCoun, R. J. (1998). Biases in the interpretation and use of research results. Annual Review of Psychology, 49, 259–287. Miller, J. D. (2004). Public understanding of, and attitudes toward, scientific research: What we know and what we need to know. Public Understanding of Science, 13, 273–294. Munro, G. D., Leary, S. P., & Lasane, T. P. (2004). Between a rock and a hard place: Biased assimilation of scientific information in the face of commitment. North American Journal of Psychology, 6, 431–444. Richard, F. D., Bond, C. F., & Stokes-Zoota, J. J. (2001). “That’s completely obvious . . . and important”: Lay judgments of social psychological findings. Personality and Social Psychology Bulletin, 27, 497–505.

MASNICK AND ZIMMERMAN Shapin, S. (1992). Why the public ought to understand science-in-the-making. Public Understanding of Science, 1, 21–30. Wong, L. Y. (1995). Research on teaching: Process-product research findings and the feeling of obviousness. Journal of Educational Psychology, 87, 504–511. Yates, G. C. R. (2005). “How obvious”: Personal reflections on the database of educational psychology and effective teaching research. Educational Psychology, 25, 681–700. Zimmerman, C. (2000). The development of scientific reasoning skills. Developmental Review, 20, 99–149. Zimmerman, C. (2007). The development of scientific thinking skills in elementary and middle school. Developmental Review, 27, 172–223.

Correspondence regarding this article should be directed to Amy M. Masnick, Hofstra University, Hauser Hall, Psychology Department, Hofstra University, Hempstead, NY, 11549. E-mail: [email protected]