Abstract
In this paper we investigate ignorance in relation to our capacity to justify our beliefs. To achieve this aim we specifically address scientific misconceptions, i.e. beliefs that are considered to be false in light of accepted scientific knowledge. The hypothesis we put forward is that misconceptions are not isolated false beliefs, but rather form part of a system of inferences—an explanation—which does not match current scientific theory. We further argue that, because misconceptions are embedded in a system, they cannot be rectified simply by replacing false beliefs with true ones. To address our misconceptions, we must rather act on the system of beliefs that supports them. In the first step of our analysis, we distinguish between misconceptions that are easy to dispel because they represent simple errors that occur against the background of a correct explanatory apparatus and misconceptions that are, on the contrary, very difficult to dispel because they are the product of pseudo explanations. We show that, in the latter case, misconceptions constitute an integral part of an incorrect explanation and the reasons that support such misconceptions are deeply misleading. In the second step, we discuss various approaches that have been adopted to address the problem of misconceptions. Challenging the notion that directly addressing and criticizing specific misconceptions is an effective approach, we propose that critical thinking is the most fruitful means to deal with misconceptions. We define the core competences and knowledge relevant for the practice of critical thinking and discuss how they help us avoid misconceptions that arise from accepting beliefs that form part of a mistaken explanation.
Similar content being viewed by others
Notes
In the literature, the approach to ignorance that considers the possession of true justified beliefs as a necessary condition for having knowledge and, by contrast, the absence of one of these conditions as sufficient to establish ignorance is called the Standard View of Ignorance (Le Morvan 2011, 2012, 2013; Le Morvan and Peels 2016). Recently, epistemological research has developed a new approach called the New View of Ignorance in which the role assigned to justification and to the capacity to explicitly offer reasons in support of our beliefs has been weakened and in which a person is considered to be ignorant primarily in the case in which s/he holds false beliefs. (On the so-called New View see Peels 2010, 2011, 2012; Le Morvan and Peels 2016). On the one hand, the New View denies—contrary to what we suggest—that justification plays a central function in determining whether we are or are not ignorant about some topic. However, on the other hand, it takes a step towards our approach: it not only places stronger emphasis on the cognizing powers of the subject than the "standard view" but also considers the possibility of distinguishing between various degrees of ignorance. In fact, by discussing the impossibility of providing a complete justification for a belief, it makes it possible to consider ignorance as a continuum rather than in a categorical manner, distinguishing between various degrees in which a person can be said to be ignorant as for the reasons s/he has to hold his/her beliefs (for a conception of ignorance as an epistemic status that comes in degrees and ignorance as the incapacity to adequately and completely answer questions concerning our beliefs cf. Nottelmann 2016). Of course, we cannot expect that people are able to provide a complete justification for their beliefs and even the issue of what an acceptable justification should look like is controversial. However, especially when considering beliefs concerning (relatively simple) phenomena that have a widely accepted , we can take a practical stance and assume that scientific theories which are commonly accepted by the scientific community at a given time provide us a measure for which beliefs should be considered to be true or false at that time and what a justification of them would ideally look like. On this basis, we can also assess (or approximate) the distance between scientific knowledge and individual beliefs as well as the distance between the way scientific theories justify specific pieces of knowledge and the way in which people justify their beliefs about the same phenomena.
In fact, it is not impossible for an individual to hold a belief that is inconsistent with other beliefs s/he holds, and yet it would be irrational for him/her to do so. As Davidson argues: “Strictly speaking, then, the irrationality consists not in any particular belief but in inconsistency within a set of beliefs.” (Davidson 1985/2004, p. 192) According to a widely shared view in psychology developed by Leon Festinger (1957), inconsistencies cannot be psychologically accepted by the subject who will make every possible effort to rationalize and thus resolve them. In the same vein, Davidson (1982/2004, 1986/2004) points out how the inconsistencies we are sometimes victim of can be explained only by postulating a kind of compartmentalization of the mind. However, generally people seek internal congruency among their beliefs and coherence plays a pivotal role in the way we interpret human thinking (on this cf. also Thagard 2000).
In epistemology, justification has been viewed in various ways. Justification might be conceived as being linear: in this case an individual belief is proven to be true by a set of other beliefs and those other beliefs are proven to be true by another set and so on, until we reach some beliefs that are based on experience and are therefore considered—if not indisputable—at least well-grounded: as BonJour (1985, p. 26) formulates it, “sufficient by itself to satisfy the adequate justification condition for knowledge”. Alternatively, justification can be viewed holistically as an interferential network of beliefs that are interconnected within a system and providing mutual support but are supported by experience only altogether as a whole. In this case, what justifies a belief is primarily its coherence with the system (this form of holism is commonly discussed in relation to Quine 1951; for a discussion of different justification models cf. Elgin 2005; van Cleve 2005). While strong forms of holism would consider a belief to be justified only if it is coherent with the whole system of beliefs that includes it, more moderate forms of holism support the view that justification depends on some chunks of this system. Typically, they also embrace some weak form of foundationalism in which some beliefs are considered more basic than others because they are closer to experience, i.e. observational, and thus serve as a foundation for others. (On moderate vs. strong holism from the point of view of Quine’s philosophy and on Quine’s later arguments in favor of a moderate form of holism cf. De Rosa and Lepore 2004). To numerous epistemologists, the idea that we can always apply linear models of justification that lead us to some fundamental beliefs appeared to be implausible in the light of the complexity of our system of knowledge. For this reason, they argued for a holist picture of knowledge in which beliefs are connected to each other within an inferential network and mutually sustain each other (cf. van Quine and Ullian 1970; Bonjour 1985; Harman 1993; Thagard 2007). At the same time, the view that all the beliefs of a complex beliefs system are also involved in the justification of each appeared to be too extreme as well as problematic from an epistemological point of view. In fact, this implies that, when even one single belief in the system is changed, all others must be modified accordingly. This excessive interdependence of beliefs makes the system as a whole too instable (cf. Fodor and Lepore 1992, chap. 2). For this reason, many epistemologists have considered a moderate form of holism as the most plausible option. And yet, independently of which view of the structure of knowledge we favor and thus of which is the specific model of justification we prefer, at least some principles of inferential justification can be considered to be shared by all these models. Indeed, independently of whether we think that our beliefs are structured “like a building that rests upon a foundation” or "like a web where the strength of any given area depends on the strength of the surrounding areas" (Steup 2018), we always presuppose that beliefs form a congruent structure and that their relationships are somehow explanatory. We will say something more on this last factor below, but—since we will consider scientific theory as a benchmark to assess misconceptions—we will mainly just assume that the inferential relationships presupposed by scientific theories are explanatory, i.e. that they are form part of an appropriate explanation.
Even though this remains implicit in his paper, Reichenbach’s analysis is inspired by a specific model of explanations, i.e. by the Hempel and Oppenheim’s (1948) Deductive-Nomological Model. However, his description is general enough to also be compatible with other positions on what a consists of. The Deductive-Nomological Model of explanation does not explicitly rely on the notion of causation. But many advocates of this model argue that it still captures the causal component of explanations since “all causal claims imply the existence of some corresponding regularity (a “law”) linking cause to effect” (Woodward 2017, cf. this article also for a brief discussion of the main modes of explanation that are currently under debate).
Reichenbach suggests that—when people do not have the means to develop an actual explanation—they try to account for phenomena by analogy with something else they understand better: since human experience is something everybody has firsthand knowledge of, people usually resort to analogies with human experience. Reichenbach intuition on this is confirmed by a number of contemporary, empirical studies showing that people with a poor understanding of the physical word have a strong tendency to anthropomorphize. They tend to explain physical phenomena using the same principles they would use to explain the behavior of human agents and thus they project human-like characteristics onto non-human things (Epley et al. 2007; Willard and Norenzayan 2013; Lindeman and Svedholm-Häkkinen 2016). And yet, the opposite also occurs, even if more rarely: people who exhibit a poor knowledge of the human mind and of social dynamics, but have a better comprehension of mechanisms and physical systems tend to interpret human phenomena according to non-human but better known mechanical principles (Lindeman and Svedholm-Häkkinen 2016).
References
Aarnio, K., & Lindeman, M. (2005). Paranormal beliefs, education, and thinking styles. Personality and Individual Differences, 39(7), 1227–1236. https://doi.org/10.1016/j.paid.2005.04.009.
Bailin, S. (2002). Critical thinking and science education. Science and Education, 11(4), 361–375. https://doi.org/10.1023/A:1016042608621.
Bailin, S., Case, R., Coombs, J. R., & Daniels, L. B. (1999a). Conceptualizing critical thinking. Journal of Curriculum Studies, 31(3), 285–302. https://doi.org/10.1080/002202799183133.
Bailin, S., Case, R., Coombs, J. R., & Daniels, L. B. (1999b). Common misconceptions of critical thinking. Journal of Curriculum Studies, 31(3), 269–283. https://doi.org/10.1080/002202799183124.
Bannink, F. P. (2007). Solution-focused brief therapy. Journal of Contemporary Psychotherapy, 37(2), 87–94. https://doi.org/10.1007/s10879-006-9040-y.
Bensley, D. A., & Lilienfeld, S. O. (2015). What is a psychological misconception? Moving toward an empirical answer. Teaching of Psychology, 42(4), 282–292. https://doi.org/10.1177/0098628315603059.
Bensley, D. A., Lilienfeld, S. O., & Powell, L. A. (2014). A new measure of psychological misconceptions: Relations with academic background, critical thinking, and acceptance of paranormal and pseudoscientific claims. Learning and Individual Differences, 36, 9–18. https://doi.org/10.1016/j.lindif.2014.07.009.
BonJour, L. (1985). The structure of empirical knowledge. Cambridge: Harvard University Press.
Brosnan, M., Ashwin, C., & Lewton, M. (2017). Brief report: Intuitive and reflective reasoning in autism spectrum disorder. Journal of Autism and Developmental Disorders, 47(8), 2595–2601. https://doi.org/10.1007/s10803-017-3131-3.
Brosnan, M., Lewton, M., & Ashwin, K. (2016). Reasoning on the autism spectrum: A dual process theory account. Journal of Autism and Developmental Disorders, 46, 2115–2125.
Browne, N. M., & Keeley, S. M. (2007). Asking the right questions. Upper Saddle River: Prentice-Hall.
Burke, B. L., Sears, S. R., Kraus, S., & Roberts-Cady, S. (2014). Critical analysis: A comparison of critical thinking changes in psychology and philosophy classes. Teaching of Psychology, 41(1), 28–36. https://doi.org/10.1177/0098628313514175.
Conception. (2011). In Merriam-Webster.com. Retrieved August 17, 2019, from https://www.merriam-webster.com/dictionary/conception.
Cottrell, S. (2005). Critical thinking skills. Developing effective analysis and argument. New York: Palgrave.
Davidson, D. (1982/2004). Paradoxes of irrationality. In D. Davidson (Ed.), Problems of irrationality (pp. 169–187). Oxford/New York: Oxford University Press.
Davidson, D. (1985/2004). Incoherence and irrationality. In D. Davidson (Ed.), Problems of irrationality (pp. 189–198). Oxford/New York: Oxford University Press.
Davidson, D. (1986/2004). Deception and division. In D. Davidson (Ed.), Problems of irrationality (pp. 199–212). Oxford/New York: Oxford University Press.
De Martino, B., Harrison, N. A., Knafo, S., Bird, G., & Dolan, R. J. (2008). Explaining enhanced logical consistency during decision making in autism. Journal of Neuroscience, 28(42), 10746–10750. https://doi.org/10.1523/JNEUROSCI.2895-08.2008.
De Rosa, R., & Lepore, E. (2004). Quine’s meaning holisms. In R. F. Gibson (Ed.), The Cambridge companion to Quine (pp. 65–90). Cambridge: Cambridge University Press.
Dewey, J. (1910). How we think. Boston/New York/Chicago: D.C. Heath.
Dewey, J. (1933). How we think: A restatement of the relation of reflective thinking to the educative process. Lexington: D.C. Heath.
di Sessa, A. A. (2006). A history of conceptual change research: Threads and fault lines. In R. K. Sawyer (Ed.), The Cambridge handbook of the learning sciences (pp. 88–108). Cambridge: Cambridge University Press.
Elgin, C. (2005). Non-foundationalist epistemology: Holism, coherence, and tenability. In M. Steup & E. Sosa (Eds.), Contemporary debates in epistemology (pp. 156–167). New York/London: Blackwell.
Ennis, R. H. (1985). A logical basis for measuring critical thinking skills. Educational Leadership, 43(2), 44–48.
Epley, N., Waytz, A., & Cacioppo, J. T. (2007). On seeing human: A three-factor theory of anthropomorphism. Psychological Review, 114(4), 864–886.
Facione, P. A. (1990). Critical thinking: A statement of expert consensus for purposes of educational assessment and instruction. Research Findings and Recommendations Prepared for the Committee on Pre-College Philosophy of the American Philosophical Association, ERIC Document ED315423.
Festinger, L. (1957). A theory of cognitive dissonance. Stanford: Stanford University Press.
Firestein, S. (2012). Ignorance: How it drives science. Oxford: Oxford University Press.
Fischer, K. M. (1983). Amino acids and translations: A misconception in biology. In H. Helm & J. D. Nowak (Eds.), Proceedings of the international seminar on misconceptions in science and mathematics (pp. 407–419). Ithaca: Cornell University Press.
Fodor, J. A., & Lepore, E. (1992). Holism: A shopper’s guide. Oxford: Blackwell.
Furnham, A., & Hughes, D. J. (2014). Myths and misconceptions in popular psychology: Comparing psychology students and the general public. Teaching of Psychology, 41(3), 256–261.
Gardner, R., & Brown, D. L. (2013). A test of contemporary misconceptions in psychology. Learning and Individual Differences, 24, 211–215. https://doi.org/10.1016/j.lindif.2012.12.008.
Garnett, P. J., & Treagust, D. F. (1992a). Conceptual difficulties experienced by senior high school students of electrochemistry: Electric circuits and oxidation-reduction equations. Journal of Research in Science and Teaching, 29(2), 121–142.
Garnett, P. J., & Treagust, D. F. (1992b). Conceptual difficulties experienced by senior high school students of electrochemistry: Electrochemical (galvanic) and electrolytic cells. Journal of Research in Science and Teaching, 29(10), 1079–1099.
Gil, F. (2000). La conviction. Paris: Flammarion.
Gilbert, J. K., & Watts, D. M. (2008). Concepts, misconceptions and alternative conceptions: Changing perspective in science education. Studies in Science Education, 10(1), 61–98.
Gingerich, W. J., & Eisengart, S. (2004). Solution-focused brief therapy: A review of the outcome research. Family Process, 39(4), 477–498. https://doi.org/10.1111/j.1545-5300.2000.39408.x.
Goris, T. & Dyrenfurth, M. (2010). Students’ misconception in science, technology, and engineering. In ASEE Illinois/Indiana section conference. Retrieved September 10, 2019 from http://ilin.asee.org/Conference2012/Papers/Goris.pdf.
Govier, T. (1989). Critical thinking as argument analysis? Argumentation, 3(2), 115–126. https://doi.org/10.1007/BF00128143.
Govier, T. (2010). A practical study of argument. Cengage: Wadsworth.
Gregory, T. R. (2009). Understanding natural selection: Essential concepts and common misconceptions. Evolution: Education and Outreach, 2(2), 156–175.
Guzzetti, B. J. (2000). Learning counter-intuitive science concepts: What have we learned from over a decade of research? Reading and Writing Quarterly, 16(2), 89–98.
Halpern, D. F. (2014). Thought and knowledge. An introduction to critical thinking. New York: Psychology Press.
Hare, W. (1979). Open-mindedness and education. Kingston: McGill-Queen’s University Press.
Hare, W. (2001). Bertrand Russell and the ideal of critical receptiveness. Skeptical Inquirer, 25(3), 40–44.
Harman, G. (1993). Meaning holism defended. In J. A. Fodor & E. Lepore (Eds.), Holism: A consumers update (pp. 163–171). Amsterdam: Rodopi.
Hempel, C., & Oppenheim, P. (1948). Studies in the logic of explanation. Philosophy of Science, 15, 135–175.
Herron, J. D. (1990). Research in chemical education: Results and directions. In M. Gardner, J. G. Greeno, F. Reif, A. H. Schoenfaled, A. A. di Sessa, & E. Stage (Eds.), Toward a scientific practice of science education (pp. 31–54). Hillsdale: Erlbaum.
Hitchcock, D. (2017). On reasoning and argument: Essays in informal logic and on critical thinking. Dordrecht: Springer. https://doi.org/10.1007/978-3-319-53562-3_30.
Hitchcock, D. (2018a). Critical thinking. In E. N. Zalta (Ed.), The Stanford encyclopedia of philosophy. Retrieved September 10, 2019, from https://plato.stanford.edu/archives/fall2018/entries/critical-thinking/.
Hitchcock, D. (2018b). Assessment. Supplement to critical thinking. In E. N. Zalta (Ed.), The Stanford encyclopedia of philosophy. Retrieved September 10, 2019, from https://plato.stanford.edu/entries/critical-thinking/assessment.html.
Irwin, H. J. (2009). The psychology of paranormal belief. A researcher’s handbook. Hatfield: University of Hertfordshire Press.
Kahane, H. (1989). The proper subject matter for critical thinking courses. Argumentation, 3(2), 141–147.
Kahneman, D. (2011). Thinking, fast and slow. New York: Farrar, Strauss & Giroux.
Kendeou, P., & van den Broek, P. (2005). The effects of readers’ misconceptions on comprehension of scientific text. Journal of Educational Psychology, 97(2), 235–245. https://doi.org/10.1037/0022-0663.97.2.235.
Kikas, E. (2004). Teachers’ conceptions and misconceptions concerning three natural phenomena. Journal of Research in Science Education, 41(5), 432–448.
Kim, J. S. (2008). Examining the effectiveness of solution-focused brief therapy: A meta-analysis. Research on Social Work Practice, 18(2), 49–64. https://doi.org/10.1177/1049731507307807.
Kirby, G. (2018). Wacky and wonderful misconceptions about our universe. Berlin/Heidelberg: Springer.
Kowalski, P., & Taylor, A. (2009). The effect of refuting misconceptions in the introductory psychology class. Teaching of Psychology, 36(3), 153–159.
Kuczmann, I. (2017). The structure of knowledge and students’ misconceptions in physics. AIP Conference Proceedings, 1916, 050001. https://doi.org/10.1063/1.5017454.
Le Morvan, P. (2011). On ignorance: A reply to Peels. Philosophia, 39(2), 335–344.
Le Morvan, P. (2012). On ignorance: A vindication of the standard view. Philosophia, 40(2), 379–393.
Le Morvan, P. (2013). Why the standard view of ignorance prevails. Philosophia, 41(1), 239–256.
Le Morvan, P., & Peels, R. (2016). The nature of ignorance: Two views. In R. Peels & M. Blaauw (Eds.), The epistemic dimensions of ignorance (pp. 12–32). Cambridge: Cambridge University Press. https://doi.org/10.1017/9780511820076.002.
Lindeman, M., & Aarnio, K. (2007). Superstitious, magical, and paranormal beliefs: An integrative model. Journal of Research in Personality, 41(4), 731–744.
Lindeman, M., & Svedholm-Häkkinen, A. M. (2016). Does poor understanding of physical world predict religious and paranormal beliefs? Applied Cognitive Psychology, 30(5), 736–742. https://doi.org/10.1002/acp.3248.
Manza, L., Hilperts, K., Hindley, L., Marco, C., Santana, A., & Vosburgh Hawk, M. (2010). Exposure to science is not enough: The influence of classroom experiences on belief in paranormal phenomena. Teaching of Psychology, 37(3), 165–171.
Maynes, J. (2015). Critical thinking and cognitive bias. Informal Logic, 35(2), 183–203.
McLean, C. P., & Miller, N. A. (2010). Changes in critical thinking skills following a course on science and pseudoscience: A quasi-experimental study. Teaching of Psychology, 37(2), 85–90.
Nottelmann, N. (2016). The varieties of ignorance. In R. Peels & M. Blaauw (Eds.), The epistemic dimensions of ignorance (pp. 33–56). Cambridge: Cambridge University Press. https://doi.org/10.1017/9780511820076.003.
Özmen, H. (2004). Some student misconceptions in chemistry: A literature review of chemical bonding. Journal of Science Education and Technology, 13(2), 147–159. https://doi.org/10.1023/B:JOST.0000031255.92943.6d.
Peels, R. (2010). What is ignorance? Philosophia, 38(1), 57–67.
Peels, R. (2011). Ignorance is lack of true belief: A rejoinder to Le Morvan. Philosophia, 39(2), 345–355.
Peels, R. (2012). The new view on ignorance undefeated. Philosophia, 40(4), 741–750.
Pennycook, G., Cheyne, J. A., Seli, P., Koehler, D. J., & Fugelsang, J. A. (2012). Analytic cognitive style predicts religious and paranormal belief. Cognition, 123(3), 335–346. https://doi.org/10.1016/j.cognition.2012.03.003.
Posner, G., Strike, K., Hewson, P., & Gertzog, W. (1982). Accommodation of a scientific conception: Toward a theory of conceptual change. Science Education, 66(2), 211–227.
Potvin, P., & Cyr, G. (2017). Toward a durable prevalence of scientific concept: Tracking the effects of two interfering misconceptions about buoyancy from preschoolers to teachers. Journal of Research in Science Teaching, 54(9), 1121–1142.
Pressman, M. R. (2011). Common misconceptions about sleepwalking and other parasomnias. Sleep Medicine Clinics, 6(4), 13–17.
Quine, W. V. O. (1951). Two dogmas of empiricism. Philosophical Review, 60, 20–43.
Rainbolt, G. W., & Dwyer, S. L. (2012). Critical thinking. The art of argument. Boston: Wadsworth.
Reichenbach, H. (1951/1968). The rise of scientific philosophy. Berkeley/Los Angeles: University of California Press.
Russell, B. (1960). Our knowledge of the external world. New York: Mentor.
Sanger, M. J., & Greenbowe, T. J. (1997). Common students’ misconceptions in electrochemistry: Galvanic, electrolytic, and concentration cells. Journal of Research in Science Teaching, 34(4), 377–398.
Siegel, H. (1989). Epistemology, critical thinking, and critical thinking pedagogy. Argumentation, 3(2), 127–140.
Siegel, H. (2009). Open-mindedness, critical thinking, and indoctrination: Hommage to William Hare. Paideusis, 18(1), 26–34.
Simpson, W. D., & Marek, E. A. (1988). Understandings and misconceptions of biology concepts held by students attending small high schools and students attending large high schools. Journal of Research in Science Teaching, 25(5), 361–364.
Smith, J. P., di Sessa, A. A., & Roschelle, J. (1994). A constructivist analysis of knowledge in transition. Journal of the Learning Science, 3(2), 115–163.
Smithson, M. (1989). Ignorance and uncertainty. Emerging paradigms. New York/Berlin: Springer.
Stark, E. (2012). Enhancing and assessing critical thinking in a psychological research methods course. Teaching of Psychology, 39(2), 107–112. https://doi.org/10.1177/0098628312437725.
Stein, M., Larrabbee, T. G., & Barman, C. R. (2008). A study of common beliefs and misconceptions in physical science. Journal of Elementary Science Education, 20(2), 1–11.
Steup, M. (2018). Epistemology. In E. N. Zalta (Ed.), The Stanford encyclopedia of philosophy. Retrieved September 10, 2019, from https://plato.stanford.edu/archives/win2018/entries/epistemology/.
Sumner, W. G. (1906). Folkways. A study of the sociological importance of usage, manners, customs, mores, and morals. Boston: Ginn.
Taylor, A., & Kowalski, P. (2004). Naive psychological science: The prevalence, strength and sources of misconceptions. Psychological Record, 54(1), 15–25.
Taylor, A. K., & Kowalski, P. (2012). Students’ misconceptions in psychology: How you ask matters… sometimes. Journal of the Scholarship of Teaching and Learning, 12(3), 62–72.
Taylor, A. K., & Kowalski, P. (2014). Student misconceptions: Where do they come from and what can we do? In V. A. Benassi, C. E. Overson, & C. M. Hakala (Eds.), Applying science of learning in education: Infusing psychological science into the curriculum (pp. 259–273). Washington: Society for the Teaching of Psychology.
Thagard, P. (2000). Coherence in thought and action. Cambridge: MIT Press.
Thagard, P. (2007). Coherence, truth, and the development of scientific knowledge. Philosophy of Science, 74(1), 28–47.
Todd, C. (2018). Fitting feelings and elegant proofs: On the psychology of aesthetic evaluation in mathematics. Philosophia Mathematica, 26(2), 211–233. https://doi.org/10.1093/philmat/nkx007.
Tversky, A., & Kahneman, D. (1983). Extension versus intuitive reasoning: The conjunction fallacy in probability judgment. Psychological Review, 90(4), 293–315. https://doi.org/10.1037/0033-295X.90.4.293.
van Cleve, J. (2005). Why coherence is not enough: A defense of moderate foundationalism. In M. Steup & E. Sosa (Eds.), Contemporary debates in epistemology (pp. 168–180). New York/London: Blackwell.
van Quine, W., & Ullian, J. S. (1970). The web of belief. New York: Random House.
Willard, A. K., & Norenzayan, A. (2013). Cognitive biases explain religious belief, paranormal belief, and belief in life’s purpose. Cognition, 129(2), 379–391. https://doi.org/10.1016/j.cognition.2013.07.016.
Wilson, J. A. (2018). Reducing pseudoscientific and paranormal beliefs in university students through a course in science and critical thinking. Science and Education, 27(1–2), 183–210. https://doi.org/10.1007/s11191-018-9956-0.
Woodward, J. (2017). Scientific explanation. In E. N. Zalta (Ed.), The Stanford encyclopedia of philosophy. Retrieved September 10, 2019, from https://plato.stanford.edu/archives/fall2017/entries/scientific-explanation/.
Wynn, L. L., Foster, A. M., & Trussell, J. (2009). Can I get pregnant from oral sex? Sexual health misconceptions in e-mails to a reproductive health website. Contraception, 79(2), 91–97.
Zohar, A., Weinberger, Y., & Tamir, P. (1994). The effect of the biology critical thinking project on the development of critical thinking. Journal of Research in Science Teaching, 31(1), 183–196. https://doi.org/10.1002/tea.3660310208.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Authors are listed alphabetically, this article was truly cooperative.
Rights and permissions
About this article
Cite this article
Dellantonio, S., Pastore, L. Ignorance, misconceptions and critical thinking. Synthese 198, 7473–7501 (2021). https://doi.org/10.1007/s11229-019-02529-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11229-019-02529-7