Social Theories of Reasoning
Dissertation
Presented in Partial Fulfillment of the Requirements for the Degree Doctor of
Philosophy in the Graduate School of The Ohio State University
By
Paul David Robinson, M.Sc.
Graduate Program in Philosophy
The Ohio State University
2020
Dissertation Committee
Richard Samuels, Advisor
Neil Tennant
Declan Smithies
Copyrighted by
Paul David Robinson
2020
2 Abstract
This dissertation consists of four independent papers that address topics concerning the nature and function of reasoning. Together, they draw on a wide range of paleontological, archaeological, and ethnographic evidence, to constrain evolutionary theorizing about reasoning. Further, they draw on empirical results from literatures on persuasion and reasoning, to evaluate hypotheses about the characteristic features of reasoning. On these foundations, I argue against the increasingly influential view that the capacity for reasoning evolved in a social context, for public practices of persuasion and self- justification. The principal thesis of this dissertation is that the capacity for reasoning evolved for the private practice of generating true beliefs from rule-like cultural knowledge transmitted through stories, to solve novel problems in the natural environment.
ii Dedication
To Alan Weir
iii Acknowledgments
I wish to thank the following people. My advisor, Richard Samuels, who provided invaluable feedback. He is the second author of the fourth chapter in this dissertation, which was published in the Routledge Series in Philosophy of Mathematics and Physics, and I am grateful to him and to Routledge for permission to include it here. Dan Sperber and Tania Lombrozo, both of whose research I draw on, provided encouragement during the critical early stages of the project. Neil Tennant, Declan Smithies, and Kevin Scharp, helped structure the project, and ensured I didn’t lose sight of the big picture. I am fortunate that Neil Tennant is also my academic grandfather, having mentored my first philosophy teacher, Alan Weir, to whom this dissertation is dedicated. Alan has supported me throughout my studies, which have taken me all over the globe (though, thankfully, not yet to an Independent Scotland). I am grateful to all my teachers, who helped me draw from multiple disciplines: the archaeologists, Aren Maeir and Amit Dagan; the cognitive scientists, Nick Chater, Dave Lagnado, and Tom Lawson; the social psychologists, Duane
Wegener, Kentaro Fujita, and Russ Fazio. I have been shaped by many philosophical discussions with Gonga, Billy, Geezer, Jackie, Rab, Marty, Big Gilbert, and Wee Jim. I have had constant support from my parents, David and Brenda, and my American family,
Akram and Monda. Finally, I am most grateful to Lydia, who “deserves a medal”.
iv Vita
2001 ……………………………….. A-level Mathematics, Physics, Biology Portora Royal School
2005 ……………………………….. B.A. Philosophy Department of Philosophy Queen’s University Belfast
2008 ……………………………….. M.Sc. History and Philosophy of Science Department of Logic and Scientific Method London School of Economics
2009 ……………………………….. M.Sc. Cognitive and Decision Sciences Department of Psychology and Language Sciences University College London
2012 to present …………...... Graduate Teaching Associate Department of Philosophy The Ohio State University
Publications
Robinson, P. D. & Samuels, R. (2018). Reasoning, rules, and representation. In S. Bangu (ed.) Naturalizing Logico-Mathematical Knowledge: Approaches from Psychology and Cognitive Science (Routledge Series in Philosophy of Mathematics and Physics), pp. 30-51.
Fields of Study
Major Field: Philosophy
Interdisciplinary Specialization: Cognitive and Brain Sciences
v
Table of Contents
Abstract ...... ii
Dedication ...... iii
Acknowledgments ...... iv
Vita...... v
List of Tables ...... ix
List of Figures ...... x
Introduction ...... 1
1. Theoretical Background ...... 2
2. Overview ...... 7
Chapter 1 ...... 8
Chapter 2 ...... 9
Chapter 3 ...... 10
Chapter 4 ...... 11
3. Methodology ...... 12
Chapter 1. The Evidence is Not on My Side: Reason, Evolution, and Bias ...... 13
1. Introduction ...... 13
vi 2. The Rationalization-First View ...... 18
2.1 The Theory ...... 19
2.2 The Empirical Case...... 21
3. Does the Rationalization-First View Explain Myside Bias?...... 26
3.1 Persuasion ...... 26
3.2 Information-Gain ...... 34
3.3 Deception...... 37
4. From Persuasion to Justification ...... 39
5. The Inquiry-First View ...... 43
6. Concluding Remarks ...... 49
Chapter 2. Cracking the Enigma: Cultural Knowledge, Social Learning, and Private Reason ..... 51
1. Introduction ...... 51
2. The Rationalization-First View of Reason ...... 55
2.1 Exposition of the Theory ...... 55
2.2 Criticism 1: The Paleo-Anthropological Context ...... 59
2.3 Criticism 2: The Scope of Reason ...... 66
3. A New Inquiry-First View of Reason ...... 69
4. Empirical Characteristics of Reason ...... 80
5. Concluding Remarks ...... 84
Chapter 3. Wason Confirmed: Why Confirmation Bias is Not Myside Bias in Disguise ...... 86
1. Introduction ...... 86
2. Confirmation Bias ...... 90
2.1 The Four Card Selection Task ...... 90
2.2 The 2-4-6 task ...... 105
vii 3. Myside Bias...... 113
4. Concluding Remarks ...... 121
Chapter 4. Reasoning, Rules, and Representation ...... 123
1. Introduction ...... 123
2. The Virtues of Intentional Rule-Following Accounts of Inference ...... 128
3. The Regress ...... 132
3.1 The General Strategy ...... 133
3.2 Establishing Condition 2 ...... 134
3.3 The Regress Within Reach ...... 136
3.4 The Regress Regimented ...... 138
4. Sub-Personal Processes, Revenge, and the Strengthened Regress ...... 140
4.1 Getting Sub-Personal...... 140
4.2 The Revenge Regress ...... 141
4.3 Regress Strengthened ...... 143
5. Rejecting the Strengthened Regress ...... 145
5.1 The “Primitivist” Strategy ...... 146
5.2 Primitive Processes and Reflexes ...... 147
5.3 Primitive Processes and Stored Program Computers ...... 148
5.4 Primitive Processes and Cognitive Architecture...... 149
6. Counter-Arguments ...... 150
6. Conclusion...... 156
References ...... 157
viii List of Tables
Table 1: Rationalization-first views of reason ...... 5
Table 2: Inquiry-first views of reason ...... 5
Table 3. Card selections predicted by strategies in the negations paradigm ...... 101
ix List of Figures
Figure 1: Card faces in the selection task ...... 90
Figure 2: Card-turning scenarios in the selection task ...... 91
Figure 3: Canonical card selections made in the selection task ...... 93
Figure 4. Card faces in the selection task explicit negations paradigm...... 104
Figure 5. Performance of selection task strategies with respect to falsification...... 110
Figure 6. Performance of selection task strategies, restricted to relevant cases ...... 112
Figure 7. Arguments cited for and against one’s view (Perkins) ...... 114
Figure 8. Arguments cited for and against one’s view (Toplak & Stanovich) ...... 117
x Introduction
Of all the faculties of the human mind, it will, I presume, be admitted that reason stands at the summit. Only a few persons now dispute that animals possess some power of reasoning.
— Charles Darwin, The Descent of Man (1871/1874)
Neither can I easily bring myself to the idea that man’s reasoning faculties & above all his moral sense, cd. ever have been obtained from irrational progenitors, by mere natural selection … This seems to be doing away altogether with the Divine Image which forms the insurmountable distinction between man & brutes.
— Leonard Jenyns, letter to Charles Darwin (4 Jan, 1860)
The topics addressed in this dissertation fall squarely within a longstanding debate regarding the nature of reasoning (what reasoning is) and the function of reasoning (what reasoning is for). Although the debate has been especially prominent within the evolutionary and cognitive sciences, it is also one to which empirically oriented philosophers have contributed. It is in this spirit that I approach the topics in this dissertation.
1 The psychological phenomenon at issue is what some philosophers have called
“active” reasoning (e.g. Broome, 2013). Minimally characterized, active reasoning is a person-level process, involving reflection on explicitly represented reasons to draw a conclusion. Henceforth, all uses of the term “reasoning” should be taken to refer to active reasoning. Thus, this dissertation is not concerned with unconscious, sub-personal, inference. Nor is it restricted to any particular kind of reasoning, such as logical reasoning, probabilistic reasoning, analogical reasoning, or causal reasoning. Rather, my focus is on reasoning tout court.
In contemporary discussions of reasoning, “function” is taken to be synonymous with primary adaptive function, unless otherwise qualified. To specify the primary adaptive function of a trait is to identify fitness-enhancing biological effects of the trait that led it to persist in the population. The principal thesis of this dissertation is that the function of reasoning is the private generation of true beliefs to solve novel problems in the natural environment.
In the remainder of this introduction, I provide a theoretical background for this dissertation (section 1), an overview of its four chapters (section 2), and some brief comments on methodology (section 3).
1. Theoretical Background
The primary goal of this dissertation is to critically assess the recent “social turn” in the study of reasoning. That reasoning evolved for a social function was first proposed in detail by the linguist, philosopher, and cognitive anthropologist, Dan Sperber, in his research on
2 the evolution of human capacities for metarepresentation and verbal communication
(Sperber, 2000, 2001a). According to his “argumentative theory of reasoning”, reasoning evolved not for the “individual Cartesian thinker” to generate knowledge (Sperber, 2001a, p. 408) but rather to produce and evaluate arguments that individuals use to manipulate others through persuasion.
Around the same time, the social psychologist Jonathan Haidt proposed his highly influential social intuitionist model of moral judgment, according to which reasoning tends to justify the reasoner’s intuitions because it evolved to protect and enhance the reasoner’s reputation, rather than to search for the truth (Haidt, 2001, 2012a).
Recent work by Mercier and Sperber incorporates both persuasion and self- justification into an overarching theory of reason:
Reason, we argue, has two main functions: that of producing reasons for justifying oneself, and that of producing arguments to convince others. These two functions rely on the same kinds of reasons and are closely related. (Mercier & Sperber, 2017, p. 8)
Since both persuasion and self-justification are social practices, they refer to this new theory as the “interactionist” account of reason. The interactionist account is increasingly popular, with many researchers in philosophy, psychology, and the cognitive sciences finding the core claims compelling (see, e.g., De Neys, 2018a; Dutilh Novaes, 2018;
Evans, 2019; Sterelny, 2018).
The interactionist account is typically contrasted with a supposed “intellectualist” tradition which holds that reason evolved to “improve individual cognition and arrive on
3 one’s own at better beliefs and decisions” (Mercier & Sperber, 2017, p. 330). However, it is important to distinguish the function of reasoning and the contexts reasoning evolved to operate in. Given the nature of the functions Mercier and Sperber posit, the relevant contexts will always be social. Nevertheless, the two can come apart. For example, the capacity to reason may have evolved for individuals to gain knowledge through group argumentation. Here the context is social (interactionist), but the function is epistemic
(intellectualist). Thus, I introduce the terms “rationalization-first” and “inquiry-first” to partition theories of reason. According to rationalization-first views, reason evolved to support the reasoner’s pre-existing beliefs and decisions for the purposes of persuasion and/or self-justification (Table 1).1
According to inquiry-first views, reason evolved to test the reasoner’s pre-existing beliefs and decisions, and/or to discover new truths (Table 2). There are prima facie plausible asocial versions of the inquiry-first view, according to which reason evolved for
1 “Rationalization” was adopted as a technical term in Freudian psychology, where it originally referred to the practice of offering socially acceptable motives for one’s actions that hide the repressed true motives (Lundholm, 1933; Taylor, 1923). However, it was also commonly used to cover implicit justifications of one’s beliefs and attitudes (Hollitscher, 1939; Jones, 1908; Piddington, 1928; Robinson, 1921; Walker, 1928). Some communication scholars further expanded the concept to include a type of persuasion that falls between genuine deliberation and emotional appeal (Oliver, 1936). I use “rationalization” to refer to the post hoc practice of attempting to cite normative reasons in favor of an actual or pretended judgment or decision (where these result in beliefs, attitudes, intentions, etc.), typically under the guise of those reasons also explaining why the relevant conclusion was reached.
4 lone individuals to attain epistemic goals through private deliberation. This dissertation develops and defends one such account.
SOCIAL ASOCIAL
PERSUASION SPERBER (2001) NA
SELF-JUSTIFICATION HAIDT (2001) NA
BOTH MERCIER & SPERBER (2017) NA
Table 1: Rationalization-first views of reason.
SOCIAL ASOCIAL
TESTING
DISCOVERY ROBINSON (2020)
BOTH INTELLECTUALISM
Table 2: Inquiry-first views of reason.
Inquiry-first views face genuine challenges. First, there is the challenge of explaining why the relevant epistemic goals were fitness-enhancing, and why they were best achieved through reasoning. Answers to these questions are often gestured towards rather than explicitly spelled out (e.g. Berwick & Chomsky, 2016, p. 166). Although it is clear that reasoning – used in conjunction with the more recent ability to read and write – underpins modern achievements such as science and engineering and perhaps even the occasional
5 doctoral dissertation, it is not a trivial task to show that reasoning was highly beneficial in ancestral environments.
Second, there is the challenge of explaining what appear to be irrational tendencies of reasoning. For instance, within psychology there is a growing consensus that reasoning tends to justify rather than correct mistaken intuitions (De Neys, 2018b). This is often taken to be one manifestation of a wider “confirmation bias” – or, as Mercier and Sperber prefer to call it, “myside bias” – that reasoning seems to display. Such a bias is especially problematic for asocial versions of the inquiry-first view, because they hold that reasoning evolved to operate without critical feedback from interlocutors. For instance, in an interview for a popular audience that has been viewed online over a quarter of a million times, Haidt states:
Something we need to talk about here is what's called the confirmation bias. That is, you might think that our reasoning is designed to find the truth. And if you want to find the truth, you should look on both sides of a proposition. But in fact what happens is, when someone gives you a proposition, our minds, we send them out, we sent them out to do research for us. But it’s research, like, as a lawyer does, or as a press secretary would do, it’s like, “Find me one piece of evidence that will support this claim that I want to make”. And if I can find one piece of evidence, I'm done. I can stop thinking. Well, that's the way we've been for … hundreds of thousands of years. (Haidt, 2012b)
Since the notion of confirmation bias is well-known even outside of academia, and is often cited to explain pressing societal issues such as increased levels of polarization, the claim that confirmation bias conflicts with inquiry-first views is rhetorically powerful.
6 Moreover, it is difficult to see how the kinds of epistemic functions that characterize inquiry-first views can be reconciled with “motivated reasoning” (Kunda, 1990) and
“motivated skepticism” (Ditto & Lopez, 1992; Taber & Lodge, 2006), i.e., the tendency to judge preference-consistent information as more valid or accurate than preference- inconsistent information, due to subjecting preference-inconsistent information to a greater degree of critical reasoning. For although considerations of truth play some role, the process is “best characterized as a compromise between the wish to reach a particular conclusion and the plausibility of that conclusion given the available data” (Ditto & Lopez
1992, 569).
The secondary goal of this dissertation is to meet these challenges by developing an inquiry-first view of reason that explains the empirical features of reason, while remaining faithful to the paleo-anthropological evidence concerning human evolution.
2. Overview
Each chapter in this dissertation is written as an independent paper. However, the topics addressed form a natural progression: from critiquing the empirical case for the rationalization-first view, to developing my own version of the inquiry-first view, to reconsidering the received wisdom about confirmation bias and myside bias, to defending a kind of rule-following account of the nature of reasoning that dovetails with (though is not presupposed by) my account of the function of reasoning.
7 Chapter 1
This chapter argues that the existence of myside bias does not support the rationalization- first view over the inquiry-first view.
As glossed by Mercier and Sperber, myside bias results in people being good at formulating arguments which support their beliefs, but bad at formulating arguments which counter their beliefs. Mercier and Sperber claim that the rationalization-first view explains myside bias, since producing arguments that support one’s beliefs facilitates persuasion and self-justification.
My response is as follows. Empirical research on communication suggests that if the function of argument production is persuasion, then people should be good at anticipating objections to their beliefs. What allows the rationalization-first view to explain myside bias is not the function of reason it posits, but rather the auxiliary hypothesis that on average it was cost-effective to offload to interlocutors the task of formulating counterarguments. Ethnographic studies of hunter-gatherer societies suggest that this auxiliary hypothesis is implausible.
However, even if the auxiliary hypothesis is granted, inquiry-first views can mimic the rationalization-first strategy for explaining myside bias. Social versions of the inquiry- first view can claim that myside bias evolved to enable a division of argumentative labor that enables groups to pursue epistemic goals efficiently. Asocial versions of the inquiry- first view can claim that myside bias evolved as a cost-effective means of checking whether there are good grounds for one’s beliefs.
8 Chapter 2
This chapter argues that: (i) the rationalization-first view is based on an inaccurate reconstruction of ancestral environments, and fails to capture inquisitive reasoning; whereas (ii) my version of the inquiry-first view is based on an accurate reconstruction of ancestral environments, and captures argumentative reasoning.
Mercier and Sperber propose that the capacity to reason evolved to overcome an information bottleneck in verbal communication that mainly resulted from the threat that senders of messages were engaging in deception. My response is as follows. Given the small-scale, egalitarian nature of the societies that existed when verbal communication evolved, the risk of deception was negligible. Moreover, a reasoning mechanism adapted for backwards inference, from pre-existing beliefs to arguments in favor of those beliefs, is fundamentally ill-suited for the kind of forwards inference from premises to previously unconsidered conclusions that enables science and engineering, for example, as well as a host of other distinctly human achievements.
I propose that the capacity to reason instead evolved to generate true beliefs for the purpose of solving novel problems in the natural environment, for which the reasoner has no intuitions. Drawing on ethnographic studies of hunter-gatherer societies, I argue that the fitness of individuals in the relevant ancestral environments was enhanced by reasoning taking as input rule-like (generalizable) cultural knowledge transmitted through stories. Since under this account reason is a problem-solver, it is a highly flexible system that can serve a wide variety of motives and goals. In particular, reasoning may
9 be used to persuade others, justify oneself, or find fault with views one doesn’t want to believe. Thus, my account captures argumentative reasoning. Further, since I do not claim that reasoning evolved to test and correct the reasoner’s pre-existing beliefs, my account is compatible with the existence of myside bias and motivated reasoning, even without adding an auxiliary hypothesis of the sort discussed in chapter 1.
Chapter 3
This chapter argues that there is good evidence for confirmation bias with respect to hypothesis-testing, but no good evidence for myside bias with respect to one’s beliefs.
Mercier claims that confirmation bias holds relative to the beliefs of the reasoner but not relative to mere hypotheses the reasoner is testing, and is thus better described as
“myside bias”. My response is as follows. When people test a hypothesis that takes the form of a conditional statement, rather than search for falsifying cases, they tend to unintentionally search for instances of the conditional which, if discovered, confirm the hypothesis. It is in this sense that a bias towards confirming evidence exists.
In contrast, there is no good evidence that people tend to unintentionally search for evidence that favors their beliefs. The classic myside bias effect is that when people are tasked with coming to a conclusion on a controversial topic, and are then asked to explain the reasoning for their conclusion, they tend to provide more arguments for their view than arguments against it. I argue that these effects can be explained away as reflecting why the (unbiased) reasoner came to the conclusion they did.
10 Chapter 4
This chapter, which is co-authored with Richard Samuels, argues that no vicious regress results from supposing that reasoning essentially involves following rules that are explicitly represented.
What we call “intentional rule-following” accounts of the nature of reasoning make the following two commitments. First, reasoning essentially involves following rules concerning the premises from which one reasons. Second, all rule-following involves
(conscious or unconscious) intentional states which represent the rules being followed. A number of philosophers – e.g. Boghossian, Broome, and Wright – argue that such theories fall prey to a vicious regress. The problem is generated by supposing that all rule-following involves grasping what the rule states, forming a belief that its antecedent is satisfied, and drawing the conclusion the act required by its consequent is required. In other words, all rule-following involves reasoning.
Our response is as follows. We posit a level of primitive processing mechanisms that take rule-representations as inputs. The primitive processes they subserve will be rule-guided in the thin sense that a rule-representation is causally implicated in the process. However, those primitive processes are not rule-guided in the sense that they involve further rule-guided or inferential sub-processes. Thus, if non-primitive processes—such as those involved in reasoning—ultimately decompose into primitive ones, no regress is generated.
11 3. Methodology
Research on the nature and function of reasoning has tended to proceed in disciplinary isolation. Although previous work has cited empirical data from the “psychology of reasoning” literature, I also draw on empirical data from the “attitudes and persuasion” and “communication studies” literatures. Further, although empirical claims are routinely made about ancestral environments, evidence is rarely cited.2 This dissertation aims to make a unique contribution to the literature by using paleontological, archaeological, and ethnographic data to constrain evolutionary theorizing about the function of reasoning.
The notion of empirical confirmation at play in this dissertation is the comparative confirmation of one theory relative to another, in the light of all the available evidence, given the assumption that human evolution was a wholly naturalistic affair.
Due to the interdisciplinary ethos of this dissertation, a number of terminological clarifications are in order. I use “attitude” to refer to an evaluative summary of an object, rather than a mental state towards a proposition. With respect to reports of meta- analyses, “r” refers to the correlation co-efficient and “k” refers to the number of studies surveyed. Finally, “my” and “ky” abbreviate “millions of years ago” and “thousands of years ago”, respectively.
2 A similar observation is made by Sterelny (2018, p. 4).
12 Chapter 1 The Evidence is Not on My Side: Reason, Evolution, and Bias
1. Introduction
In recent years a number of influential researchers in philosophy, psychology, and cognitive science have converged on the idea that the human capacity to reason evolved for the attainment of socially strategic goals through persuasion and self-justification (e.g.
Haidt, 2012a; Mercier & Sperber, 2017). Call this the “rationalization-first” view of reason.
Rationalization-first because typically these public practices of persuasion and self- justification are attempts to cite normative reasons in favor of a belief or decision that the reasoner has already made, often under the guise of those reasons explaining how the relevant conclusion was reached. Rationalization-first because the capacity to reason is held to have emerged and persisted within the Homo sapiens lineage mainly due to the fitness-enhancing effects of persuasion and self-justification. Given that these are the primary adaptive functions of reason, ceteris paribus they should be the tasks to which our reasoning system naturally defaults.
13 The rationalization-first view is often contrasted with the supposedly dominant idea that the primary adaptive function and modus operandi of reason is to attain epistemic goals, such as avoiding false beliefs and gaining true beliefs (e.g. Haidt, 2012a, p. 74; Mercier & Sperber, 2017, p. 218). Call this the “inquiry-first” view of reason. Inquiry- first because attaining these goals involves attempting to ascertain the facts about some issue. Applied to a preexisting belief or decision, inquiry is a form of testing through the weighing of normative reasons concerning the relevant conclusion. Alternatively, in the absence of a preexisting belief or decision, inquiry is a form of discovery that searches for relevant evidence and seeks to abduce to a new conclusion based on a consideration of the normative implications of the gathered evidence.
Why suppose that the rationalization-first view is preferable to the inquiry-first view? Here is one oft-cited line of argument. Empirical research shows that reasoning displays a characteristic known as “confirmation bias”. Though the term “reasoning” gets used in a variety of ways, the phenomenon of relevance here is not sub-personal inference but rather a person-level activity which delivers a verbalizable awareness of something to the effect that “P is a reason for Q ”. To a first approximation, confirmation bias is the tendency to unwittingly search for evidence in ways which favor information that supports rather than undermines one’s pre-existing belief or working hypothesis (Klayman, 1995;
Nickerson, 1998). The existence of such a bias is puzzling if reason evolved to facilitate the attainment of epistemic goals, since it can lead people to be overconfident that their beliefs are correct. However, confirmation bias appears to facilitate persuasion and self-
14 justification, since it helps the reasoner build a case. Thus, in his bestseller The Righteous
Mind, the social psychologist Jonathan Haidt writes:
confirmation bias is a built-in feature (of an argumentative mind), not a bug that can be removed … each individual reasoner is really good at one thing: finding evidence to support the position he or she already holds. (2012a, p. 105f.)
Sometimes the claim is framed in terms of “myside bias”, which, in this context, is taken to be a restricted form of confirmation bias that holds relative to the reasoner’s beliefs or decisions rather than mere hypotheses they happen to be testing.3 Thus conceived, myside bias will result in people being better at formulating arguments which support their beliefs and decisions, than formulating arguments which undermine their beliefs and decisions. Hence, in The Enigma of Reason the cognitive scientists Hugo Mercier and Dan
Sperber state:
A lot of evidence shows that reasoning has a myside bias… This is pretty much the exact opposite of what you should expect of a mechanism that aims at improving one’s beliefs through solitary ratiocination … [However] If the function of reasoning, when it produces reasons, is to justify one’s actions or to convince others, then it should have a myside bias. (2017, p. 218)
Arguably, the rationalization-first view and the associated claim that it explains myside bias are becoming the mainstream position.4 For instance, they have been used as a
3 Henceforth, I refer only to “myside bias”. 4 For pushback see critical reviews of The Enigma of Reason by Dutilh Novaes (2018) and Sterelny (2018).
15 framework for understanding System 2 in dual-process theories of reasoning, and the tendency to defend rather than correct mistaken initial responses in experimental tasks
(De Neys, 2018a; Evans, 2019). It has also been suggested that they help elucidate cases of confabulation (Ganapini, 2019), group-level scientific inquiry (Peters, 2019), the propagation of pseudoscientific beliefs (Blancke, Boudry, & Braeckman, 2019), and the reliability of our intellectual faculties (Bland, 2018).
Suppose that myside bias exists in the sense outlined above. What does this empirical phenomenon tell us about the function of reasoning? My goal in this paper is to show that the existence of myside bias does not support the rationalization-first view over the inquiry-first view. To explain myside bias, the rationalization-first view needs to be supplemented by an auxiliary hypothesis that is not warranted by the empirical evidence.
But even if this auxiliary hypothesis is granted, serious theoretical problems remain. For as I will show, the resulting explanation of myside bias undermines core components of the rationalization-first view concerning the evolutionary history and adaptive functions of reason. When this is appreciated, it becomes clear that if the inquiry-first view is supplemented with similar auxiliary hypotheses, it provides an explanation of myside bias which is at least as plausible as the explanation provided by the rationalization-first view.
Here’s how I proceed. I begin by reconstructing the standard version of the rationalization-first view, according to which the ability to produce and evaluate arguments evolved to overcome an information bottleneck which resulted largely from the threat of being misinformed through deception. Proponents of the rationalization-first
16 view claim that since the capacity to produce arguments evolved for senders of messages to persuade the receivers of those messages, the view predicts that people will be good at producing arguments in support of their beliefs but relatively bad at producing arguments that undermine their beliefs.5
The main burden of this paper is to develop my objections to this claim (section 3).
I show that if the adaptive function of argument production is persuasion, then ceteris paribus people should be good at anticipating objections to their beliefs. Mercier and
Sperber appear to suggest that it was advantageous for senders to offload this task to receivers, because anticipating objections is cognitively demanding. Based on paleontological, archaeological, and ethnographic evidence, I argue that this auxiliary hypothesis depends on a cost-benefit analysis that is unlikely to hold for the environment of evolutionary adaptedness. However, even if the auxiliary hypothesis is granted, the resulting explanation of myside bias undermines two core components of the rationalization-first view. The first is that argument evaluation evolved to help receivers accept genuine information and reject misinformation. The second is that argument production evolved in part to help senders deceive receivers with arguments in favor of views the senders don’t hold. Although these two problems do not suffice to show that the rationalization-first explanation of myside bias is incorrect, they do substantially weaken its plausibility. In section 4, I examine whether the problems can be overcome by
5 Of course, over time any given individual will act as a sender and as a receiver.
17 hypothesizing that justification rather than persuasion is the primary adaptive function of argument production. Although this modification has some independent merit, I show that the relevant problems resurface for the rationalization-first view.
Finally, I sketch alternative explanations of myside bias under the inquiry-first view. On the one hand, we can imagine social versions of the inquiry-first view, according to which the capacity to reason evolved for the attainment of epistemic goals through the public trading of arguments. Such views can explain myside bias as resulting from a cost- effective division of argumentative labor that is efficient at reaching the truth, without the need to accommodate an evolved capacity for deception. On the other hand, we can imagine asocial versions of the inquiry-first view, according to which the capacity to reason evolved for the attainment of epistemic goals through private deliberation.6 Such views can explain myside bias as resulting from a cost-effective heuristic that tests the reasoner’s beliefs by checking whether there are good grounds to hold them. I thus conclude that the existence of myside bias fails to adjudicate between the rationalization-first view and the inquiry-first view.
2. The Rationalization-First View
The adaptive function of any trait is wedded to its evolutionary etiology. The rationalization-first view represents a thoroughgoing departure from traditional ways of
6 That there is no plausible asocial version of the rationalization-first view follows from the nature of the functions it posits.
18 thinking about reason, based on considerations relating to the social aspects of human evolution. In this section I clarify what the proposal is.
2.1 The Theory
The rationalization-first view was first proposed in detail by Dan Sperber in his work on the evolution of human capacities for verbal communication (Sperber, 2000, 2001a). For communication to be evolutionarily stable in a population the benefits of both sending and receiving messages must, on average, outweigh their costs. In particular, if sending messages had on average been more costly than beneficial then the practice would have ceased, resulting in the complementary practice of receiving messages also ceasing (and vice versa). The verbal exchange of messages allows senders to manipulate how receivers behave, and receivers to gain useful information from senders. This practice is often highly beneficial for both sender and receiver, as when it is used to coordinate cooperative ventures. However, it can also be highly detrimental for a receiver to be misinformed by a sender.
Sperber suggested that in the early stages of the history of verbal communication the threat of being misinformed generated selective pressure for mechanisms of “epistemic vigilance”. One set of mechanisms evaluate the source of the information based on cues relating to the competence and honesty of the sender. Another set evaluate the content of the information based on how it coheres internally and with respect to some relevant subset of the receiver’s beliefs. The major factor driving the evolution of these mechanisms
19 was not the risk that senders were mistaken, but rather the risk that senders were engaging in deception:
the major problem posed by communicated information has to do not with the competence of others, but with their interests and their honesty. While the interests of others often overlap with our own, they rarely coincide with ours exactly. In a variety of situations, their interests are best served by misleading or deceiving us. It is because of the risk of deception that epistemic vigilance may be not merely advantageous but indispensable if communication itself is to remain advantageous. (Sperber et al., 2010, p. 360)
At this point a vigilance-persuasion “arms race” developed. To the extent that epistemic vigilance delivered “true positives” that blocked misinformation, it deprived senders of the benefits of deception. On the other hand, when epistemic vigilance delivered “false positives” that blocked genuine information, it deprived both senders and receivers of the benefits that would have resulted from cooperation based on the shared information. This
“information bottleneck” generated selective pressure for senders to take advantage of epistemic vigilance through “honest displays” in which they present considerations in favor of the target message that the receiver either already accepts or will accept more readily than the target message in isolation:
The argumentative use of reasons helps genuine information cross the bottleneck that epistemic vigilance creates in the social flow of information. It is beneficial to addressees by allowing them to better evaluate possibly valuable information that they would not accept on trust. It is beneficial to communicators by allowing them to convince a cautious audience. (Mercier & Sperber, 2017, p. 194)
20 The evolution of the ability to produce arguments in support of a target message stimulated the flow of genuine information.7 However, it also provided an additional means by which senders could deceive receivers, through specious arguments:
sophistry is a way to use the “honest display” strategy of argumentation in a dishonest way and thereby make it more advantageous for the communicator. In other words, sophistry contributes to making argumentation adaptive. (Sperber, 2001a, p. 411)
Senders engaging in argumentative deception generated further selective pressure for the ability to critically evaluate arguments. In turn, this led to a sharpening of the ability to produce persuasive arguments. And so on. In this way, a specialized cognitive system
(“reason module”) dedicated to the production and evaluation of arguments evolved for persuasion and information-gain.8
2.2 The Empirical Case
In evolutionary approaches to cognition, research is often described as being either theory- driven or observation-driven (Lewis, Al-Shawaf, Conroy-Beam, Asao, & Buss, 2017; Reeve
& Sherman, 2007). A theory-driven “top-down” approach specifies conditions in ancestral environments that would have impacted biological fitness (i.e., hypothetical selection pressures), and then postulates the evolution of a psychological trait which increased
7 Mercier (2020) prefers the term “open vigilance” to “epistemic vigilance” because the mechanisms are as much about accepting genuine information as rejecting misinformation. 8 This was dubbed the “argumentative theory of reasoning” (Sperber et al., 2010).
21 fitness in that environment (i.e., a hypothetical adaptation). In contrast, an observation- driven “bottom-up” approach specifies a currently existing psychological trait, and then postulates selection pressures in past ancestral environments which would render the trait an adaptation. Although the latter approach is often criticized as “just-so storytelling”, qua science both the top-down and bottom-up approaches must be used to generate novel predictions, which may concern either the nature of the trait (its causes and effects) or the nature of ancestral environments (as revealed by archaeology, paleontology, and the like).
The early development of the rationalization-first view represents a top-down approach. Indeed, when the theory was first proposed by Sperber around the turn of the millennium, no empirical reasoning research was cited. Instead, Sperber used the theory to predict that reasoning will be facilitated in argumentative contexts:
logically equivalent tasks will yield significantly better performance when they are presented in a context where subjects are scrutinizing arguments plausibly aimed at persuading them than when they are evaluating these arguments in the abstract (as happens in most experimental tasks). (Sperber, 2000, p. 136)
By the turn of the next decade, Sperber – with his new collaborator Hugo Mercier – began to link the argumentative theory to results in the psychology of reasoning literature. The finding that groups outperform individuals on average in experimental tasks on reasoning
(e.g. Moshman & Geil, 1998) was taken to confirm the prediction that argumentative contexts should facilitate performance (Mercier & Sperber, 2009, p. 261). The rationale for this prediction is that the ability to evaluate arguments evolved to facilitate the acceptance
22 of genuine information and the rejection of misinformation. Thus, people will be
“demanding so as not to be deceived by poor or fallacious arguments into accepting false ideas” and “objective so as to be ready to revise our ideas when presented with good reasons” (Mercier & Sperber, 2017, p. 332):
DEMANDING: People are persuaded by an argument only when it is good.
OBJECTIVE: People are persuaded by an argument when it is good.
Compared to a single individual, a group of individuals will tend to generate more potential solutions, and more arguments for and against any given solution. Since everyone in the group tends to accurately evaluate arguments, the correct solution is more likely to win out.9 But as Mercier and Sperber seem to recognize (2017, pp. 235; 332),
DEMANDING and OBJECTIVE are also likely to hold if reasoning evolved to facilitate epistemic goals. Hence, the evidence in favor of DEMANDING and OBJECTIVE does not support the rationalization-first view over the inquiry-first view.
However, around this time a new empirical claim began to emerge. This was the claim that the rationalization-first view explains a generalized form of confirmation bias:
when we try to persuade someone that something is true (or false), a confirmation (or disconfirmation) bias may help us achieve our goal. (Mercier & Sperber, 2009, p. 163)
9 There are caveats, of course. For example, if everyone in the group starts out with the same opinion, then “group polarization” may result.
23 The term “confirmation bias” has been used to refer to a wide variety of phenomena. Early research using the 2-4-6 rule-discovery task (Wason, 1960) and four-card selection task
(Wason, 1966) suggested that people tend to test hypotheses in ways that unintentionally privilege supporting evidence at the expense of falsifying evidence, perhaps due to cognitive limitations such as the difficulty of holding more than one hypothesis in mind or forming mental representations of hypothesis-inconsistent information (Evans, 1989).
Mercier and Sperber claim to explain a different kind of phenomenon that holds when what is at stake is not an arbitrary hypothesis but rather the views of the reasoner.
In the classic “myside bias” paradigm, participants are asked to think about a controversial issue and come to a conclusion if possible. When explaining the reasoning for their conclusion, participants tend to provide more (myside) arguments for their view than (otherside) arguments against it (Perkins, 1989). This effect holds even when participants are instructed at the outset to produce arguments on both sides of the issue.
Furthermore, people find it easier to provide counterarguments to opposing beliefs than to their own (Kuhn, 1991, pp. 171; 276). The explanation given by Sperber and Mercier is that since reason evolved to allow senders to persuade receivers, it will search for and output arguments that tend to favor the sender’s beliefs:
MYSIDE BIAS: People produce arguments that tend to favor their beliefs.
24 These arguments can favor the sender’s beliefs directly by concluding to the truth of the sender’s beliefs, or indirectly by concluding to the falsity of an opposing belief (2017, pp. 9,
218, 235):
What they find difficult is not looking for counterevidence or counterarguments in general, but only when what is being challenged is their own opinion. Reasoning does not blindly confirm any belief it bears on. Instead, reasoning systematically works to find reasons for our ideas and against ideas we oppose. (Mercier & Sperber, 2017, p. 218)
Although this may appear to represent a bottom-up approach, as we have seen, the rationalization-first view had already been proposed on independent grounds. However, neither was the existence of myside bias a novel prediction of the theory. Rather, myside bias effects were known and subsequently explained by the theory.10 Crucially, Mercier and Sperber add that MYSIDE BIAS should not hold if reasoning evolved to facilitate epistemic goals (2017, pp. 235; 332). They conclude that the data “unambiguously supports” their theory over what they take to be its main rival (Mercier & Sperber, 2017, pp. 203; 275). Thus, the existence of myside bias is held to provide comparative confirmation of the rationalization-first view relative to the inquiry-first view.
10 Sometimes it is claimed that the existence of myside bias is a “retrodiction” of the theory. However, it is more common to use “retrodiction” to refer to the derivation of unknown data gathered in the past.
25 3. Does the Rationalization-First View Explain Myside Bias?
In this section I show that myside bias does not follow from the adaptive functions posited by the rationalization-first view. Rather, the explanatory work is shouldered by auxiliary assumptions that are empirically unwarranted.
3.1 Persuasion
Suppose the primary adaptive function of argument-production is persuasion. Intuitively, experience shows that a failure to anticipate counterarguments is often fatal to a sender’s attempt to persuade a receiver. For example, my attempt to persuade an open-minded acquaintance to support the introduction of single-payer healthcare to the United States floundered due to my inability in the spur-of-the-moment to deal with their objection that this would increase wait times. Moreover, as I will show, beyond these kinds of anecdotal observations the social science literature on persuasion and argumentation provides robust evidence that anticipating counterarguments aides persuasion. Thus, prima facie, we should expect people to be relatively good at thinking of counterarguments to their views.
Call messages that present at least one myside argument “one-sided” when they do not include any counterargument, and “two-sided” when they do. The communication of a two-sided message may be interpreted as a sign that the sender is uncommitted, uncertain, or confused (Baron, 1995). As a result, asking people to author a message with the explicit goal of persuading an audience sometimes suppresses the communication of counterarguments (Nussbaum & Kardash, 2005; Felton, Crowell, & Liu, 2015). However,
26 the empirical research to-date is clear that two-sided messages tend to be more persuasive than one-sided messages when the relevant counterarguments are rebutted by the sender
(refutational two-sided messages), rather than merely dismissed or deemed to be outweighed by myside arguments (non-refutational two-sided messages). The two landmark meta-analyses investigating this issue, Allen (1998) and O’Keefe (1999), both found that refutational two-sided messages were more persuasive than one-sided messages (r = .074, k = 43; r = .077, k = 42), whereas non-refutational two-sided messages were less persuasive than one-sided messages (r = – .087, k = 26; r = – .049, k = 65).11
Subsequent studies corroborate these meta-analytic results. For example, Wolfe et al. (2009) found that written arguments were more persuasive when the author raised and adequately addressed objections to them. Relatedly, studies of mock jury trials find that the impact of negative information about defendants is reduced when it is preemptively addressed by the defense attorney (e.g. Howard, Brewer, & Williams, 2006).
Expert communicators appear to be sensitive to these two-sidedness effects. A content analysis of published essays, editorials, and opinion pieces in which professional authors
11 The mean effect sizes are small even by newly lowered benchmarks (e.g. Bosco, Aguinis, Singh, Field, & Pierce, 2015). However, they are typical in relation to the persuasion literature specifically. O’Keefe (2009) notes that for message variations which permit meta-analytic conclusions the largest mean effect size is .14. In the context of an evolutionary arms race we should expect effects that are not easily perceptible. Small effects can have a substantial impact when they are repeatedly manifested (Abelson, 1985), as they would be in argumentative exchanges over the course of an individual’s lifetime.
27 were aiming to persuade the general public revealed that they tended to include counterarguments for rebuttal (Wolfe & Britt, 2008, p. 2f.). In contrast, studies which show that the goal of persuasion tended to suppress the use of counterarguments typically used student participants who were merely aiming to fulfil academic assignments. Moreover, practicing attorneys and legal reasoning textbooks almost uniformly recommend that defense teams utilize the technique of “stealing thunder” by pro-actively addressing negative information (Williams et al., 1993).
So far we have considered the merits of the sender pro-actively raising and rebutting counterarguments. In argumentation theory this strategy is known as
“prolepsis” (Walton, 2009). In any particular case, whether the sender should voluntarily communicate counterarguments that come to mind may depend on situational variables such as how compelling the counterarguments are, how likely the receiver is to independently think of the counterarguments, and whether the receiver has a strong pre- existing opinion about the issue. But even when prolepsis is not used, there are two further ways in which anticipating counterarguments might be highly beneficial for the sender.
First, if those counterarguments are subsequently raised by the receiver, the sender will have had more time to think of a suitable framing and rebuttal. And, second, if the sender is aware of multiple myside arguments that establish the target conclusion, the sender can adopt a risk-averse strategy of choosing to communicate what they deem to be the least objectionable line of reasoning. Thus, for example, if a cleric is seeking to persuade a biologist that God exists it may be prudent for the cleric to argue from the
28 premise that the laws of physics are fine-tuned for the existence of life rather than from the premise that certain biological systems are irreducibly complex, even if in the cleric’s experience the latter “intelligent design argument” strikes most people as more intuitively compelling. Although the persuasive effects of these two kinds of strategies have not been studied directly, research on face-to-face conversations by Waldron and colleagues suggests that communication goals such as persuasion are facilitated by the formulation of concrete message plans that take into account the receiver’s perspective, since they allow senders to guide the conversation and better deal with contingencies as they arise
(Waldron, Caughlin, & Jackson, 1995).
Research findings about what is persuasive to well-educated Homo sapiens in the
20th and 21st centuries should not be carelessly generalized to our evolutionary past.
However, the factors proposed to underlie two-sidedness effects would also have been highly relevant in ancestral environments:
• Framing: Two-sided messages allow the sender to frame counterarguments in
favorable ways before they occur to the receiver.
• Reactance: Two-sided messages put less psychological pressure on the receiver
to accept the conclusion, thus reducing psychological reactance to the message.
• Effort: Refutational two-sided messages ease the argumentative burden of the
receiver, thus reducing processing effort and negative feelings.
29 • Reputation: Refutational two-sided messages avoid the receiver publicly raising
the relevant counterargument(s), thus reducing defensiveness due to
reputational concerns.
• Credibility: Refutational two-sided messages increase the perceived credibility
of the sender because they signal that the sender is both competent enough to
consider counterarguments and honest enough to communicate both sides of
the issue.
Indeed, if reasoning evolved to solve the adaptive problem of facilitating communication in cases where the message cannot initially be taken on the basis of trust, then enhancing sender credibility would have been especially relevant.12
In summary, the ability to anticipate counterarguments enables four basic kinds of argumentative strategies. The counterarguments may be pro-actively communicated in two-sided messages that allow the sender to either (i) rebut the counterarguments or (ii) minimize their importance. Alternatively, the sender can keep the counterarguments private at the outset and either (iii) spend additional time thinking of a good response in case the receiver raises them or (iv) avoid opening gambits that are likely to lead to the receiver raising them. Thus, if the function of argument-production is persuasion, prima
12 Interestingly, O’Keefe (1999, p. 230) found increased judgments of sender credibility for both refutational (r = .11) and non-refutational (r = .08) two-sided messages.
30 facie we should expect that in argumentative contexts people will be good at anticipating counterarguments to their views.
Instead, Mercier and Sperber claim their theory implies that people should “have trouble finding counterarguments to their favorite theories” (2017, p. 218). Although they never explicitly explain why this is, their rationale seems to be as follows. For senders, on average it was beneficial to offload the task of formulating counterarguments to the receiver and instead use the deliverances of the reason module, which specializes in producing myside arguments, to respond to those counterarguments. This offloading allowed senders to conserve resources that were used more productively for other tasks.
It also cost the sender little because “failing to convince one’s interlocutor right away carries little cost” (2017, p. 322 f.).
However, the cost-benefit analysis provided by Sperber and Mercier rests on an implausible picture of the relevant ancestral environments. Estimates of when syntactical spoken language emerged vary widely, from (very) roughly 500ky with archaic humans such as Homo antecessor or Homo heidelbergensis (de Boer, 2017; Fitch, 2017) to ca. 80ky with anatomically modern Homo sapiens (Berwick & Chomsky, 2016). For present purposes, it suffices to note that during this time span, and up until the Neolithic revolution ca. 12ky, humans were organized into nomadic band societies. A wide range of paleontological, archaeological, and ethnographic evidence suggests that bands had a fluid membership ranging from a handful to a maximum of around 50 people who camped together, divided labor, and shared food (Hill et al., 2011; Layton, O’Hara, & Bilsborough,
31 2012). In turn, each band belonged to a widely dispersed clan within which individuals could move freely, join a different band, or form a new band. Simulation models suggest that for a clan to be self-sustaining over multiple generations it needs to number at least
150, while for it to be socially cohesive in the absence of any hierarchical political authority it needs to number at most 500 (Dunbar, 2016; White, 2017).
As Mercier and Sperber (2017, p. 227) recognize, failing to initially state a strong case often carries a significant cost for the sender in terms of reputation. Senders who blurted out weak or easily objectionable arguments would have developed a reputation for incompetence. Given the small size of bands, in which every individual knew every other individual, having such a reputation may have seriously hampered future opportunities to persuade others.
Furthermore, in a wide variety of situations the iterated nature of the argumentative exchange would have been highly limited. Ethnographic studies consistently show that hunting groups keep verbal communication to a bare minimum so as not to alert prey (Hindley, 2015). More generally, any group that is away from the security of their base camp – whether they are hunting, scavenging, foraging for water or plants, or journeying to another camp – need to avoid making noises that will alert predators. At pivotal moments decisive leadership may be called for, as when the group is attacking or being attacked. Moreover, even in scenarios that offer both time and safety for a group, such as when a large number of people congregate around a camp fire, the iterative nature of the exchange will be limited by the participant’s energy and willingness
32 to engage. Tempers may flare, especially when there are disputes within a camp. In band societies, such disputes are often resolved not by argumentation but rather by people joining a different band for a period of time, until tempers cool (Lee, 1979, p. 372;
Turnbull, 1965, p. 106; Woodburn, 1982, p. 435).
Finally, even if there is time and energy for an iterative exchange between interlocutors, the responses from the receiver may be evasive. In competitive scenarios involving bargaining or arbitration – forms of communication which Sperber claims contributed most to the selective pressure for epistemic vigilance – the receiver may have an incentive not to “show their hand” and provide feedback that will aid the sender (cf.
Mercier & Sperber, 2017, p. 322). Furthermore, persuasion can be achieved through means other than argumentation. A sender’s argument may be met not with a counterargument but rather with ridicule, ad hominem attacks, browbeating, and so forth.
This is not to deny that there were ever iterated exchanges in which counterarguments were offered and responded to, leading to rational persuasion.
However, it appears to be a rare occurrence in contemporary hunter-gatherer societies, and there are no independent grounds to hold that in ancestral environments this kind of argumentation would have occurred sufficiently often to render myside bias adaptive.
Although it is tempting to imagine that life in small-scale band societies is inherently more communal than life in agrarian or industrialized societies, it should be borne in mind that contemporary hunter-gatherers spend much of their time alone. For example, to make foraging more efficient, band members typically hunt alone for hours or days, and form a
33 hunting party only if they spot suitable prey and require help tracking or killing it
(Marlowe, 2010, p. 227). Our evolutionary ancestors would thus likely have had ample time to think alone and, in particular, to anticipate objections to what they planned to say at camp.
For these reasons, I submit that the cost-benefit analysis provided by Sperber and
Mercier is empirically unwarranted. However, even if it is granted that it was adaptive for senders to offload the task of producing counterarguments to receivers, I will now show that the resulting explanation of myside bias faces two theoretical problems relating to information-gain and deception.
3.2 Information-Gain
Thus far, we have been considering the issue of myside bias from the perspective of a sender. But we can also look at it from the vantage point of a receiver. Suppose the primary adaptive function of argument evaluation is to decrease acceptance of costly mis- information (DEMANDING) and increase acceptance of beneficial information (OBJECTIVE).
This function of reasoning requires that argument evaluation be accurate. However, if receivers display myside bias in responding to senders then they will tend to pile up what they take to be confirming evidence for myside arguments and disconfirming evidence for otherside arguments. This problem is especially pressing because the psychology of reasoning literature on “disconfirmation bias” suggests that this kind of phenomenon – producing more reasons to favor than to disfavor myside arguments, and more reasons to disfavor than to favor otherside arguments – gives rise to a tendency to overrate the
34 strength of myside arguments and underrate the strength of otherside arguments
(Edwards & Smith, 1996).
Nevertheless, Mercier and Sperber claim that under their theory we should expect that when presented with an otherside argument, people tend to be good at formulating objections to it and bad at formulating defenses of it (Mercier & Sperber, 2017, p. 218).
They suggest that people intuitively evaluate arguments accurately, but subsequently search for further reasons to reject otherside arguments. If such reasons are found, this leads to a final all-things-considered rating of argument strength that is lower than their initial intuitive evaluation (Mercier, 2017, p. 109f.). The hypothesis that there is a two- stage process of evaluation is an empirical one, and ought to be empirically tested.
However, note that postulating a two-stage process is irrelevant with respect to
DEMANDINGNESS and OBJECTIVITY, since it the final all-things-considered judgment that determines whether the argument is actually accepted and hence whether the information bottleneck is eased.
Mercier and Sperber might respond that, for receivers, on average it was beneficial to offload the task of providing support for otherside arguments to the sender; and since the sender’s reason module was disposed to provide rebuttals to objections raised by the receiver, over time the receiver’s final evaluation moved from underrating the strength of
35 the otherside argument to accurately evaluating it.13 It might be wondered why receivers would outsource argumentative labor to senders who they believed may be misinformed or trying to deceive them. However, setting this issue aside, the main problem with the proposal is that it would lead to what I will call an adversarial spiral. Suppose I receive an otherside argument and am disposed to produce objections to it. In turn, my interlocutor will be disposed to either produce objections to my objections, or to produce new arguments for the original conclusion. In the former case, I will be disposed to object to their objections; in the latter case, I will be disposed to object to the new arguments.
And so on.
Sometimes argumentation does involve this kind of dynamic. Think of debates concerning identity politics, or, indeed, the kind of morally saturated and emotionally charged issues that participants are typically asked to think about in myside bias experiments. Notoriously, such debates tend to lead not to agreement but rather to polarization (Lord, Ross, & Lepper, 1979) and in extreme cases even to conflict (Kennedy
& Pronin, 2008). If this is how reasoned persuasion works in general, then it is misleading to claim that people are “ready to revise our ideas when presented with good reasons” and
“accept even challenging reasons” (Mercier & Sperber, 2017, pp. 332; 235). If argument-
13 This cost-benefit rationale is similar to the one discussed in section 3.1 and is subject to the same kinds of problems concerning the iterative nature of argumentation. However, these problems are perhaps less pressing for receivers.
36 evaluation evolved so that receivers are persuaded when and only when the sender simply runs out of things to say, then reasoning is spectacularly inefficient at overcoming the informational bottleneck.
Typically a receiver is in a better position than the sender to know whether the argument being offered coheres with their beliefs. Hence there may be supporting arguments that the sender is not in a position to make. The accurate evaluation of arguments is better served by receivers who are able to think of both the pros and cons of otherside arguments, rather than by receivers who operate with a disposition to defend their view alongside a blindspot resulting from myside bias. As with cases in which senders are aiming to persuade, it may not always be wise for vigilant receivers to voluntarily communicate every argument they think of. For example, perhaps they should refrain from explaining why they think the sender’s argument is good and instead have the sender focus on responding to their misgivings. Nevertheless, the ability to think of both myside and otherside arguments can break the adversarial spiral and make argumentation more efficient.
3.3 Deception
According to the rationalization-first view, argument-production evolved to facilitate not just honest communication but also deceptive communication. Indeed, this is a central assumption that underwrites Mercier and Sperber’s account of how reasoning evolved.
The benefit of persuasion for the sender is producing some effect in the receiver that serves the sender’s interests. Hence, to benefit from communication senders should (ceteris
37 paribus) “communicate the information most likely to produce the intended effect in the addressee, regardless of whether it is true or false” (Sperber et al., 2010, p. 360). An important aspect of the persuasion-vigilance “arms-race” was that the benefits of deception generated selective pressure for an ability to efficiently produce arguments in favor of views senders publicly endorsed but privately held to be unjustified or false. Yet,
Mercier and Sperber claim that under their theory we should expect people to be bad at finding arguments for views we don’t hold:
Finding arguments for a position we do not support, or even one we disagree with, is difficult. It takes skills and training. We may be lawyers, but only when it comes to defending beliefs and decisions we actually endorse. (Mercier & Sperber, 2017, p. 219f.)
Mercier and Sperber thus seem to face a dilemma: either it is not difficult to produce arguments with which to deceive people, which seems to run counter to their explanation of myside bias, or it is difficult to produce arguments with which to deceive people, which seems to run counter to a core component of their account of how reasoning evolved.
On Mercier and Sperber’s behalf, it might be suggested that although the reason module evolved to take genuine beliefs as inputs it can facilitate deception through a kind of pretense in which it is “tricked” into treating states of make-belief as genuine beliefs. In this vein Mercier and Sperber do allow that the tendency to rationalize one’s views can sometimes be overcome:
individuals may develop some limited ability to distance themselves from their own opinion, to consider alternatives and thereby become more objective … But
38 this is an acquired skill and involves exercising some imperfect control over a natural disposition that spontaneously pulls in a different direction. (Mercier & Sperber, 2011, p. 72)
However, regardless of how plausible this kind of model may be, invoking it only pushes the problem back a level: if it were correct then experimental participants who are tasked with thinking on both sides of an issue, or with providing arguments in favor of alternative views, should be able to use “make-believe” to spark the reason module into producing otherside arguments. The importance of deception in the evolution of reasoning, then, provides further grounds to expect that under the rationalization-first view people should be good at producing otherside arguments.
4. From Persuasion to Justification
According to the rationalization-first view, reasoning evolved under pressure from multiple factors including not just persuasion but also self-justification. Traditionally, persuasion has been treated as the main factor. Thus, it is natural to ask whether it is plausible that self-justification was the main factor and, if so, whether revising the rationalization-first view in this way can help it better explain myside bias.
The idea that self-justification often drives reasoning was proposed by Haidt (2001,
2012a) in his highly influential “social intuitionist” model of moral judgment. According to this model, reasoning is typically a post hoc practice of producing arguments that justify the sender’s initial intuitive judgments, so as to achieve social goals such as making a good impression and having smooth social interactions. Likewise, Mercier and Sperber (2017,
39 p. 8) now hold that the reason module evolved not just to provide myside arguments but also to provide what might be called “myside explanations” that favor the sender:
our goal is not to give an objective sociological or psychological account of our actions and interactions; it is to achieve beneficial coordination by protecting and enhancing our reputation and influencing the reputation of others … The explanatory use of reasons, we suggest, is in the service of its justificatory use… (Mercier & Sperber, 2017, p. 186)
Just as the adaptive function of the production of reasons qua arguments is persuasion rather than gaining the truth, under the rationalization-first view of reasoning the adaptive function of the production of reasons qua explanations is self-justification rather than gaining an accurate understanding.
No account of how this supposed trait developed has been provided in the literature.
Perhaps the relevant evolutionary history might be fleshed out as follows. The ability to construct and keep track of reputations facilitates mutual understanding and hence cooperation. Clearly, it is often detrimental for a person to have a bad reputation or an insufficiently established reputation. Moreover, reputations can be inaccurate. It can be especially detrimental for an observer to think that a “bad actor” is in fact dependable and benevolent towards them. In particular, the bad actor – through words or deeds – may be actively deceiving the observer about why they behaved as they did. With the rise of verbal communication, people developed an ability to overcome mistrust by providing explanations of their behavior. In turn, it was in the interests of observers to develop an
40 ability to accurately evaluate these verbally communicated explanations. This allowed observers to trust people they previously had a negative view of.
Assume for the sake of argument that such a module did develop. Recall that, according to Mercier and Sperber, if the function of reasoning is “to justify one’s actions or to convince others” then it “should display myside bias”. Presumably, the idea is that the person sending the explanations should be good at finding explanations they think justify their views and actions, but bad at finding explanations that they think undermine their views and actions. More generally, they should be good at finding explanations that undermine their rivals, and bad at finding explanations that justify their rivals.
Prima facie, the trigger conditions of a module dedicated to this kind of self- justification better correspond to the context of experimental reasoning tasks than the trigger conditions of a module dedicated to persuasion. In particular, myside bias paradigms typically ask participants to explain why they reached the conclusion they did, and participants are given no indication that their conclusion disagrees with the views of the researcher. It is more plausible that participants interpret the situation as one in which they should provide an explanation of their reasoning that maintains their reputation as rational and knowledgeable, than as one in which they should persuade the researcher to share their view.
Nevertheless, for an explanation to function as an effective self-justification the sender must persuade the receiver that the explanation is correct. Persuading a receiver need not involve sending an argument that the explanation is correct. Sometimes, the
41 mere act of communicating an explanation that coheres with the beliefs of the receiver will be enough to persuade them. Thus, a hunter who has spent a week with a different band might explain to their partner that they were tracking a herd of giraffes in the area. If this is a common and widely accepted practice, then their partner may simply accept the explanation.
However, this cannot be the kind of case which led the ability to reason to emerge and persist in the population. Under the rationalization-first view, the ability to produce explanations evolved not for accuracy, to help a receiver achieve understanding, but rather for self-justification, to overcome a trust barrier with the receiver. Moreover, the sender may in fact be deceiving the receiver. If these are the adaptive problems underpinning the evolution of the ability to reason, then in general senders should expect the receiver to have thought of alternative explanations. For example, the hunter’s partner might suspect that they visited the camp to keep in touch with a potential mate. In that case, the objections I developed in the section 3 will arise again, with “explanations” being substituted in for “arguments”. If the ability to reason evolved for senders to provide explanations that justify their views and actions, then it would be surprising if people are relatively bad at anticipating alternative explanations that undermine their reputation.
Furthermore, if an alternative explanation is cited by the receiver, then we should expect the exchange to lead to argumentation. In our previous example, the hunter might attempt to persuade their partner by arguing that the explanation is correct because no one had caught an animal for weeks and they need meat, or the hunter may argue that
42 the alternative explanation is incorrect because the potential mate is in a relationship with someone in the other camp. Thus, the objections in section 3 will return in their original form. The partner will be disposed to pile up objections to the hunter’s argument; in turn, the hunter will be disposed to pile up considerations in favor of their view; and so on. The functions of reason are better served if both parties do not suffer from myside bias.
5. The Inquiry-First View
In the evolutionary sciences, the standard way to make a case that a trait is an adaptation is to show that the trait exhibits hallmarks of “special design”. Although there is debate about the evidentiary standards this requires, there is widespread agreement that the trait should perform the relevant function with sufficient precision, efficiency, and economy to reasonably rule out chance and rival theories of how the trait evolved
(Andrews et al., 2002; Williams, 1966). However, due to genetic and environmental constraints, it is likely that no adaptation conforms to an optimal design for its function
(Dawkins, 2006). The human eye, for instance, has a blindspot because by historical accident the nerves of photocells protrude outwards towards the light and thus have to connect to the brain by looping back through a gap in the retina. Natural selection is a tinkerer, not an engineer.
It might be objected that all I have shown so far is that myside bias is not optimal for the kinds of functions posited by the rationalization-first view; but perhaps the ability to reason was initially based on a genetic mutation that linked, say, a language system
43 and a belief system, in such a way that only myside arguments were produced, and this sufficed for persuasion and self-justification. In itself, this is a fair point. However, recall that the empirical case for the rationalization-first view rests on the claim that it can explain myside bias whereas the inquiry-first view cannot. In the previous two sections my aim was to show that the rationalization-first explanation relies on empirically implausible auxiliary hypotheses, and that myside bias undermines other roles that reason is supposed to have under the view. In this section, my aim is to further argue that if the inquiry-first view is granted the same kinds of auxiliary hypotheses and the same leeway concerning non-optimality, then it can explain myside bias at least as well as the rationalization-first view.
In examining what a plausible inquiry-first view might look like, we should clearly distinguish between the adaptive function of reasoning and the context reasoning evolved to operate in. A theory according to which reasoning evolved for individuals (not groups) to attain epistemic goals may postulate either that this was mainly achieved through group deliberation, or that it was mainly achieved through solitary ratiocination.
Consider first a social version of the inquiry-first view. According to the rationalization-first view, the threat of being misinformed – mainly through deception – generated selective pressure for epistemic vigilance. However, perhaps epistemic vigilance was driven by the threat of being misinformed mainly through incompetence. Whereas deception is typically beneficial to the sender but costly to the receiver, undetected incompetence is typically costly to the receiver and the sender. Under the rationalization-
44 first view, the fitness of senders was increased because senders used arguments to manipulate receivers, for the attainment of socially-strategic goals. To achieve these goals the relevant argument merely had to be persuasive, regardless of whether it had a true conclusion. However, if reason evolved for the attainment of epistemic goals such as gaining true beliefs and avoiding false beliefs, then the production of arguments increased the fitness of senders not through persuading receivers tout court, but rather through persuading receivers when and only when the conclusion of the relevant argument was true.
Imagine a hunter claiming to their band that giraffes will be found to the north.
The band members are skeptical, so the hunter argues that there are plentiful trees to the north. In response, the band members point out that there are also trees to the east. The hunter then argues that there is also plentiful water to the north, but not to the east.
However, the band members object that the water is likely to attract lions, which the giraffes will want to avoid. That the mechanisms which underpin reasoning evolved a myside bias under these kind of conditions can be explained by exactly the same cost- benefit assumptions made by proponents of the rationalization-first view. Group deliberation enables a division of cognitive labor, in which senders offload the work of producing otherside arguments to the receivers. This outsourcing of labor to receivers makes more sense if there is a negligible risk that they are trying to deceive the sender.
Moreover, if reasoning evolved to attain epistemic goals, then there is no need to suppose that reasoners will be good at thinking up otherside arguments for the purposes of
45 deception; for example, that the hunter-gatherer should be able to readily produce arguments which persuade the band members that the giraffes are north, while privately believing that they are east.
However, the problem remains that accurate argument evaluation does not seem to be well served by myside bias. Moreover, the cost-benefit assumptions are subject to the same objections I outlined in the previous section. In band societies, the frequency of iterated argumentative exchanges would likely have been highly limited, and blurting out ill-considered claims or arguments would likely have had significant reputational consequences. If the risk of incompetence drove the evolution of reasoning, it is more plausible that people would have developed the ability to think well for themselves without relying on others.
Consider, then, an asocial version of the inquiry-first view in which the adaptive function of reasoning is the attainment of epistemic goals through solitary ratiocination.
Clearly, such a theory will avoid objections relating to the need to persuade others, whether honestly or deceptively. It may also avoid objections relating to the need to be good at evaluating arguments that receivers present verbally to senders. For why suppose that typically people would consciously “evaluate” their own arguments at all, let alone in a way that would transfer to the arguments that others communicate to them? It is factors such as the fluency of the private reasoning process which determine the degree of confidence the reasoner experiences with respect to any conclusion that is drawn (cf.
Ackerman & Thompson, 2017).
46 An asocial inquiry-first view might hold that reason evolved for the individual to gain true beliefs, but not to test for and correct false beliefs. For example, reason might be a mechanism that exploits cultural knowledge that has been transmitted by trusted members of the community, to come to novel conclusions when faced with situations for which the reasoner has no intuitions. In that case, there are no grounds to suppose that the reasoner should be good at formulating arguments that run counter to their beliefs.
However, even if we build into the asocial inquiry-first view the hypothesis that reason evolved for the individual to test for and correct false beliefs, myside bias can still be plausible explained. Clearly, an asocial theory of reason view cannot introduce the kinds of cost-benefit assumptions that accompany social versions of the rationalization- first and inquiry-first views. However, it can introduce similar cost-benefit assumptions.
First, myside bias might partially reflect an efficient stopping rule for deliberation. As with all psychological processes, reasoning must stop at some point, and when it does should be a function of the perceived importance of the issue and how satisfactory the conclusion arrived at is deemed to be. In his pioneering studies of myside bias, Perkins suggested that participants use a “makes sense epistemology” such that they judge the conclusion true in virtue of it making (often superficial) sense in the light of their beliefs:
Such a thinker’s reasoning is dominated by a strategy of cognitive load minimization … the first model we deliberately generate that makes sense often serves perfectly well. When it does not, and we are dealing with a situation in practical terms, we quickly discover that failing through experience. (Perkins, Allen, & Hafner, 1983, p. 187)
47 This makes sense epistemology concerns inquiry in the sense of discovery. It is a cost- effective means of initial belief fixation.
Myside bias might also serve to efficiently test the truth of preexisting beliefs. Given that the reasoner has such a belief, it will be cognitively demanding to think in terms of opposing hypotheses. A less demanding method for testing the truth of the belief is simply to try to find supporting arguments. Indeed, Koriat et al. (1980) found that when participants were asked to consider two alternatives by writing down reasons for and against each alternative, there was in both cases a bias towards arguments for (the researchers termed this a bias towards “positive evidence”). If such a search does not easily return compelling arguments, then the belief should be doubted (cf. Nestler, 2010;
Schwarz, 2004). At that point, inquiry should revert from testing to discovery, with the aim of coming to a new belief.
Note that this is not merely the thesis that people evolved to be cognitive misers
(Stanovich, West, & Toplak, 2016), including when they reason. The cognitive miser, when faced with a task, tends to save energy by defaulting to a cognitive mechanism with low-computational expense. For example, rather than engage their reasoning system, the cognitive miser might passively accept information when listening to their favorite talk show host, or rely on their intuitions when faced with a problem. However, the cognitive miser has the ability to engage cognitive mechanisms of high-computational expense when they are sufficiently motivated to do so. Mercier and Sperber point out that people display myside bias even when they are motivated to come to a balanced conclusion. Thus,
48 they suggest that the reasoning mechanism itself evolved a myside bias because on average the resources needed to produce otherside arguments was better used for other tasks. My suggestion is that a proponent of the asocial version of the inquiry-first view can make the same move, since it is more efficient to test a belief by checking whether there are good grounds to hold it, than to try to formulate arguments against it. Just as myside bias is not optimal for persuasion or information-gain, neither is it optimal for testing one’s beliefs. However, myside bias is satisficing for this proposed function of reason.
6. Concluding Remarks
I have argued that, contra Sperber and Mercier, the existence of myside bias fails to adjudicate between the rationalization-first view and the inquiry-first view. To explain the existence of myside bias, the rationalization-first view must invoke auxiliary hypotheses that are based on a cost-benefit analysis which is rendered implausible by paleontological, archaeological, and ethnographic evidence concerning human evolution. When granted the same or very similar cost-benefit assumptions, both social and asocial versions of the inquiry-first view can generate explanations of myside bias that are at least as plausible.
However, the rationalization-first and inquiry-first views are not the only options.
For example, it may be that the sort of reasoning at issue is an evolutionary by-product of adaptations for comprehension, learning, planning, problem solving, and the like. While these kinds of views may make predictions based on how the relevant exaptation is spelled out, they make no predictions based on the adaptive function of reasoning since they deny
49 that reasoning has an adaptive function. Furthermore, an inquiry-first view has yet to be worked out in the kind of detail that Mercier and Sperber provide for the rationalization- first view. I have some thoughts on how such an inquiry-first view of reasoning might go; but that is work for another day.
50 Chapter 2 Cracking the Enigma: Cultural Knowledge, Social Learning, and Private Reason
1. Introduction
In recent years, Dan Sperber and Hugo Mercier have formulated an important challenge for theories of human reason. On the one hand, reason seems to be a domain-general and multi-purpose “superpower” that sets our species intellectually and technologically far apart from the rest of the animal kingdom. Based on this idea, they claim, there is a long- standing tradition which takes the pursuit of truth and knowledge by lone individuals to be the modus operandi of reason:
Most of the philosophers and psychologists we talked to … see reason as a means to improve individual cognition and arrive on one’s own at better beliefs and decisions. (Mercier & Sperber, 2017, p. 330)
Call this an “inquiry-first” view of reason, since the attainment of epistemic goals involves attempting to ascertain the facts about the issue at hand.
On the other hand, scientific studies suggest that people display seemingly irrational biases when reasoning. For example, research on “myside bias” shows that
51 people are better at producing arguments for their views than against their views
(Perkins, 1989). As a result, they may become overconfident and reluctant to give up unwarranted beliefs (Koriat et al., 1980). This is problematic for inquiry-first views because if reason evolved for the attainment of epistemic goals such as gaining true beliefs and avoiding false beliefs, it appears to be ill-adapted to its primary function. Hence we are faced with the “enigma of reason” – reason is both a superpower and badly flawed.
Sperber and Mercier aim to crack the enigma by offering a new account of the nature of reason (what reason is) and the function of reason (what reason is for). First, they propose that the ability to reason is based on a garden-variety cognitive module – i.e. a domain-specific mechanism individuated by its evolutionary etiology – dedicated to processing representations of the form < P is a reason for Q >.14 Second, they propose that this “reason module” is triggered by the detection of disagreements or situations in which reputations are at stake, and evolved for the attainment of socially-strategic goals through the restricted purposes of public acts of persuasion and self-justification. That reason displays a myside bias can then be explained as a feature of the module that allows it to rapidly build a case in the reasoner’s favor. Hence, “rather than being a towering superpower”, reason “fits quite naturally among other human cognitive capacities and, despite apparent evidence to the contrary, is well adapted to its true function” (Mercier &
14 I use angled brackets to denote representations. Mercier and Sperber (2017, p. 81) characterize representations by their information-bearing functional role rather than their format or location.
52 Sperber, 2017, p. 5). Call this a “rationalization-first” view of reason, since acts of persuasion and self-justification are attempts to cite reasons in favor of a pre-existing belief or decision.
My aim in this chapter is to develop a new version of the inquiry-first view which cracks the enigma of reason while capturing the traditional idea of reason as a domain- general and general-purpose superpower that pursues truth. According to my account, the primary adaptive function of reason is to generate true beliefs through private deliberation. More precisely, I claim that reason evolved to take rule-like cultural knowledge gained through social learning, and use it to generate solutions to novel problems for which the reasoner has no intuitions. I argue that in comparison to rationalization-first views, my account better fits how human societies functioned in ancestral environments and better captures the empirical characteristics of reasoning.
Here’s how I proceed. I begin (section 2) by reconstructing Mercier and Sperber’s theory, according to which the capacity to reason evolved to overcome an information bottleneck in verbal communication that mainly resulted from the threat that senders of messages were trying to deceive receivers. In response, I appeal to an important kind of evidence that is strangely absent in Mercier and Sperber’s work, viz., paleontological, archaeological, and ethnographic evidence concerning the conditions that existed in ancestral environments. I argue that given the kinds of small-scale, egalitarian societies that existed during the era in which verbal communication evolved, selective pressure for a module dedicated to persuasion and self-justification didn’t exist because the risk of
53 deception was negligible. I further argue that a module adapted for the kind of public argumentative reasoning required for persuasion and self-justification is fundamentally ill-suited to give rise to the kind of private inquisitive reasoning required for reaching a previously unconsidered conclusion. This is significant, I claim, because what underwrites reason’s status as a domain-general and multi-purpose superpower is the ability to apply cultural knowledge to novel problems concerning which we have no intuitions. It is this ability that enables science and engineering, for example, as well as a host of other distinctly human achievements.
Next, I develop a new inquiry-first account of the nature and function of reason
(section 3). Based on ethnologies of hunter-gatherer societies, I provide evidence that the biological fitness of individuals was enhanced by the capacity to use cultural knowledge attained through social learning mechanisms, to understand the environment and guide decision making. I argue that the reasoning system is triggered by the detection of novel problems for which the individual has no intuitions, and takes as input cultural knowledge in the form of rules. The function of the reasoning system is then to generate true beliefs that allow the reasoner to better understand and solve the problems they are faced with.
Since the relevant rules can be about any subject matter, reason is domain-general. And since the relevant problems can concern any issue, reason is multi-purpose. Thus, I provide a naturalistically acceptable grounding for the traditional idea of reason as a superpower. Furthermore, my account recaptures argumentative reasoning, in the sense
54 that persuading an audience and defending one’s reputation are problems that may trigger inquisitive reasoning.
In section 4, I sketch an empirical case for my account. I mimic the strategy used by Mercier and Sperber, which appeals to the evolutionary etiology and adaptive functions of reason to explain the kinds of biases that reason displays. Under my account, although reason evolved to generate true beliefs about novel problems, it did not evolve for the testing and correction of mistaken beliefs. Moreover, since under my account reason is responsive to anything deemed a problem, it should be a highly flexible system that may be put to work in the service of a wide variety of motives, including motives to rationalize one’s beliefs. Thus, I claim that my account is neutral with respect to the existence of myside bias. However, I argue that my account better explains the continued impact of discredited beliefs (belief perseverance). For, contra Sperber and Mercier, belief perseverance is not driven by the generation of arguments that support the belief (myside bias), but rather by the generation of explanations based on the belief. This is what we should expect if reason evolved not for persuasion, but rather to use cultural knowledge to generate true beliefs which help the reasoner solve problems in the natural environment.
2. The Rationalization-First View of Reason
2.1 Exposition of the Theory
Sperber and Mercier’s account of reason is based on a theoretical reconstruction of the evolutionary history of verbal communication. Communication allows animals to share
55 information and manipulate the behaviour of others. For a communication system to be evolutionarily stable in a species, on average the benefits of both sending and receiving messages must outweigh their costs (Dawkins & Krebs, 1978). By sharing information, individuals can facilitate cooperation in the pursuit of complementary goals. However, communication is vulnerable to the spread of misinformation. This occurs when a sender is mistaken or purposely misleads the receiver.
Sperber and colleagues propose that during the early stages of the evolution of verbal communication, the risk of receivers being misinformed by senders generated selective pressure for cognitive mechanisms dedicated to “epistemic vigilance” (Sperber,
2001a; Sperber et al., 2010). Epistemic vigilance has two components. First, the receiver may evaluate the trustworthiness of the sender, based on cues relating to competence and honesty. Second, the receiver may evaluate the plausibility of the communicated claim, based on how the content coheres internally and with respect to a relevant subset of the receiver’s pre-existing beliefs.
Here’s how reason enters the picture (Mercier & Sperber, 2011; Sperber, 2001a).
To the extent that epistemic vigilance blocked the transmission of misinformation, it was costly to senders who aimed to deceive receivers. Moreover, to the extent that epistemic vigilance blocked the transmission of accurate information, it was costly to both senders and receivers. Due to this information bottleneck, senders were subject to selective pressure for the ability to manipulate epistemic vigilance by pre-emptively presenting
56 arguments for their claims, using premises the receiver either already accepted or accepted more readily than the claim in isolation.
The introduction of arguments to verbal communication stimulated the flow of information. However, it also enabled senders to misinform receivers in a more sophisticated way, through specious arguments. As a result, receivers were subject to selective pressure for the ability to accurately evaluate arguments. In turn, this led to a sharpening of the sender’s ability to produce persuasive arguments. And so on. Thus, the two abilities – to produce and evaluate arguments – are mutually adapted.
These argumentative practices are held to be implemented by an evolved cognitive module (Mercier & Sperber, 2009). Modules are autonomous computational devices that automatically perform specific tasks on a restricted range of domain-specific inputs
(Sperber, 1994, 2001b). Although people are not conscious of how modules process information, they are conscious of the outputs of some modules. Furthermore, a conscious output may also be intuitive in the sense that it is accompanied by a feeling of confidence that it is correct, despite the person being unaware of why they take it to be correct.
However, due to the human capacity for meta-representation, people also have intuitions about intuitions, including intuitions about the considerations that justify their intuitions. For Sperber and Mercier, reasoning is the meta-representational process of drawing a conclusion by consciously attending to (purported) reasons in favor of the conclusion. Suppose a sender has an intuitive belief that Q, and wishes to persuade a skeptical receiver that Q is true. This situation triggers the reason module to search for
57 evidence that supports the message the sender wants to communicate (2017, 289). If successful, it will output an intuition that P is a reason for Q , which can then be used as an argument for Q. If either the sender or the receiver accepts Q on the basis of the argument, they will have a “reflective” or “reasoned” belief that Q rather than an intuitive belief that Q, since they will be aware of why they take Q to be true.
In recent work, Mercier and Sperber (2017, 130) propose that the reason module also underwrites the ability to justify ourselves. The rationale is that any argument for the correctness of a belief or action can be used as a justification for holding the belief or performing the action, and vice versa. Having a good reputation increases biological fitness because it facilitates cooperation, especially with non-kin in the pursuit of long-term goals.
Just as the adaptive function of producing arguments is persuasion rather than inquiry, the adaptive function of producing explanations is not accuracy but rather self- justification.
In summary, here’s what Mercier and Sperber tell us about how the hypothesized reason module operates. We have a passing sense of the inputs, outputs, and trigger conditions. According to Mercier and Sperber (2017, 280), the proper domain of reasoning is “disagreements between oneself and others”. Thus, the reason module is triggered by the “detection of a clash of ideas”. Furthermore, the module is also triggered by the detection of situations in which reputations are at stake (2017, 123). Representations of the form < P is a reason for Q > can be verbally communicated as arguments to persuade or as explanations to justify oneself. Once activated, the module either takes < Q > as an
58 input and outputs < P is a reason for Q > as an intuition, or takes < P is a reason for Q > as an input and outputs an evaluation of it as an intuition (2017, 148). The capacity to reason, then, derives from a module that outputs intuitions in the same manner as other modules that make up our animal minds, except that the reason module outputs intuitions about representations. Moreover, since this module is “a cognitive mechanism aimed at justifying oneself and convincing others” (2017, 10), it will display a myside bias. Thus, the enigma is solved. Or so it would appear.
2.2 Criticism 1: The Paleo-Anthropological Context
As we have seen, Sperber and Mercier’s reconstruction of the environment of evolutionary adaptiveness for reason is based on an informal game-theoretical model of how verbal communication developed. According to the model, the major factor that drove the evolution of reason was not the risk that senders were mistaken, but rather the risk that senders were engaging in deception:
the major problem posed by communicated information has to do not with the competence of others, but with their interests and their honesty. While the interests of others often overlap with our own, they rarely coincide with ours exactly. In a variety of situations, their interests are best served by misleading or deceiving us. It is because of the risk of deception that epistemic vigilance may be not merely advantageous but indispensable if communication itself is to remain advantageous. (Sperber et al., 2010, p. 360)
The model does not presuppose any particular account of how the capacity for language originally emerged and persisted in the hominid line. For example, it is compatible with
59 the theory that language initially evolved to enable complex inferential transitions in inner thought (Berwick & Chomsky, 2016; Fitch, 2011) or community bonding through stories, songs, jokes, and the like (Dunbar, 2017). However, it does presuppose that at some point in the evolution of language the verbal exchange of information about the world was sufficiently fitness-enhancing that epistemic vigilance and argumentation became adaptive.
Sperber and Mercier do not provide an analysis of the kinds of interactions which supposedly gave rise to the relevant selection pressures, beyond suggesting that bargaining and arbitration are prime examples:
Communication among people, in particular close kin, on matters where their interests are aligned does not require epistemic vigilance towards the risk of deception in order to be advantageous to all participants. On the other hand, several typical forms of human communication such as bargaining or arbitration could hardly be practiced in a manner advantageous to all in the long run without vigilance. (Sperber, 2013, p. 63)
However, as I will show, this proposal faces two major problems in the light of our best theories about how human societies functioned in ancestral environments. First, practices such as bargaining and arbitration are unlikely to have given rise to an informational bottleneck that became an adaptive problem. And, second, even if such an adaptive problem did arise, it is more plausible that it was solved by the development of moral norms and the threat of punishment.
60 Paleontological and genetic evidence suggests that our species (Homo sapiens) emerged from Homo erectus, or descendants of Homo erectus such as Homo antecessor and Homo heidelbergensis, in Africa, by at least 300–200ky (Brown et al., 2012; Richter et al., 2017). Populations of Homo sapiens gradually expanded beyond Africa, with the main dispersal occurring at 70–60ky and Australia being reached by boat around 50–40ky at the latest (Nielsen et al., 2017). Estimates of when syntactical spoken language emerged vary widely, from 500ky with Homo antecessor or Homo heidelbergensis (de
Boer, 2017; Fitch, 2017) to 80ky with anatomically modern Homo sapiens (Berwick &
Chomsky, 2016). During this period of the Paleolithic era, before the 12ky Neolithic revolution which introduced agriculture and permanent settlements, humans were organized into nomadic band societies (Grove, Pearce, & Dunbar, 2012).
A wide range of paleontological and ethnographic evidence suggests that bands had a fluid membership ranging from a handful to a maximum of around 50 people who camped together, divided labor, and shared food (Hill et al., 2011). In turn, each band belonged to a widely dispersed clan within which individuals could move freely, join a different band, or form a new band. Indeed, bands may have evolved within pre-existing clans as crystallized “fission-fusion parties” of the kind seen within chimpanzee societies, with the entire range of the community accessible within a day’s walk (Layton et al., 2012).
The fluidity of band membership within the clan allowed individuals to buffer local resource scarcity and find spouses. Simulation models suggest that for a clan to be self-
61 sustaining over multiple generations it needs to number at least 150, while for it to be socially cohesive it needs to number at most 500 (Dunbar, 2016; White, 2017).
Ethnographic studies of 21st century hunter-gatherer populations report that they are egalitarian (Woodburn, 1982). Since resources are not stored in nomadic societies, opportunities for greed and theft rarely arise. Individual hunts are usually unsuccessful, so whenever meat is procured it is shared within the band. Gift exchanges are also used to establish and maintain relationships which may be critical for survival in the future
(Winterhalder, 1986). Movement between bands, which is commonplace and encouraged, serves as a mechanism for avoiding and resolving conflicts. For example, often individuals will simply announce where they are going to hunt or set up camp, with others then being free to make their own decision about whether to join or not (Woodburn, 1982, p. 444).
This freedom of movement helps to even out access to resources within the clan’s territory.
When disputes do break out within a band, typically one of the disputants will simply move to a different band until tempers calm or the issue loses importance (Lee, 1979, p.
372; Turnbull, 1965, p. 106; Woodburn, 1982, p. 435).
If deception is harmful to receivers within the band, then the sender runs the risk of being ostracized from the group. In Paleolithic hunter-gatherer societies, ostracism is likely to have been fatal (Marshall, 1976, pp. 196; 287). This is not to say that deception never occurred. However, many commonplace forms of deception benefit the receiver. For example, a sender might produce arguments to reassure a receiver that they are in the
62 right during a dispute, that they are a good hunter, and so forth. In these kinds of cases involving white lies, it harms the receiver to exercise vigilance.
Might long-distance trade have given rise to epistemic vigilance? Paleolithic trade networks are typically inferred from lithic site-to-source distances. For example, although obsidian has been used for flaked stone-tool manufacture for at least 1.7 million years, evidence for long-distance transfer of obsidian and other raw materials does not become widespread until 50ky (Ambrose, 2012). Assuming that the human ability to reason was in place before the main dispersal out of Africa 70–60ky, this counts against the hypothesis that the ability to reason was driven by long-distance trade. However, recently a 200ky site in Kenya was found to contain obsidian from three different sources located 25 km,
140 km and 166 km away (Blegen, 2017). A modern band typically ranges over a 20km radius annually, and foraging zones are estimated to have a 40km radius at maximum
(Brooks et al., 2018; Gamble, 1993). This tentatively suggests that during the early history of anatomically modern Homos sapiens there were long-distance trade networks.
However, economic transactions are likely to have been infrequent, and materials such as obsidian, pigments, shells, beads, clothes, and so forth, carry their value on their face. Whereas a trader might know that an animal comes from a sickly stock, or that a complicated machine has a flaw, there is nothing to conceal about simple material objects.
Of course, in any exchange promises of future payment may be subject to doubt. But it is unclear why the exchange of material objects – prior to the rise of more sophisticated
Neolithic economies based on herding livestock, plant agriculture, and land ownership –
63 would often have involved deferred payments. Thus, there is little reason to suppose that
Paleolithic trade gave rise to doubts concerning the accuracy of communicated messages, to an extent that significantly impacted fitness.
At first glance, interactions that provided opportunities for self-justification are easier to imagine. For example, foragers might creatively explain away unsuccessful hunts, or boastfully recount the decisions that led to successful hunts. Having a good reputation is likely to have been fitness enhancing in ancestral environments.
Ethnographic studies of band societies, in which people have few personal possessions, have found that character traits such as honesty and generosity are less important than traits such as being a hard worker and a good forager (Smith & Apicella, 2019). Men and women with better reputations as foragers have more friends and are more preferred as camp co-residents than people with poorer reputations (Marlowe, 2010, p. 251). Moreover, the trait that women most value in a potential mate is that of being a good forager, and men who have better hunting reputations have more friends, younger wives, and more children, than men with poorer reputations (Marlowe, 2010, p. 215). Since food is shared in band societies, a good hunting reputation may primarily function as a sign of health and intelligence, and hence overall mate quality.
Nevertheless, within the clan everyone knows each other and bands can communicate relatively easily. Hence, reputations would likely have been well grounded in personal experience and difficult to manipulate. For example, in the ethnographic studies cited above, men with better hunting reputations had higher overall food returns
64 and higher hourly return rates than men with poorer reputations (Marlowe, 2010, p.
215).15 Experiments have shown significant positive correlations between hunting reputation and skills required for hunting such as aim, upper-body strength, and the ability to recognize animal vocalizations (Stibbard-Hawkes, Attenborough, & Marlowe,
2018). Moreover, often a man who is interested in a woman as a potential mate is required to live with her band for a period, and if they do not prove themselves as a hunter they will not be accepted (Marshall, 1959). Within small band societies, experience speaks for itself. Thus, despite the likely importance of reputation in ancestral environments, there is no evidence that self-justification would have had a significant impact on fitness.
In summary, given our best theories of the kinds of conditions that held in ancestral environments during the evolution of verbal communication, there was no selection pressure for the evolution of a reason module dedicated to persuasion and self- justification. The risk of deception may have been more pressing in the Neolithic world of permanent settlements and agriculture, than in the Paleolithic world of nomadic bands and foraging. However, the Neolithic period began around 12ky, which is long after the main exodus out of Africa, and is in any case too recent to allow sufficient time for the development of a sophisticated pan-cultural cognitive capacity such as reason.
15 However, hunting success is not easy to estimate. For example, Hill and Kintigh (2009) calculate that if hunters acquire meat on < 27% of hunting days, a sample of > 600 hunting days is required to estimate a hunter’s mean daily acquisition rate with ± 20% accuracy.
65 2.3 Criticism 2: The Scope of Reason
I aim to have shown that there is little empirical warrant for supposing that a module evolved for public practices of persuasion and self-justification. But even granting that such a module exists, it is unclear how it alone can account for the wide range of private reasoning practices that humans seem to engage in.
Mercier and Sperber (2017, p. 168) suggest that the reason module came to be triggered by the presence of conflicting intuitions within an individual, as well as mentally simulated disagreements with an imagined audience. As they point out, it is commonly allowed that the trigger conditions of modules only imperfectly capture their original target, and sometimes mismatches will be advantageous. For example, a module that evolved for the detection of snakes may become biased towards false positives, such as being triggered by the sight of an exposed tree root. In a similar fashion, the benefits of solitary argumentative reasoning may have biased the reason module towards being triggered by doxastic disagreements even in the absence of an actual interlocutor.
More problematic is whether the reason module can allow for inquisitive reasoning, where there are no initial intuitions to rationalize. Mercier and Sperber appear to treat such cases as a kind of problem-solving which involve the application of step-by-step methods, rather than reasoning as such:
This ability to understand and apply step-by-step methods is a hugely important aspect of human psychology. It plays a major role in the development and transmission of cultural skills, including specialized problem-solving skills. The psychologist Keith Stanovich has argued that the ease with which people acquire
66 and apply such methods (or “mindware”) correlates with individual differences in rationality. Still, reasoning doesn’t consist in applying such methods, and in general, it doesn’t need them. (Mercier & Sperber, 2017, p. 171)
According to their definition, reasoning tout court consists in “attending to reasons for adopting new conclusions” (Mercier & Sperber, 2017, p. 52). But note that “new conclusion” is ambiguous. The reason module is a mechanism for backwards (abductive) inference from an intuition that Q to an intuitive argument for Q. If the argument is judged to be good, then Q may be accepted on those grounds, in which case it will be held as a newly formed reflective belief (Mercier & Sperber, 2009). However, no account has been provided of how this mechanism can be exapted for forwards inference, in the absence of any intuition Q, to a previously unconsidered conclusion.16
This is important because, prima facie, it is inquisitive reasoning, rather than argumentative reasoning, which is responsible for the “superpower” that gives rise to the enigma of reason. There is a growing recognition that what sets the human species so far apart from other species is our heightened capacity to learn not just from our observations and interactions with the environment, but also from each other (Heyes, 2018; Tomasello,
2001). This “social learning” enables the creation and transmission of culture, and often gives rise to a “ratchet effect” such that successive generations improve on the work of previous generations, resulting in artifacts that are too sophisticated to be invented whole-
16 A similar point is made by Evans (2019); see also Sterelny (2018, p. 5f.).
67 cloth by any single community (Tennie, Call, & Tomasello, 2009). Arguably, we are now totally reliant on cultural knowledge for our survival:
During eons of relying on large bodies of cumulative cultural knowledge, our species became addicted to this cultural input … Neither our intelligence nor domain-specific psychological abilities fire up to distinguish edible from toxic plants or to construct watercraft, bone awls, snow houses, canoes, fishhooks, or wooden sledges. (Henrich, 2015, p. 33)
Social learning allows the individual to go beyond associative inferences based on personal experience, by reasoning to novel conclusions based on a consideration of the normative implications of cultural knowledge that has been gleaned from others. This process was supercharged by the development of writing systems around 6-5ky, which allowed humans to transmit vast amounts of information and led to religion, business, politics, and science as we know them today (Bottero, Herrenschmidt, Vernant, & Zabbal, 2000).
This role of reasoning dovetails well with conceptions of reasoning that are influential in recent philosophical discussion – what might be called “intentional rule- following” theories of reasoning (Boghossian, 2014). Such theories make a pair of commitments. First, they suppose that reasoning essentially involves following rules concerning the premises from which one reasons. In addition, they suppose that all rule- following involves (conscious or unconscious) intentional states which represent the rules being followed. The generality of rules explains why inferences often exhibit similar patterns across cases. Moreover, since the rules are represented, we can combine them and even use the rules themselves as premises. Finally, that rules are followed helps
68 explain the sense in which reasoning is a person-level activity that is subject to social and rational norms, since the reasoner can err by either failing to correctly follow a rule or by correctly following an inappropriate rule.
Although this is not the place to defend intentional rule-following theories of reasoning, it is worth noting that they can naturally model the application of cultural knowledge to problems for which we have no intuitions: the premises refer to the particular situation, and the rules are supplied by cultural knowledge gained through social learning. Thus, the conception of inquisitive reasoning outlined above has independent merit. In the following section, I provide an account of how the capacity for this kind of reasoning might have evolved.
3. A New Inquiry-First View of Reason
My suggestion is that reasoning evolved to put cultural knowledge to use. Since the kind of reasoning we are interested in involves attending to explicitly represented, verbalizable, reasons to draw a conclusion, prima facie we should expect the reasoning system to primarily draw on rule-like cultural knowledge that is linguistically transmitted. If my account of the function of reason is correct, then we should expect to see this kind of knowledge transmission in the educational practices of contemporary hunter-gatherers.
However, perhaps surprisingly, ethnologies of modern hunter-gatherer societies tend to conclude that there is scant evidence for overt teaching (Boyette & Hewlett, 2018;
Garfield, Garfield, & Hewlett, 2016; Lew-Levy, Reckin, Lavi, Cristóbal-Azkarate, & Ellis-
69 Davies, 2017; MacDonald, 2007). Instead, these studies find that children and adolescents mainly learn through self-guided observation and play. Technical skills such as digging roots, food preparation, tool manufacture, and weapon use, are gained through imitation.
These skills are honed through role-playing games, in which animal or adult human behavior is mimicked. For example, a study of the !Kung (Ju/’hoan) hunter-gatherers of
East Africa reports that ecological knowledge and tracking skills are gained with “little formal instruction” from adults:
Play hunting can begin as early as age 3 … Little girls participate in these games too, and frequently a play group of four to six kids pretends to cut up, cook, and serve an imaginary animal in the same way that North American children have an imaginary tea party … tracking skills, however, are acquired through the on- going study of nature as the young boys and girls learn to identify the hundreds of plants and animals in their environment. Studying animal tracks is a major pastime of the older boys. (Lee, 1979, p. 236)
Older children are brought on foraging expeditions by their parents, and while verbal instructions are sometimes given during these trips, they tend to be limited and brief
(Hewlett, 2016). Typically, parents will simply use gestures to direct their children. For example, a hunter may point to a track, and perhaps state something about the animal that made it, but will not explain how that conclusion was arrived at (MacDonald, 2007).
Moreover, although theoreticians commonly suppose that cooperative hunting requires verbal planning and commands, actual strategies used on the ground – such as hunters spreading out into a large circle and then progressively tightening it, while making loud
70 noises to direct prey towards the center – tend to be intuitive and can be coordinated through gesture (Coon, 1971; Endicott, 1988, p. 123).
Clearly, hunter-gatherer technologies and practices have likely changed significantly since language began to evolve. For instance, although there is evidence for hunting spears at 400ky (Dennell, 1997; Thieme, 1997), the earliest known evidence for a bow and arrow is at just 71ky (Brown et al., 2012). Until quite recently in our evolutionary history, Homo sapiens have relied on stone tools, and these may have required a greater degree of coordination and cooperation to manufacture and wield. Indeed, it has been suggested that the reliance of humans on lithic technologies created selection pressure for the evolution of language as a means to transmit the relevant manufacturing skills and knowledge (Laland, 2018, p. 200). Might manufacturing stone tools have required overt teaching?
The earliest known lithic technologies associated with the Homo genus – Oldowan cores and flakes – date to 2.6my and were used as axes, hammers, knives, and scrapers
(Leakey, Tobias, & Napier, 1964; Semaw et al., 2003). These tools were often produced using the freehand percussion method, in which a stone core is held in one hand and struck with a stone hammer. A major advance in sophistication came with the introduction of
Acheulian bifacial hand axes, cleavers, and picks, the earliest known examples of which date to 1.7my (Diez-Martín et al., 2015). These were much larger than Oldowan tools and featured neatly worked symmetrical sides, which resulted in sharp edges and points.
Nevertheless, Oldowan and Acheulian techniques are alike in treating flakes as a useful
71 by-product of shaping the core. The next technological advance came with the introduction of Levallois prepared core techniques, in which the core itself is treated as a means to an end, being carefully worked in multiple stages so that a thin and sharp flake of predetermined size and shape can eventually be detached from it. These prepared core techniques led to a decline in handaxes and their replacement by points and blades which could be used to make composite tools such as spears. Importantly, the emergence of
Levallois tools coincides with the emergence of archaic Homo sapiens, with the earliest
Homo sapiens fossils being found alongside both Acheulian and Levallois tools (Richter et al., 2017; Tryon, McBrearty, & Texier, 2005).
A number of experimental studies have focused on whether language, or at least some form of proto-language, is necessary to transmit the skills and knowledge required for the manufacturing of stone tools. Ohnuma et al. (1997) and Putt et al. (2014) investigated the production of Levallois flakes and Acheulean flakes respectively, and found no statistically significant differences between gestural instruction and verbal instruction. Morgan et al. (2015) and Lombao et al. (2017) investigated the production of
Oldowan and Acheulean bifaces, and found that both gestural instruction and verbal instruction were superior to mere imitation and emulation. However, although they also reported that verbal instruction was superior to gestural instruction, these differences did not reach statistical significance. Finally, noting that previous studies included gesture in their verbal instruction conditions, Cataldo et al. (2018) investigated the production of
Oldowan and found that speech alone is inferior to gesture alone and to gesture-plus-
72 speech. In summary, all five studies found that stone tools can be manufactured without verbal instruction, and any facilitation of performance through verbal instruction is dependent on gesture.
Although lithic technologies do not appear to require language, they do require a host of other cognitive capacities that are highly relevant to inquisitive reasoning. For example, knapping requires the flexible use of strategies held in long-term memory, which guide action routines and subroutines (Alperson-Afil, Goren-Inbar, Herzlinger, & Wynn,
2020; Putt, Wijeakumar, Franciscus, & Spencer, 2017). Prepared core techniques in particular require the knapper to form a plan, carefully monitor the core while keeping in mind the intended form of the flake, and make knapping decisions that will have implications several steps ahead in the process. Thus, it has been argued that Levallois tool manufacture resulted in selective pressure for an increase in working memory capacity and the development of a long-term working memory system, both of which underlie the capacity for modern day expert cognition and performance (Wynn & Coolidge,
2011).
Like a knapper, a reasoner may work towards a goal by retrieving from long-term memory strategies which guide step-by-step procedures for achieving sub-goals (Lee &
Johnson-Laird, 2013; Newton & Roberts, 2005; Schaeken et al., 1999). To take a simple example, suppose your goal is to find the sum of the digits from 1 to 100. Here we have no intuitions, and must rely on the arithmetical knowledge we were taught in school. One strategy is to add 1 and 2 (a sub-goal), store the answer in working memory, then add 3 to
73 the previous answer (another sub-goal), store that answer in working memory, and so on.
An alternative strategy is to break the numbers up into 50 pairs which each add to 101 – i.e. (1 + 100), (2 + 99), (3 + 98), (4 + 97), and so on – and then multiply 50 by 101, perhaps by multiplying 50 by 100 and then adding 50. Or to take a more practical example, consider the choice method known as “elimination by aspects” (Tversky, 1972). Suppose you are choosing which car to purchase. You select the characteristic of cars that is most important to you, such as automatic transmission, and eliminate all cars which do not possess this characteristic. Next you select a further characteristic, such as a maximum price of $30k, and eliminate all cars which do not possess that characteristic. And so on, until one car is left. Both everyday and expert reasoning make use of a myriad of such consciously deployed strategies, which can be regarded as involving sequences of mental actions performed to solve practical and theoretical problems.
However, this leaves unresolved the issue of how Homo sapiens came to possess the ability to arrive at conclusions by attending to the normative implications of linguistically transmitted cultural knowledge. Thankfully, all is not lost for my account of reason. For although there is no evidence for overt teaching in hunter-gatherer societies, either past or present, previous studies have overlooked the pedagogical significance of stories (Scalise Sugiyama, 2017). For example, a study of the Chabu forager-farmers of
Ethiopia found that adolescents learned ecological and hunting knowledge by listening to stories told to them by their fathers (Dira & Hewlett, 2016). Likewise, Lee reports that
!Kung adolescents learn about hunting from stories:
74 Before they actually go on a hunt, !Kung boys have listened to dozens of hunts described in the minutest detail in the storytelling around the campfire. This is a major component of their socialization as hunters. This vast body of knowledge is a treasure house of lore and information about animals and how to kill them. And the boys listen intently. (Lee, 1979, p. 236)
This is corroborated by Marshall (1976, p. 130), who notes that the !Kung “recount over and over memorable episodes of past hunts”.
Wiessner (2014) collected conversations among the !Kung and found that the overwhelming majority of stories were told at night around a campfire, with 81% of all night talk being dedicated to stories. The earliest evidence for regular controlled use of fire dates to 400–300ky (Sandgathe & Berna, 2017), which is just prior to the earliest evidence for Homo sapiens. The use of fire for cooking reduced the need for chewing and digestion, resulting in the evolution of smaller teeth, guts, and intestines (Wrangham & Conklin-
Brittain, 2003). Eating cooked food also reduced the risk of bacterial infection and increased net energy gain, which facilitated brain expansion (Carmody & Wrangham,
2009). Fire is a source of heat, which facilitated the movement of humans into colder regions outside of Africa. Fire is also a source of light, which extends the number of active hours of the day. However, since fire does not provide enough light for work, most of this extra time is taken up with social activities, such as singing and acting. Although gesture- based communication across a flickering fire may have led to the development of artistic performances, it would have been inadequate for sustained sharing of information. Thus,
75 it has been suggested that verbal communication evolved to enable group conversation around the campfire (Dunbar, 2014).
Since stories are often a source of fun and entertainment, it is easy to overlook their impact on biological fitness. First, as we have seen, stories transmit ecological and technological knowledge. Of course, the relevant knowledge need not be restricted to hunting. Consider the following example, taken from the !Kung stories transcribed by
Wiessner (2014).17 In this tale about a trip to a neighboring camp, a recipe is described:
the people fetched water, fetched water and said: “Why don’t you people look for look for the eland chyme, why don’t you look for it so you can cook it into porridge?” Early the next morning when the sun had not yet risen, they fetched water, fetched water, brought water. Some put the pots on the fire and while others prepared the chyme. They put eland fat into small tortoise shells, put it in there, and poured hot water into the shell to melt the fat. (Wiessner, 2014, supplementary material)
Stories also transmit knowledge about social norms and practices. Wiessner (2014) found that most !Kung stories focused on marriage and kinship. Similarly, Smith et al (2017) collected stories across seven different hunter-gatherer societies and found that 70% pertained to prescribing social norms and coordinating behavioral expectations (39% dealt with natural phenomena, while 34% dealt with resource use). A familiar theme of such
17 The examples cited in this section are selected from the very small sample of stories that have been translated and published. Although they may not be the best examples of knowledge transmission, it is important to remain wedded to actual practices when possible.
76 stories is that people who are not cooperative and peaceful suffer negative consequences, including being ejected from the community. For example, in the following story a woman leaves her poorly behaved husband for another man:
After she kicked him out she went away to the west. She went and married a man called Ku 'oan who was short and nicely built and who could kill an eland now and then. He would kill an eland now and then, and he would get a bag of ostrich eggshells and bring them to his wife. She went and choose that man herself. She, /Xoan, choose him all by herself and married him. When she married him her parents immediately agreed: “Yes”. People were living on the fat eland meat and Ku... did not beat or mistreat his wife. (Smith et al., 2017, supplementary material)
Knowledge of social norms and practices promotes interpersonal cohesion within the band.
Moreover, since stories often feature people in other bands, they may strengthen social bonds within the extended clan by creating a “virtual community” that exists in the imagination.
This transmission of knowledge through the medium of language supplies the materials for reasoning. For example, consider the following report from a !Kung camp:
In discussing fighting of kudu, one man described the sound one hears of their horns clashing, and how, if one hears the sound, one can approach to shoot them. The same man, a very enthusiastic and busy hunter also described how one tracks infant kudu, showing that it sleeps away from its mother, and that one can follow it to where it is hidden, and kill it by hitting it. (Jones & Konner, 1976, p. 340)
77 This is not information about a particular kudu, but rather rule-like information about all kudu. Such rule-like information can be consciously deployed as a heuristic to guide action.
Suppose you discover the tracks of an infant kudu while foraging, and wonder whether they are worth following. You may recall the elder telling you that infant kudu sleep away from their mothers. This knowledge, which takes the form of a generic, functions as a default rule, and you conclude that this kudu will sleep away from its mother. Hence, you decide that you should follow the tracks.
I agree with Mercier and Sperber that reasoning involves attending to reasons for drawing a conclusion. Thus it is worth noting that the ability to create and understand stories, which interpret the world in a reflective way, facilitates an awareness of something being a reason. For example, in a tale of moving to another camp, one man shares that he and his host acted as they did because they were scared:
He put our things down, because we were afraid of each other. He was afraid of me and I was afraid of him. (Wiessner, 2014, supplementary material)
Stories explain why people act as they do. Understanding that a mental state can be a reason for bodily action provides some basis for eventually understanding that a belief can be a normative reason to perform the mental action of accepting a conclusion.
Stories transmit knowledge indirectly, through narrative explanations that entertain the audience. Smith et al (2017) found that skilled storytellers were almost twice as likely as unskilled storytellers to be chosen as campmates, and had an additional 0.53 living offspring compared to non-skilled storytellers. However, in discussions with the
78 !Kung, Jones and Konner (1976) found that storytellers did not take license with the facts and readily admitted their ignorance. Since stories are shared with campmates, including close kin, in general it is not in the storyteller’s interests to mislead the audience. Thus, there is no need for the audience to be especially skeptical about the stories.
I propose that the ability to critically evaluate evidence may instead have arisen to test one’s own hypotheses about the environment. For example, tracks are sometimes ambiguous or unclear, and the hunter must formulate a hypothesis about their meaning and continually weigh the evidence as the tracking progresses (Liebenberg, 2012, pp. 71;
102). This process is illustrated by the following !Kung hunt:
They began following a gemsbok spoor which, the man said, was made the same morning. After about twenty minutes the man stopped and said, “No, it was made last night,” and abandoned the spoor. Asked what made him change his mind, he indicated a single gemsbok hoofprint with a mouse track inside it, that is, super-imposed on it. Since mice are nocturnal, the gemsbok print must have been left during the night. (Jones & Konner, 1976, p. 342)
However, non-human animals are also responsive to cues in the environment. In general, we should expect a wide range of perceptual and cognitive processes to involve the assessment of hypotheses. The weighing of evidence is mainly intuitive rather than reflective, with the exceptions being the rare cases which happen to bring to mind linguistically transmitted knowledge.
79 4. Empirical Characteristics of Reason
In the previous section, I have illustrated how the biological fitness of individuals in ancestral environments is likely to have been enhanced by cultural knowledge transmitted through stories. Reasoning, I claim, has a key role in this process, by taking rule-like inputs and delivering true beliefs about problems in the natural environment.
Thus, my account is a version of what I call the “inquiry-first view” of reasoning.
Mercier and Sperber rule out inquiry-first views on the grounds that people tend to be poor at correcting mistaken intuitions:
Reason rarely questions reasoners’ intuitions, making it very unlikely that it would correct any misguided intuitions they might have. This is pretty much the exact opposite of what you should expect of a mechanism that aims at improving one’s beliefs through solitary ratiocination. (Mercier & Sperber, 2017, p. 218)
For example, De Neys et al. (2017) found that, across 12 reasoning experiments in which participants gave an initial intuitive response followed by a final considered response, further deliberation typically did not help the participant find the correct answer:
• In 15% to 42% of trials participants gave the correct response initially and finally. • In 48% to 76% of trials participants gave an incorrect response initially and finally. • In 7% to 10% of trials participants initially gave an incorrect response and then a correct final response.
De Neys and colleagues thus conclude that the main role of reasoning in these experiments is to construct post-hoc justifications of intuitive responses.
80 Mercier and Sperber interpret this as one manifestation of a wider phenomenon of
“myside bias”. In the classic myside bias paradigm, participants are asked to think about a controversial issue and come to a conclusion if possible. When explaining the reasoning for their conclusion, participants tend to provide more arguments for their view than arguments against it (Perkins, 1989). This effect holds even when participants are instructed at the outset to produce arguments on both sides of the issue (Stanovich &
West, 2008). Thus, it appears that people find it easier or more natural to search for evidence that supports rather than undermines their view. According to Mercier and
Sperber, myside bias is predicted by rationalization-first views since it helps the reasoner build a case to persuade others or justify themselves, but is in tension with the epistemic goals posited by inquiry-first views.
However, note that inquiry can improve beliefs in two ways: it can test information to avoid or weed out false beliefs, or it can use information to generate or bolster true beliefs. Experimental reasoning tasks are often set up so that participants are cued to have a mistaken intuitive response. However, my view does not predict that people will be naturally inclined to test their intuitions. Under my version of the inquiry-first view, the function of reasoning is not to test information but rather to generate true beliefs from cultural knowledge when faced with novel situations for which we have no intuitions. My view instead predicts that people will tend to recognize that evidence undermines their beliefs, when presented with such evidence. Studies on reasoning universally agree that
81 when participants are presented with the correct answer (e.g. during debriefing) they tend to admit that it is the correct answer.18
Of course, if participants do wish to check their intuitive beliefs, they can use task- relevant knowledge to come a reasoned conclusion, which can then be compared with their intuitive belief. However, participants often do not possess task-relevant knowledge.
Indeed, participants who have expertise – training in logic, for example – are typically excluded from studies. Given that experimental tasks require a response, it is unsurprising that such participants would tend to rationalize their intuitions as best they can.
Under my view, the trigger for reasoning is the detection of a novel problem for which we have no intuitions. Thus, my account predicts that reason should be a flexible system that can be exapted to serve a wide variety of motives and goals. We might say that although the adaptive function of reason is to apply cultural knowledge to generate true beliefs, reasoning as such is neither a scientist nor a lawyer but rather a handy-man for hire.19 For example, the task of defending one’s reputation, and the task of persuading others, are problems that might initiate reasoning. In this way, my view of reason is more
18 An example: in Wason’s famous four-card selection task, very few participants correctly select the two cards that can falsify the given rule. However, when participants are instead asked to turn over the cards and state which of them falsify the rule, nearly everyone gets the answer correct (Wason, 1966, p. 146). 19 I owe this metaphor to Richard Samuels.
82 general than that of rationalization-first views. Moreover, reasoning is often motivated not by a desire to ascertain the truth about an issue, or to justify oneself, or to persuade others, but rather by a desire to defend cherished beliefs or attack unsettling propositions.
Thus, my account captures what is commonly referred to as “motivated reasoning”
(Kunda, 1990). Some of these motivations may have even been adaptive in ancestral environments. However, there is a crucial distinction to be made between reason evolving for such goals, and evolved motives using reason – and perhaps other cognitive systems as well – to attain such goals.
Finally, Mercier and Sperber claim that their account predicts the tendency for beliefs to continue to influence thought and behavior after the reasoner learns they are unwarranted. This “belief perseverance” effect, they suggest, results from reason operating outside of its natural social environment, so that it piles up supporting arguments which favor the relevant belief, without any pushback from an interlocutor:
Participants were asked to distinguish between real and fake suicide notes, and were told how well they’d done. They were then left to think about their performance for a little while. During this time, they thought of many reasons why the feedback made sense: they had always been very sensitive, they knew someone who was depressed, and so on. Then the participants were told that in fact the feedback had been completely bogus, bearing no relationship whatsoever with their actual performance. But it was too late. Participants had found many reasons to buttress their beliefs, so that even when the initial support was removed, the beliefs stood on their own. (Mercier & Sperber, 2017, p. 242)
83 Hence, even though participants are told that the belief is unwarranted, they will have found their own supporting arguments in the meantime.
As we have seen, my view can allow that we are sometimes motivated to build a one- sided case. However, this is not the commonly accepted cause of the perseverance effect.
Rather, initially the belief is used to generate a causal explanation, so that when the belief is subsequently discredited, the belief will rightly be given up but the causal explanation remains and can subsequently be drawn on (Anderson, Lepper, & Ross, 1980). For example, if participants are initially told that risky people are less/more successful as firefighters, they tend to generate plausible explanations of why this is so, e.g., risky firefighters save more people because they fight fires that others will not. When they are subsequently debriefed that the initial belief has no basis in evidence, they will drop the belief but unwittingly continue to be influenced by the explanation they generated. Thus, the perseverance effect does not result from participants thinking up new supporting arguments, but rather from participants thinking up causal explanations. This fits well with my account of reason, since socially transmitted knowledge is used to understand the world, with reasoning serving to apply it to novel situations rather than search for further evidence.
5. Concluding Remarks
In this paper I have outlined a new naturalistic account of human reason, according to which the function of reason is to generate true beliefs. Private reason is a superpower,
84 but only when it is fueled by cultural knowledge, gained through social learning. I have attempted to sketch the ways in which this account makes sense of biases that reason displays. However, many questions remain. Do some kinds of problems trigger reasoning more readily than others? Are there multiple reasoning systems that might be revealed through dissociations? Much work remains to be done as part of a wider research project to flesh out and empirically evaluate the theory.
85 Chapter 3 Wason Confirmed: Why Confirmation Bias is Not Myside Bias in Disguise
1. Introduction
In the early 1960s the psycholinguist Peter Wason reinvigorated the study of the psychology of reasoning by introducing a novel research project focused on whether people naturally attempt to falsify their hypotheses and, if so, how. To this end, he devised the
“2-4-6 task”, in which participants try to discover a rule by formulating and testing predictions, and the “four-card selection task”, in which participants judge the evidential relevance of potential tests for a given rule. Based on his studies, Wason famously concluded that people have a tendency to search for evidence in ways which unwittingly favor information that supports rather than falsifies their hypotheses (Wason, 1966). This was later termed “confirmation bias” (Mynatt, Doherty, & Tweney, 1977).
86 Issues relating to confirmation bias have been extensively studied in the six decades since Wason’s pioneering work.20 However, the literature has increasingly shifted away from his original interpretation of performance on the 2-4-6 and four-card selection tasks. Building on this trend, recently it has been argued that the notion of confirmation bias should be replaced by the notion of “myside bias”. For instance, in a review of the literature, the cognitive scientist Hugo Mercier writes:
most of the conventional wisdom about the confirmation bias is wrong – starting with its name … there is no such thing as a general tendency to confirm whatever one thinks about, only a tendency to find arguments that support one’s own views – a myside bias. (Mercier, 2017, p. 99f.)
This represents a sea change in how confirmation bias is thought about. The core of the new conception is that rather than searching for evidence in ways which unwittingly favor information that supports whatever proposition they happen to be testing, people unwittingly seek out evidence that favors propositions they believe.
Getting clear about what the bias amounts to may have practical value for attempts to debias reasoning. But the new conception also bears on weighty theoretical issues concerning the nature and evolutionary origins of the human capacity to reason. For
20 The notion of confirmation bias has sometimes been expanded to include phenomena such as “pseudo- diagnosticity” (Doherty, Mynatt, Tweney, & Schiavo, 1979), “selective exposure” (Jonas, Schulz-Hardt, Frey, & Thelen, 2001), and “motivated reasoning” (Hahn & Harris, 2014). In this paper, I restrict attention to the original phenomenon associated with Wason’s studies.
87 example, a number of prominent researchers who take myside bias to be a characteristic feature of reasoning have proposed that the best explanation of this purported fact is that reason evolved for attaining socially strategic goals through persuasion and self- justification (Haidt, 2012a; Mercier & Sperber, 2017).
My goal is to show that, when we sharpen the theoretical issues and the right distinctions are made, the new conception is seen to rely on a serious misunderstanding of the evidence. Wason was largely right back in the 1960s, and some missteps have occurred in the progression of the literature. Notably, there is a persistent failure to distinguish between effects and processes. In this paper I show that there is evidence for both confirmatory effects and myside effects. However, although there is also evidence that a process which searches for confirming evidence tends to operate when people test hypotheses, there is scant evidence that a process which searches for arguments that support one’s view tends to operate when people deliberate. This is significant because proponents of the theory that reasoning evolved for persuasion and self-justification hold that there is a cognitive module which has the adaptive function of producing myside arguments.
Here’s how I proceed. I begin by examining Wason’s tasks. The confirmation bias interpretation of the selection task faces two main objections. The first is that potential falsification is not avoided by participants. The second alleges that participants are not searching for confirming cases, but rather are guided by unconscious linguistic comprehension heuristics which under certain conditions inadvertently direct attention
88 towards confirming cases, with subsequent reasoning serving to rationalize this selectivity
(ala myside bias). In response, I argue that people tend to search for instances of their hypothesis – which, if discovered, confirm the hypothesis – and show one way in which this “positive” search strategy is biased towards confirmation despite the possibility of falsification.
I then turn to the 2-4-6 task. Here the main objection has been that in typical real- world scenarios, the statistical structure of the environment ensures that falsification is more likely when searching for cases that are compatible rather than incompatible with the hypothesis. In response, I show that if attention is restricted to test outcomes that participants deem relevant to the hypothesis, there does exist a confirmation bias effect that is robust across all environments.
Finally, I turn to reason generation tasks, which are often cited as demonstrating the existence of myside bias. In these studies, participants consider a controversial claim
φ and are invited to conclude either φ or not-φ. The classic finding is that when asked to cite their reasons, participants tend to provide more arguments for their view than against their view. I argue that even if this is taken to be a myside bias effect, it can be plausibly explained away as a side-effect of an unbiased reasoning process that led the participant to adopt the view.
89 2. Confirmation Bias
In this section I provide a detailed analysis of the 2-4-6 task (Wason, 1960) and the four- card selection task (Wason, 1966). Although expositions typically present the tasks in chronological order, I use my analysis of the selection task to shed light on the 2-4-6 task.
2.1 The Four Card Selection Task
Suppose you are shown four cards with a revealed face and a hidden face, and are told that each card has a letter on one side and a number on the other. For example, the revealed faces may display A, D, 4, and 7 (Figure 1).
A D 4 7
Figure 1: Example card faces in the four card selection task.
Which cards would you need to turn over to find out whether someone who utters the following statement about these four cards is lying?
“If a card has a vowel on one side, then it has an even number on the other side.”
90 Take a moment to answer before reading on. If you are like most participants you will have selected either card A alone or cards A and 4. However, the correct answer is A and
7 since either of these cards may falsify the statement (Figure 2).21
Wason noted that people tend not to accept that a conditional with a false antecedent is true (Johnson-Laird & Tagart, 1969). Thus, participants are likely to think that if the conditional is true, then there exists at least one positive instance of it – i.e. a card with a vowel and an even number. Since only the A and 4 cards may provide positive instances of the conditional when turned, Wason originally concluded that there is an
“apparent bias towards verification” of the statement (1966, p. 147).
A D 4 7 VOWEL CONSONANT EVEN ODD
ODD EVEN ODD EVEN VOWEL CONSONANT VOWEL CONSONANT
modal3selection falsifying3counter?instance verifying3positive?instance
Figure 2: Possible scenarios in the four card selection task.
21 A and 7 may not be the correct answer relative to how the participant understands the instructions. This feeds into a wider debate about rationality and the value of normative models (Elqayam & Over, 2016). Nevertheless, A and 7 is the correct answer assuming that the cards are to be turned at the same time and that the rule is restricted to the four cards, rather than a pack to which they belong.
91 The “four card selection task” has been used in over 300 published studies, perhaps making it the most researched task in the history of the psychology of reasoning (Evans,
2017). The now standard “abstract” version uses an indicative conditional with abstract content that cites instances rather than categories (e.g. “If a card has A on one side, then it has 4 on the other side”), along with instructions to determine whether the “rule” is “true or false”. By convention, across studies the four cards are denoted True Antecedent (TA),
False Antecedent (FA), True Consequent (TC), and False Consequent (FC).
A recent meta-analysis of the abstract selection task found that 11% of participants made the correct selection, which corresponds to TA and FC (Figure 3). Improved results have sometimes been found in studies that use realistic or thematic content such as “Every time I go to Manchester, I travel by train” (Wason & Shapiro, 1971). However, the results of replication studies are mixed, and facilitation effects may be due to the content cuing memories of past experiences with falsifying cases.22
22 For reviews see Evans et al. (1993) and Dominowski (1995). Performance is typically much improved when deontic rules are used, such as “If you tidy your room, then you may go out to play” (Manktelow & Over, 1991). However, since the task is to ascertain whether people are violating the rule rather than to ascertain whether the rule is true or false, these selection tasks are not directly relevant to the issue of confirmation bias.
92 50
40 45 41 39 30 TA 28 TA,3TC
PARTICIPANTS 20 23 TA,3TC,3FC
OF TA,3FC
%3 10 11 11 0 4 ABSTRACT REALISTIC
Figure 3: Canonical selections made on the four card selection task, from a meta-analysis of 55 studies using abstract material and 44 studies using realistic material. Based on data reported in Ragni et al. (2018), Table 4.
Wason rarely used the term “bias” and it appears that by “bias towards verification” he merely meant a tendency to search for positive instances of the rule. Call this a “positive test strategy”. In contrast, a “negative test strategy” searches for cases that are incompatible with the rule. Wason did not claim that participants desire to show that the rule is true. This would be odd, since the rule is abstract and arbitrarily chosen. Instead, he claimed that participants rely on a positive test strategy to examine whether the rule is true.
Participants who use a positive test strategy are trying to confirm the rule, but only in the weak sense that the strategy involves initiating a search process the goal of which is to identify positive instances which confirm the rule. There is some debate in the wider psychological literature about the extent to which goal activation, goal pursuit, and goal
93 monitoring require conscious awareness and intentional control (Marien, Custers, Hassin,
& Aarts, 2012). For the purposes of this paper, it suffices to note that the goal of the search process figures in a stopping rule for the search and a decision rule for the test strategy. If the goal is attained, the search is terminated and the rule is judged true. If the goal is not attained, the rule may be judged false. I refer to the tendency to use a positive test strategy as “confirmation biasg”, where the subscript abbreviates “goal”. Since confirmation biasg refers to the process that generates the selections, it can be invoked to explain the selections. The modal selection is TA and TC because there is a tendency for participants to search for positive instances. These participants fail to grasp that a positive test strategy filters out the class of potentially falsifying instances associated with turning FC.
Furthermore, they also often fail to grasp that turning TC is redundant, since turning TA will either conclusively verify that at least one positive instance of the rule exists or conclusively falsify the rule.
The term “bias” is also widely used to denote systematic deviation from a normative standard. Here is one natural way to quantify a bias towards confirmation in the selection task. Let the probability of falsification when turning TA and when turning FC be denoted by a and b, respectively. Further, let the probability of finding a confirming positive instance when turning TA and when turning TC be denoted by c and d, respectively. These probabilities are epistemic and do not depend on what the reverse side of the cards actually show, beyond the assumption that every possible combination that conforms to the task instructions has a non-zero chance of occurring. The ratio associated with the logically
94 correct solution (TA, FC) acts as the baseline. For the three most common selections, the ratios of the probability of finding a confirming positive instance to the probability of finding a falsifying case can be expressed as: