Log in

Arbitrating norms for reasoning tasks

  • Original Research
  • Published:
Synthese Aims and scope Submit manuscript

Abstract

The psychology of reasoning uses norms to categorize responses to reasoning tasks as correct or incorrect in order to interpret the responses and compare them across reasoning tasks. This raises the arbitration problem: any number of norms can be used to evaluate the responses to any reasoning task and there doesn’t seem to be a principled way to arbitrate among them. Elqayam and Evans have argued that this problem is insoluble, so they call for the psychology of reasoning to dispense with norms entirely. Alternatively, Stupple and Ball have argued that norms must be used, but the arbitration problem should be solved by favouring norms that are sensitive to the context, constraints, and goals of human reasoning. In this paper, I argue that the design of reasoning tasks requires the selection of norms that are indifferent to the factors that influence human responses to the tasks—which aren’t knowable during the task design phase, before the task has been given to human subjects. Moreover, I argue that the arbitration problem is easily dissolved: any well-designed task will contain instructions that implicitly or explicitly specify a single determinate norm, which specifies what would count as a solution to the task—independently of the context, constraints, and goals of human reasoning. Finally, I argue that discouraging the use of these a priori task norms may impair the design of novel reasoning tasks.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Canada)

Instant access to the full article PDF.

Fig. 1

Similar content being viewed by others

Notes

  1. Samuels et al., (2012) make a similar distinction between two things that can be evaluated: the exercises of cognitive capacities and the judgments that result from the exercises of the cognitive capacities. I’ve proposed here that we can evaluate judgments in relation to these things.

  2. My conception of cognitive success is completely different from Schurz & Hertwig’s (2019). They argue that what is rational for the exercise of a cognitive capacity is whatever maximizes the likelihood of success. This is just a consequentialist conception of cognitive norms. My proposal is that task norms are distinct from cognitive norms and that achieving task norms counts as a success for cognition that is distinct from its counting as rational.

  3. Davies et al. (1995) argue that the WST is specifiable in predicate logic, not propositional logic, so the norms of classical predicate logic should be used to categorize responses instead of classical propositional logic. Johnson-Laird & Wason (1970) recognize the same problem but suspect that it doesn’t make a significant empirical difference.

  4. Oaksford & Wakefield (2003) design a modification of the WST that partly addresses the indeterminacy problem: it specifies the probabilities of the antecedent and consequent. However, their modified task fails to address the contradiction problem: their task instructions still ask subjects to evaluate the truth of the conditional, rather than asking them which card would maximize information gain.

  5. To mitigate priming, he could provide them examples of modus ponens and other distractor inferences.

  6. I thank Reviewer 1 for raising this important challenge and for offering Skovgaard-Olsen et al. (2019) as an excellent example of one such task.

  7. For another example, see Elqayam’s (2012) proposed subjectivist modification to Stanovich & West’s (2008) task.

  8. I expect that many philosophers experience the use of such double standards.

  9. Different forms of rational analysis offer different formalizations of these inferences (Anderson, 1990; Russell, 1997; Oaksford & Chater, 2007; Howes et al., 2009, 2016; Lewis et al., 2014; Griffiths et al., 2015; Lieder & Griffiths, 2020).

  10. These slightly modified tasks are often used as controls for the CRT (see De Neys, 2012, 2014 for reviews).

  11. Subjects still fail at high rates if the antecedent and consequent propositions attribute arbitrary properties to the rule and cards (alphanumeric, geometric, color, etc.) (Wason, 1969) or if they attribute meaningful properties (Manktelow & Evans, 1979). But cognitive success can be partially restored if the antecedent and consequent propositions describe social rules in the context of enforcement (Cox & Griggs, 1982; Pollard & Evans, 1987).

  12. In principle, we could say that subjects are deploying their relevant capacity for modus tollens but are somehow robustly failing to rationally exercise it (Wason & Johnson-Laird, 1972). But this seems implausible and uncharitable (Cohen, 1981; Stich, 1990; Schurz & Hertwig, 2019).

  13. For example, they cite Putnam’s (1974) example that perturbations in the orbit of Uranus were known to contradict the prevailing Newtonian model of the solar system, but it wasn’t clear which part of the model was false until Neptune was found. By comparison, perturbations in the orbit of Mercury were also known to contradict the prevailing Newtonian model of the solar system, and it wasn’t clear which part of the model was false until Newton’s theory of gravity was supplanted by Einstein’s theory of relativity (vs. the discovery of the hypothetical planet Vulcan). So, they conclude that it wouldn’t be useful for us to acquire the capacity for falsification or modus tollens.

  14. This explains why robust failures aren’t useful for studying cognitive control: cognitive control has the function to correct errors in the exercise of cognitive capacities. But robust failure is the result of the rational exercise of irrelevant cognitive capacities, so cognitive control shouldn’t recognize the task errors as cognitive errors, and so shouldn’t intervene. Since cognitive failures aren’t failures in cognitive control, there is little information that we can gain from them about cognitive control.

  15. This is an across-task arbitration problem: we have to arbitrate between well-defined tasks, each of which has a single determinate norm. This is distinct from the problem that Elqayam & Evans (2011) raise, which is a within-task arbitration problem: we supposedly have to arbitrate across multiple norms for a single task.

  16. Klauer et al. (2007) and Ragni et al. (2018) both show that Oaksford & Chater’s (1994) model fails to sufficiently explain all 16 possible types of responses to the WST. In a sense, this isn’t surprising: Oaksford & Chater attribute the same reasoning response to every subject, which is a very strong assumption. By comparison, Klauer et al. and Ragni et al. attribute different reasoning responses to different subjects using multinomial processing tree models. This is a much weaker assumption, which significantly increases the flexibility of their models. Still, Oaksford & Chater’s model is simpler (it uses 4 parameters rather than 10), so it will be easier for me to discuss its advantages vis-à-vis the classical logic model. Hence, I’ll focus on their model, even though my point would be the same if I had focused on Klauer et al. and Ragni et al.’s model instead. I thank Reviewer 1 for pressing this point.

  17. Finally, if one has found a task norm that designs tasks that elicit very low rates of robust failure, then we should test whether the task norm corresponds to a cognitive norm by searching for non-normative modifications to the task (which hold the task norm fixed) that restore high rates of robust failure. This can correct for misleading task norms, as in the case where high rates of success can be observed on the WST under trivalent logic even though subjects are maximizing the expected information gains.

  18. For example, I recently developed a novel kind of task design that explicitly requires the use of hard norms (Dewey, 2022). During the review process for that paper, though, I received strong criticism from some reviewers for using hard norms and was directed to Elqayam & Evans (2011). As a philosopher, though, I was confident with insisting on the use of normative assumptions and held my ground, but I worried that non-philosophers might not be so confident. That convinced me to write the current paper, to offer and defend an alternative to subjectivism, soft normativism, and descriptivism.

  19. To their credit, Elqayam & Evans (2011) are sensitive to this. They recognize that the is-ought gap can be closed without creating an is-ought fallacy by adding an “implicit normative premise”, often known as a bridge premise. But they don’t seem to be sensitive to the fact that this undermines their original concern: whenever we draw normative conclusions from empirical premises, it’s always possible to rationalize our inference post hoc by adding an implicit normative bridge premise that validates our argument. For this reason, it seems quite infelicitous to accuse anyone of committing an is-ought fallacy. If we want to disagree with someone’s is-ought inferences, it’s much more felicitous to explicate their implicit normative bridge premise and then argue that the premise is false.

  20. This is just an onus-shifting argument: if someone believes that norms aren’t empirical (in some sense) but other unobservable things are, then the onus is on them to show that there is a consistent, plausible way to do this.

  21. Note that these resources are mostly focused on the norms used for evaluating moral reasoning (a favourite kind of reasoning among philosophers, including myself). However, similar arguments extend to other kinds of reasoning.

References

Download references

Acknowledgements

I thank Reviewer 1 for their extensive comments and invaluable suggestions, which have significantly improved this paper. I also thank Sara Aronowitz and Mike Oaksford for their feedback, which helped shape earlier versions of this paper.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Aliya R. Dewey.

Ethics declarations

Conflict of interest

I have no conflicts of interest to declare.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Dewey, A.R. Arbitrating norms for reasoning tasks. Synthese 200, 502 (2022). https://doi.org/10.1007/s11229-022-03981-8

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11229-022-03981-8

Keywords

Navigation