Uncategorized

[2402.09193] (Ir)rationality and Cognitive Biases in Large Language Models



Download a PDF of the paper titled (Ir)rationality and Cognitive Biases in Large Language Models, by Olivia Macmillan-Scott and Mirco Musolesi

Download PDF

Abstract:Do large language models (LLMs) display rational reasoning? LLMs have been shown to contain human biases due to the data they have been trained on; whether this is reflected in rational reasoning remains less clear. In this paper, we answer this question by evaluating seven language models using tasks from the cognitive psychology literature. We find that, like humans, LLMs display irrationality in these tasks. However, the way this irrationality is displayed does not reflect that shown by humans. When incorrect answers are given by LLMs to these tasks, they are often incorrect in ways that differ from human-like biases. On top of this, the LLMs reveal an additional layer of irrationality in the significant inconsistency of the responses. Aside from the experimental results, this paper seeks to make a methodological contribution by showing how we can assess and compare different capabilities of these types of models, in this case with respect to rational reasoning.

Submission history

From: Olivia Macmillan-Scott [view email]
[v1]
Wed, 14 Feb 2024 14:17:21 UTC (487 KB)
[v2]
Thu, 15 Feb 2024 11:09:09 UTC (487 KB)



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *