Uncategorized

[2401.08273] Large Language Models are Null-Shot Learners



Download a PDF of the paper titled Large Language Models are Null-Shot Learners, by Pittawat Taveekitworachai and 2 other authors

Download PDF
HTML (experimental)

Abstract:This paper presents null-shot prompting. Null-shot prompting exploits hallucination in large language models (LLMs) by instructing LLMs to utilize information from the “Examples” section that never exists within the provided context to perform a task. While reducing hallucination is crucial and non-negligible for daily and critical uses of LLMs, we propose that in the current landscape in which these LLMs still hallucinate, it is possible, in fact, to exploit hallucination to increase performance in performing tasks compared to standard zero-shot prompting. Experiments with eight LLMs show improvements in performance across the majority of eight datasets, including reading comprehension, arithmetic reasoning, and closed-book question answering. The observed inconsistency in increased relative performance across the LLMs also potentially indicates a different degree of inherent hallucination in each model. These differences show that it is possible to utilize null-shot prompting as a way to detect degrees of hallucination in LLMs using existing benchmarking datasets. We also perform ablation studies, including experimenting with a modified version of null-shot prompting that incorporates ideas from zero-shot chain-of-thought prompting, which shows different trends of results.

Submission history

From: Pittawat Taveekitworachai [view email]
[v1]
Tue, 16 Jan 2024 10:53:11 UTC (8,564 KB)
[v2]
Wed, 14 Feb 2024 10:34:48 UTC (8,575 KB)



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *