Uncategorized

Dynamic Benchmark on Reasoning Ability of Large Language Models via Complexity Classes



Download a PDF of the paper titled NPHardEval: Dynamic Benchmark on Reasoning Ability of Large Language Models via Complexity Classes, by Lizhou Fan and 4 other authors

Download PDF
HTML (experimental)

Abstract:Complex reasoning ability is one of the most important features of current LLMs, which has also been leveraged to play an integral role in complex decision-making tasks. Therefore, the investigation into the reasoning capabilities of Large Language Models (LLMs) is critical: numerous benchmarks have been established to assess the reasoning abilities of LLMs. However, current benchmarks are inadequate in offering a rigorous evaluation of the full extent of reasoning abilities that LLMs are capable of achieving. They are also prone to the risk of overfitting, as these benchmarks, being publicly accessible and static, allow models to potentially tailor their responses to specific benchmark metrics, thereby inflating their performance. Addressing these limitations, our research introduces a new benchmark, named NPHardEval. This benchmark is designed to evaluate the reasoning abilities of LLMs across a broad spectrum of 900 algorithmic questions, extending up to the NP-Hard complexity class. These questions are meticulously chosen to represent a wide range of complexity class below the NP-hard complexity class, offering a rigorous measure of the reasoning ability of LLMs. Through this study, we shed light on the current state of reasoning in LLMs, providing an objective and rigorous perspective through the comparison of LLMs’ performance across complex classes. Moreover, this benchmark is designed with a dynamic update mechanism, where the datapoints are refreshed on a monthly basis. Such regular updates play a crucial role in mitigating the risk of LLMs overfitting to the benchmark, promoting a more accurate and reliable assessment of their reasoning capabilities. The benchmark dataset and code of NPHardEval are available at this https URL.

Submission history

From: Lizhou Fan [view email]
[v1]
Fri, 22 Dec 2023 18:07:44 UTC (12,839 KB)
[v2]
Mon, 25 Dec 2023 06:56:50 UTC (13,244 KB)
[v3]
Fri, 12 Jan 2024 05:54:13 UTC (19,665 KB)
[v4]
Mon, 12 Feb 2024 17:30:25 UTC (21,528 KB)



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *