Back to Search Start Over

Towards Reproducible LLM Evaluation: Quantifying Uncertainty in LLM Benchmark Scores

Authors :
Blackwell, Robert E.
Barry, Jon
Cohn, Anthony G.
Publication Year :
2024

Abstract

Large language models (LLMs) are stochastic, and not all models give deterministic answers, even when setting temperature to zero with a fixed random seed. However, few benchmark studies attempt to quantify uncertainty, partly due to the time and cost of repeated experiments. We use benchmarks designed for testing LLMs' capacity to reason about cardinal directions to explore the impact of experimental repeats on mean score and prediction interval. We suggest a simple method for cost-effectively quantifying the uncertainty of a benchmark score and make recommendations concerning reproducible LLM evaluation.<br />Comment: 4 pages, 1 figure

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2410.03492
Document Type :
Working Paper