Small score gaps in model evals might just be noise. Learn how to use statistical error bars and rigor to determine if your model is actually better.

The biggest red flag in AI right now isn't a low score—it’s a high score with no error bars. We need to stop treating evals like static scores and start treating them like the scientific experiments they actually are.
Adding Error Bars to Evals: A Statistical Approach to Language Model Evaluations Evan Miller Anthropic evanmiller@anthropic.com Abstract Evaluations are critical for understanding the capabilities of large language models (LLMs). Fundamentally, evaluations are experiments; but the literature on evaluations has largely ignored the literature from other sciences on experiment analysis and planning. This article shows researchers with some training in statistics how to think about and analyze...


Ranking models by tiny margins—such as a 0.5% difference—is often misleading because these fluctuations may simply be statistical noise rather than a reflection of true capability. Evaluation datasets are finite samples pulled from a theoretical "super-population" of all possible questions. Without calculating error bars or standard error, it is impossible to know if a higher score is a significant result or if the ranks would flip if the experiment were run again with different questions or different model seeds.
The Rule of Three is a statistical guideline used when a model passes every single test in a small sample size. If you run 30 safety tests and the model never fails, it is mathematically incorrect to claim the model is 100% safe. Instead, the rule dictates that the 95% confidence upper bound for the failure rate is 3 divided by the number of tests. In a 30-test scenario, you can only say with 95% confidence that the failure rate is below 10% in the wild.
Standard statistical assumptions require that every question in a dataset be independent, but real-world benchmarks often violate this by using multiple questions based on the same document or translating the same prompt into different languages. If a model struggles with the underlying context, it will likely fail all related questions, meaning they are not independent "votes" on performance. Clustered Standard Errors account for this correlation by grouping related items, preventing researchers from underestimating uncertainty and reporting artificially small error bars.
One of the most effective ways to shrink error bars is to use continuous metrics like "logprobs" (log probabilities) instead of binary pass/fail scores. By looking at the probability the model assigned to the correct answer rather than whether it happened to sample that answer, you eliminate "within-question" variance caused by the model's internal randomness. Other strategies include resampling (averaging multiple completions for the same prompt) and averaging results across the final few checkpoints of a training run to smooth out lucky fluctuations in model weights.
Comparing two separate error bars is often too conservative; models can have overlapping confidence intervals and still show a statistically significant difference. A paired difference test evaluates both models on the exact same set of questions and focuses on the gap between their scores. Because models usually agree on which questions are difficult, their scores are positively correlated. Subtracting these correlated variables shrinks the variance of the difference, making the test much more sensitive and capable of detecting real improvements that a naive comparison would miss.
From Columbia University alumni built in San Francisco
"Instead of endless scrolling, I just hit play on BeFreed. It saves me so much time."
"I never knew where to start with nonfiction—BeFreed’s book lists turned into podcasts gave me a clear path."
"Perfect balance between learning and entertainment. Finished ‘Thinking, Fast and Slow’ on my commute this week."
"Crazy how much I learned while walking the dog. BeFreed = small habits → big gains."
"Reading used to feel like a chore. Now it’s just part of my lifestyle."
"Feels effortless compared to reading. I’ve finished 6 books this month already."
"BeFreed turned my guilty doomscrolling into something that feels productive and inspiring."
"BeFreed turned my commute into learning time. 20-min podcasts are perfect for finishing books I never had time for."
"BeFreed replaced my podcast queue. Imagine Spotify for books — that’s it. 🙌"
"It is great for me to learn something from the book without reading it."
"The themed book list podcasts help me connect ideas across authors—like a guided audio journey."
"Makes me feel smarter every time before going to work"
From Columbia University alumni built in San Francisco
