Quantifying uncert-Al-nty: Testing the accuracy of LLMs' confidence judgments
Overview
Paper Summary
Across five different tasks, LLMs demonstrated mixed metacognitive accuracy in their confidence judgments, sometimes outperforming and sometimes underperforming humans, but generally performing slightly *better*. A key finding is that several LLMs were less likely than humans to improve their metacognitive calibration after completing a task, suggesting a limitation in learning from experience. Overall, LLM confidence isn't uniformly better or worse than human confidence, varying considerably by the specific model and task.
Explain Like I'm Five
Scientists found that computers are sometimes good at knowing how sure they are about an answer, and sometimes not, but generally a little better than people. However, they don't learn from mistakes to get better at being sure like humans do.
Possible Conflicts of Interest
The authors acknowledge using LLMs for proofreading, but declare no other conflicts of interest. Funding was provided by the National Science Foundation and Carnegie Mellon University.
Identified Limitations
Rating Explanation
This paper explores a timely and relevant topic—the metacognitive abilities of LLMs—using a rigorous experimental approach. Comparing LLMs directly to humans across multiple studies is a strength. While limitations regarding sample size, task variety, and potential confounds exist, the overall methodology is solid and the findings provide valuable insights. The paper is well-written and presents a balanced perspective on LLM capabilities.
Good to know
This is the Starter analysis. Paperzilla Pro fact-checks every citation, researches author backgrounds and funding sources, and uses advanced AI reasoning for more thorough insights.
Explore Pro →