BASE MODELS KNOW HOW TO REASON, THINKING MODELS LEARN WHEN
Overview
Paper Summary
This paper proposes that advanced "thinking" Large Language Models (LLMs) don't acquire new reasoning abilities but primarily learn *when* to activate existing reasoning mechanisms already latent in simpler base models. By applying targeted "steering vectors" to base models, the researchers were able to recover up to 91% of the performance gap to dedicated thinking models on mathematical reasoning tasks, without updating the base model's weights. This suggests that pre-training instills reasoning capacity, and subsequent training teaches strategic deployment rather than fundamental skill acquisition.
Explain Like I'm Five
Your super smart AI isn't learning *new* ways to solve problems; it's just figuring out *when* to use the smart ideas it already has in its brain. We helped simpler AIs act just as smart by giving them tiny, well-timed nudges.
Possible Conflicts of Interest
Two authors, Arthur Conmy and Neel Nanda, are likely affiliated with Anthropic (as inferred from their affiliations in co-authored papers cited in the bibliography, e.g., from 2025). This paper extensively discusses and evaluates various commercial "thinking models," including Anthropic's Claude series, DeepSeek-AI's DeepSeek-R1, Google's Gemini, and OpenAI models. Their potential affiliation with Anthropic, a direct competitor and subject of analysis in the paper, constitutes a conflict of interest.
Identified Limitations
Rating Explanation
This is a strong research paper offering a novel and compelling hypothesis regarding LLM reasoning, backed by a robust methodology and significant empirical evidence. The findings have important implications for the understanding and training of LLMs. While there are acknowledged limitations regarding the LLM-as-a-judge evaluation and some variability for smaller models, these do not undermine the core scientific contribution. The potential conflict of interest is noted but is mitigated by the paper's focus on general mechanistic interpretability rather than direct product comparisons.
Good to know
This is the Starter analysis. Paperzilla Pro fact-checks every citation, researches author backgrounds and funding sources, and uses advanced AI reasoning for more thorough insights.
Explore Pro →