LIMI: Less is More for Agency
Overview
Paper Summary
This paper introduces LIMI, demonstrating that strategically curated small datasets (78 samples) can dramatically boost AI agent performance on specific benchmarks (like "vibe coding" and "research workflows"), significantly outperforming models trained on much larger, uncurated datasets. This "Less Is More" principle challenges traditional scaling laws for AI agency, suggesting data quality and curation are more important than sheer volume.
Explain Like I'm Five
Instead of feeding AI agents mountains of data, this research found that carefully choosing just a few really good examples makes them much smarter at solving problems and acting on their own.
Possible Conflicts of Interest
Yes, conflicts of interest are identified. Several authors are affiliated with SII-GAIR and ASI, which appear to be involved in the development of both the LIMI approach and some of the baseline models (e.g., GLM-4.5, GLM-4.5-Air) and the evaluation environment (SII CLI, AgencyBench). This constitutes a conflict as the researchers are evaluating their own products and frameworks.
Identified Limitations
Rating Explanation
The paper presents a compelling and significant finding that challenges the common scaling law paradigm in AI, showing that strategic data curation can yield superior agentic performance with drastically fewer samples. The methodology is well-explained, and the results are quantitatively strong on the chosen benchmarks. However, the inherent conflict of interest due to authors evaluating their own models/frameworks, the specific nature of the 'agency' tasks studied, and the potentially labor-intensive data curation process prevent a top rating.
Good to know
This is the Starter analysis. Paperzilla Pro fact-checks every citation, researches author backgrounds and funding sources, and uses advanced AI reasoning for more thorough insights.
Explore Pro →