REASONINGBANK: Scaling Agent Self-Evolving with Reasoning Memory
Overview
Paper Summary
This paper introduces REASONINGBANK, a new memory framework that helps AI agents learn from both successful and failed experiences to develop generalizable reasoning strategies. It also proposes memory-aware test-time scaling (MATTS) to enhance this learning by generating diverse experiences during tasks. The approach significantly improves agents' effectiveness and efficiency on web browsing and software engineering benchmarks compared to existing memory systems.
Explain Like I'm Five
Imagine a smart robot that keeps a diary of its good ideas and its mistakes. This paper teaches robots to write better diaries, so they can learn faster and get better at new tasks over time, just like you learn from practicing.
Possible Conflicts of Interest
A significant number of authors are affiliated with 'Google Cloud AI Research' and 'Google Cloud AI'. The experiments primarily utilize Google's own proprietary models (Gemini-2.5-flash, Gemini-2.5-pro). This constitutes a conflict of interest, as the authors are evaluating a system that leverages and potentially enhances technology developed by their employer.
Identified Limitations
Rating Explanation
The paper presents a strong technical contribution with a novel memory framework and test-time scaling method that demonstrates significant improvements in agent performance and efficiency on relevant benchmarks. The methodology is clearly described, and important limitations are acknowledged. However, the presence of a notable conflict of interest, with authors from Google extensively using and promoting Google's own AI models, slightly impacts the overall rating despite the paper's scientific merit.
Good to know
This is the Starter analysis. Paperzilla Pro fact-checks every citation, researches author backgrounds and funding sources, and uses advanced AI reasoning for more thorough insights.
Explore Pro →