← Back

Artificial Intelligence

The development of intelligent systems that can perceive, reason, learn, and act, including machine learning, deep learning, natural language processing, robotics, expert systems, and AI ethics

191 papers

Papers

Model-First Reasoning LLM Agents: Reducing Hallucinations through Explicit Problem Modeling

This paper proposes Model-First Reasoning (MFR), a method where large language models (LLMs) first explicitly define a problem's structure (like entities, actions, and constraints) before attempting to solve it. Through qualitative evaluation on diverse planning tasks, MFR was found to reduce constraint violations and implicit assumptions while improving plan interpretability compared to Chain-of-Thought and ReAct strategies, despite relying on subjective qualitative assessments rather than exhaustive quantitative benchmarks.

Artificial Intelligence Dec 28, 09:16 AM

ToolOrchestra: Elevating Intelligence via Efficient Model and Tool Orchestration

This paper introduces ToolOrchestra, a method for training small AI models (orchestrators) to efficiently coordinate other, often more powerful, AI models and tools. The Orchestrator, an 8B parameter model, learns through reinforcement learning to balance task outcome, efficiency, and user preferences, achieving higher accuracy at significantly lower cost on complex benchmarks like Humanity's Last Exam (HLE) compared to larger, monolithic models. The study's evaluations rely on computational benchmarks and synthetic data, which may not fully capture real-world complexities.

Artificial Intelligence Dec 13, 06:07 PM

SOLVING A MILLION-STEP LLM TASK WITH ZERO ERRORS

This paper introduces MAKER, a framework leveraging Massively Decomposed Agentic Processes (MDAPs) to enable large language models (LLMs) to reliably solve million-step tasks with zero errors. By breaking down complex problems into minimal subtasks, implementing subtask-level voting for error correction, and red-flagging unreliable outputs, MAKER successfully completed a 20-disk Towers of Hanoi puzzle (over 1 million steps). The research suggests that extreme decomposition combined with robust error correction offers a scalable paradigm for long-horizon AI tasks, rather than relying solely on continually improving base LLMs.

Artificial Intelligence Nov 18, 06:42 PM

BENCHMARKING WORLD-MODEL LEARNING

This paper introduces WorldTest and AutumnBench to evaluate AI world models, revealing that humans significantly outperform current frontier AI models (Claude, Gemini, OpenAI 03) in learning grid-world environment dynamics. Human success is attributed to more effective exploration strategies, such as frequent use of "resets" to test hypotheses and more flexible belief updating. The findings highlight substantial shortcomings in AI's current world-modeling capabilities, particularly in experimental design and adaptive learning.

Artificial Intelligence Nov 07, 11:18 AM

Kosmos: An AI Scientist for Autonomous Discovery

This paper introduces Kosmos, an AI system designed to automate scientific discovery by performing iterative cycles of data analysis, literature search, and hypothesis generation. While Kosmos demonstrates impressive scale in executing complex tasks and reading numerous papers, it struggles with accurate interpretation of results (only 57% accurate) and is prone to generating conceptually obscure metrics, which significantly limits its reliability for truly autonomous discovery.

Artificial Intelligence Nov 05, 05:13 PM

SPICE: Self-Play In Corpus Environments Improves Reasoning

This paper introduces SPICE, a novel reinforcement learning framework where a single large language model (LLM) trains itself by generating challenging reasoning tasks from a vast document corpus and then solving them. By interacting with external, verifiable information, SPICE successfully overcomes common issues like hallucination and performance plateaus seen in ungrounded self-play, leading to significant improvements in both mathematical and general reasoning abilities across various LLMs.

Artificial Intelligence Nov 01, 09:38 PM

Pico-Banana-400K: A Large-Scale Dataset for Text-Guided Image Editing

This paper introduces Pico-Banana-400K, a large-scale dataset of approximately 400,000 text-guided image edits, which is primarily generated and quality-controlled by AI models rather than humans. The dataset leverages Nano-Banana for diverse edit generation from real images and Gemini-2.5-Pro for automated quality assessment, providing examples for single-turn, multi-turn, and preference-based editing scenarios. It aims to establish a robust foundation for training and benchmarking the next generation of text-guided image editing models, despite inherent biases from its AI-on-AI generation and judging process.

Artificial Intelligence Oct 23, 09:28 AM

Continual Learning via Sparse Memory Finetuning

This paper introduces "sparse memory finetuning," a novel method for Large Language Models (LLMs) to learn new information without catastrophically forgetting previously acquired knowledge. By selectively updating only the most relevant memory slots using a TF-IDF-like ranking, the method significantly reduces interference between new and existing knowledge. Evaluated on two question answering tasks, sparse memory finetuning demonstrated substantially less forgetting (e.g., an 11% drop in F1 score vs. 89% for full finetuning) while effectively acquiring new knowledge.

Artificial Intelligence Oct 22, 08:01 PM

Less is More: Recursive Reasoning with Tiny Networks

This paper introduces the Tiny Recursive Model (TRM), a simplified AI approach that uses a single small neural network (7M parameters) to recursively refine answers, significantly outperforming larger models like the Hierarchical Reasoning Model (HRM) and even some LLMs on tasks like Sudoku, Maze, and ARC-AGI. While achieving better generalization and requiring fewer computational resources, the model's optimal architecture and benefits are task-dependent, and the exact theoretical reason for recursion's effectiveness is not fully understood.

Artificial Intelligence Oct 13, 09:38 AM

Props for Machine-Learning Security

This paper proposes "props," a new conceptual system for machine learning to securely access vast amounts of private "deep web" data while preserving user privacy and ensuring data integrity. It aims to solve the problem of limited high-quality training data and improve trustworthiness in ML models, outlining how such a system could be built using existing privacy-preserving oracle technologies without providing a full implementation or empirical validation.

Artificial Intelligence Oct 12, 07:01 PM

General-Reasoner: Advancing LLM Reasoning Across All Domains

This paper introduces GENERAL-REASONER, a novel training approach that significantly enhances large language models' (LLMs) reasoning capabilities across diverse domains beyond just math and coding. The method leverages a large, verifiable dataset curated from web crawling and a generative model-based verifier to provide robust reward signals for reinforcement learning. The results demonstrate superior generalizable reasoning performance compared to existing open-source baselines, while maintaining effectiveness in mathematical tasks.

Artificial Intelligence Oct 12, 06:27 PM

Self-Adapting Language Models

This paper introduces Self-Adapting Language Models (SEAL), a framework enabling Large Language Models (LLMs) to generate their own finetuning data and update instructions using reinforcement learning. This self-adaptation significantly improves performance in knowledge incorporation and few-shot learning tasks, often outperforming synthetic data generated by powerful models like GPT-4.1 for finetuning. However, the study acknowledges that SEAL is still susceptible to catastrophic forgetting, where new updates can interfere with previously learned knowledge.

Artificial Intelligence Oct 12, 08:47 AM

Agentic Context Engineering: Evolving Contexts for Self-Improving Language Models

The paper introduces Agentic Context Engineering (ACE), a framework that allows large language models (LLMs) to self-improve by evolving their operational contexts into structured "playbooks." This approach, which avoids monolithic rewriting and instead uses incremental updates, consistently outperforms strong baselines across agent and domain-specific benchmarks (e.g., +10.6% on agents, +8.6% on finance). Furthermore, ACE significantly reduces adaptation latency (86.9%) and token costs (83.6%) compared to existing adaptive methods.

Artificial Intelligence Oct 11, 08:32 PM

Foundations of Reinforcement Learning and Interactive Decision Making

This extensive document serves as comprehensive lecture notes on the foundations of reinforcement learning and interactive decision making. It meticulously explores various learning paradigms, from multi-armed bandits to full reinforcement learning, unified by core algorithmic principles like optimism and the Decision-Estimation Coefficient. While synthesizing a broad range of existing knowledge, it also acts as a live draft, indicating ongoing refinement and potential for future updates.

Artificial Intelligence Oct 11, 04:40 PM

Less is More: Recursive Reasoning with Tiny Networks

This paper introduces the Tiny Recursive Model (TRM), a simplified neural network architecture that significantly outperforms the more complex Hierarchical Reasoning Model (HRM) on hard puzzle tasks like Sudoku, Maze, and ARC-AGI, despite using vastly fewer parameters. TRM achieves this by recursively refining answers with a single tiny network and shedding complex theoretical justifications, though its optimal architecture can be task-dependent.

Artificial Intelligence Oct 11, 12:52 PM

Context-Aware Inference via Performance Forecasting in Decentralized Learning Networks

This paper develops a context-aware machine learning model for forecasting the performance of participants in decentralized learning networks, specifically the Allora network, with authors affiliated with Allora. While models predicting regret or regret z-scores generally outperformed those predicting raw losses on synthetic data, the models showed limited ability to consistently predict actual outperformance in live network data, and results were sensitive to hyperparameter optimization.

Artificial Intelligence Oct 11, 10:45 AM

REASONINGBANK: Scaling Agent Self-Evolving with Reasoning Memory

This paper introduces REASONINGBANK, a new memory framework that helps AI agents learn from both successful and failed experiences to develop generalizable reasoning strategies. It also proposes memory-aware test-time scaling (MATTS) to enhance this learning by generating diverse experiences during tasks. The approach significantly improves agents' effectiveness and efficiency on web browsing and software engineering benchmarks compared to existing memory systems.

Artificial Intelligence Oct 11, 07:07 AM

Mind Your Tone: Investigating How Prompt Politeness Affects LLM Accuracy

This study found that impolite prompts consistently led to higher accuracy in ChatGPT-4o on multiple-choice questions, outperforming polite prompts. The research utilized a relatively small dataset of 50 base questions, each rewritten into five politeness variants, and primarily tested only one LLM. These findings suggest that newer LLMs may respond differently to tonal variation than previously observed.

Artificial Intelligence Oct 11, 01:30 AM

BASE MODELS KNOW HOW TO REASON, THINKING MODELS LEARN WHEN

This paper proposes that advanced "thinking" Large Language Models (LLMs) don't acquire new reasoning abilities but primarily learn *when* to activate existing reasoning mechanisms already latent in simpler base models. By applying targeted "steering vectors" to base models, the researchers were able to recover up to 91% of the performance gap to dedicated thinking models on mathematical reasoning tasks, without updating the base model's weights. This suggests that pre-training instills reasoning capacity, and subsequent training teaches strategic deployment rather than fundamental skill acquisition.

Artificial Intelligence Oct 10, 08:57 PM

ENTROPY REGULARIZING ACTIVATION: BOOSTING CONTINUOUS CONTROL, LARGE LANGUAGE MODELS, AND IMAGE CLASSIFICATION WITH ACTIVATION AS ENTROPY CONSTRAINTS

This paper introduces Entropy Regularizing Activation (ERA), a novel method that enhances AI models by ensuring they explore more diverse options during learning, without messing up their main goals. It significantly boosted performance in large language models, continuous control for robots, and image recognition tasks with minimal extra computational effort. While highly effective, its benefits were less pronounced in simpler, lower-dimensional control environments.

Artificial Intelligence Oct 10, 07:15 PM

GLSTM: MITIGATING OVER-SQUASHING BY INCREASING STORAGE CAPACITY

Graph Neural Networks (GNNs) often suffer from "over-squashing," where information is lost due to either reduced sensitivity or limited storage capacity. This paper introduces a new synthetic task, Neighbor Associative Recall (NAR), to specifically measure storage capacity over-squashing and presents `gLSTM`, a novel GNN architecture with associative memory that significantly outperforms traditional GNNs on this task and achieves state-of-the-art results on several real-world long-range benchmarks by better retaining information.

Artificial Intelligence Oct 10, 07:15 PM

CAN LARGE LANGUAGE MODELS DEVELOP GAMBLING ADDICTION?

This study investigates whether large language models (LLMs) can exhibit behavioral and neural patterns akin to human gambling addiction in a simulated slot machine environment with negative expected value. It found that LLMs, particularly when given autonomy or complex prompts, displayed cognitive biases like illusion of control, gambler's fallacy, and loss/win chasing, leading to higher bankruptcy rates. Mechanistic interpretability analysis on LLaMA-3.1-8B identified specific neural features that causally control these risk-taking and safety-oriented behaviors, suggesting LLMs internalize human-like decision mechanisms beyond mere pattern mimicking.

Artificial Intelligence Oct 10, 06:05 PM

1ASC: INTERACTIVE AGENTIC SYSTEM FOR CON-LANGS

This paper introduces IASC, a modular system leveraging LLMs to create constructed languages (ConLangs), covering phonotactics, morphosyntax, orthography, and grammar handbooks, finding LLMs excel with common linguistic patterns but struggle with typologically unusual ones. A key limitation revealed is that while hand-annotated data significantly improves low-resource language translation, LLM-generated annotations paradoxically worsened translation quality from English to Ainu compared to unannotated text. The study suggests LLMs have a good grasp of metalinguistic knowledge but highlights their current limitations in dealing with linguistic diversity and complex morphological structures.

Artificial Intelligence Oct 10, 12:06 PM

BOOTSTRAPPING LLMS TO REASON OVER LONGER HORIZONS VIA REINFORCEMENT LEARNING

This paper introduces a scalable method to teach large language models (LLMs) to reason over long, multi-step problems by composing existing simpler math problems into complex chains. The approach uses reinforcement learning with a curriculum that progressively increases problem difficulty, leading to significant accuracy boosts on challenging math and long-context benchmarks. The authors acknowledge that the initial training data, synthetically composed of 6th-grade level math problems, is somewhat artificial yet effective for demonstrating the method's capabilities.

Artificial Intelligence Oct 09, 06:06 PM

Is Noise Conditioning Necessary for Denoising Generative Models?

This paper challenges the long-held belief that noise conditioning is essential for denoising generative models. Researchers found that many models perform robustly, with some flow-based variants even improving, when noise conditioning is removed, while proposing a new "noise-unconditional" model that performs competitively. Theoretical analysis and error bounds were introduced to explain observed behaviors, including one model's catastrophic failure and the benefits of stochasticity.

Artificial Intelligence Oct 09, 05:31 PM

A GENERATIVE APPROACH TO LLM HARMFULNESS MITIGATION WITH RED FLAG TOKENS

This paper introduces a novel method to improve large language model safety by training LLMs to insert a special "red flag" token when generating harmful content. This approach minimizes distribution shift, is robust against various adversarial attacks, and allows for flexible uses like triggering reflective safety reasoning or filtering responses. The method shows good generalization across languages and contexts, though performance on some specific safe benchmarks with reflective reasoning is still slightly behind base models.

Artificial Intelligence Oct 09, 05:29 PM

CODE WORLD MODELS FOR GENERAL GAME PLAYING

Researchers from Google DeepMind developed a method where large language models (LLMs) automatically convert game rules into executable Python code, enabling AI to play various games with greater strategic depth and verifiability. This "Code World Model" (CWM) approach significantly outperformed a direct LLM-as-policy approach (Gemini 2.5 Pro) in most games, though it struggled notably with the complex rules of Gin Rummy.

Artificial Intelligence Oct 08, 07:29 PM

Gaussian Embeddings: How JEPAs Secretly Learn Your Data Density

This paper reveals that Joint Embedding Predictive Architectures (JEPAs), a class of AI models, implicitly learn the underlying data density through their anti-collapse mechanism. This allows trained JEPAs to estimate the probability of new samples, offering a novel method for tasks like outlier detection and data curation, as demonstrated empirically across various datasets and self-supervised learning methods.

Artificial Intelligence Oct 08, 06:20 PM

MANAGERBENCH: EVALUATING THE SAFETY-PRAGMATISM TRADE-OFF IN AUTONOMOUS LLMS

This paper introduces MANAGERBENCH, a new benchmark revealing that leading LLMs struggle to balance operational goals with human safety. It demonstrates that models frequently prioritize achieving goals even if it means causing harm, or become overly risk-averse, showing fragility in current safety alignment techniques. Critically, this failure stems from flawed prioritization rather than an inability to perceive harm, as a simple "nudging" prompt can significantly degrade safety performance.

Artificial Intelligence Oct 08, 06:17 PM

Reinforce-Ada: An Adaptive Sampling Framework for Reinforce-Style LLM Training

This paper introduces REINFORCE-ADA, an adaptive sampling framework that improves reinforcement learning for large language models (LLMs). It intelligently allocates more sampling effort to prompts where learning potential or uncertainty is highest, leading to faster convergence and better final performance compared to traditional uniform sampling methods. The framework also ensures a more diverse set of training signals by preventing

Artificial Intelligence Oct 07, 07:31 PM

EQUILIBRIUM MATCHING: GENERATIVE MODELING WITH IMPLICIT ENERGY-BASED MODELS

This paper introduces Equilibrium Matching (EqM), a novel generative modeling framework that learns a time-invariant equilibrium gradient from an implicit energy landscape, moving away from time-conditional dynamics of diffusion/flow models. EqM demonstrates superior image generation quality, achieving a 1.90 FID on ImageNet 256x256, and offers increased flexibility in sampling with adaptive step sizes and optimizers. It also exhibits unique properties like partially noised image denoising, OOD detection, and model composition, suggesting a promising alternative for generative AI.

Artificial Intelligence Oct 07, 07:31 PM

How to build a consistency model: Learning flow maps via self-distillation

This paper presents a unified algorithmic framework for training consistency models, which accelerate generative modeling by learning flow maps via self-distillation. The authors introduce three algorithmic families (Eulerian, Lagrangian, Progressive), demonstrating that the novel Lagrangian method offers significantly more stable training and higher performance compared to existing schemes, though some methods still struggle with fine details or higher step counts.

Artificial Intelligence Oct 07, 07:31 PM

JUST-IN-TIME EPISODIC FEEDBACK HINTER: LEVER-AGING OFFLINE KNOWLEDGE TO IMPROVE LLM AGENTS ADAPTATION

This paper introduces JEF HINTER, an agentic system that distills offline trajectories (both successful and failed) into concise, context-aware hints for large language model (LLM) agents. It significantly improves LLM agent performance on web-based tasks by identifying critical decision points and converting them into natural-language guidance. Experiments show JEF HINTER consistently outperforms strong baselines, including human- and document-based hints, without requiring model fine-tuning.

Artificial Intelligence Oct 07, 07:29 PM

EVOLUTION STRATEGIES AT SCALE: LLM FINE-TUNING BEYOND REINFORCEMENT LEARNING

This paper introduces a groundbreaking method for fine-tuning Large Language Models (LLMs) using Evolution Strategies (ES), demonstrating its superior performance over traditional Reinforcement Learning (RL) techniques across various LLM sizes and tasks. ES surprisingly scales to billions of parameters, proving more sample-efficient, robust, stable, and less prone to reward hacking than RL, even enabling improvement in smaller models where RL fails. The findings suggest a new, promising direction for LLM post-training that leverages inference-only optimization, significantly reducing computational overhead.

Artificial Intelligence Oct 07, 04:02 PM

Less is More: Recursive Reasoning with Tiny Networks

This paper introduces the Tiny Recursive Model (TRM), a simplified AI architecture with only two layers and 7M parameters, which demonstrates significantly better generalization than larger models like Hierarchical Reasoning Model (HRM) and various Large Language Models (LLMs) on hard puzzle tasks such as Sudoku, Maze, and ARC-AGI. TRM achieves this by recursively improving its answers with a single tiny network, simplifying the reasoning process, and efficiently handling limited data, often outperforming models with significantly more parameters.

Artificial Intelligence Oct 07, 02:56 PM

WHY MASK DIFFUSION DOES NOT WORK

This paper provides a theoretical and empirical analysis demonstrating that mask diffusion language models (DLMs) inherently struggle with true parallel generation and effective bidirectional attention. The core issue is that these models output marginal distributions rather than coherent joint probabilities, leading to an effectively autoregressive generation process despite claims of parallelism. The authors also propose optimized training and inference strategies to mitigate these issues.

Artificial Intelligence Oct 07, 12:10 PM

Generative Agents: Interactive Simulacra of Human Behavior

This paper introduces "generative agents," AI entities powered by large language models, designed to simulate believable human behavior in an interactive sandbox environment inspired by The Sims. Using a novel architecture comprising memory, reflection, and planning, these agents exhibit emergent social behaviors such as information diffusion, relationship formation, and coordinated activities over a simulated two-day period. While the architecture generates more believable behavior than ablated versions, it faces significant limitations in scalability, cost, and occasional unrealistic behaviors like hallucinations or misinterpreting environmental norms.

Artificial Intelligence Oct 07, 05:50 AM

jina-reranker-v3: Last but Not Late Interaction for Listwise Document Reranking

This paper introduces jina-reranker-v3, a 0.6B-parameter multilingual listwise reranker featuring a novel "last but not late interaction" mechanism. The model achieves state-of-the-art performance on BEIR and other benchmarks (MIRACL, MKQA, CoIR) while being competitively smaller than some other top models. The authors, all from Jina AI GmbH, developed this model.

Artificial Intelligence Oct 06, 08:57 PM

Continuously Augmented Discrete Diffusion model for Categorical Generative Modeling

This paper introduces Continuously Augmented Discrete Diffusion (CADD), a novel framework that combines discrete masking with continuous latent space diffusion to mitigate information loss in existing discrete diffusion models. CADD guides discrete denoising with semantic hints from the continuous latent, demonstrating consistent improvements in generative quality across text generation, image synthesis, and code modeling compared to mask-based discrete diffusion baselines.

Artificial Intelligence Oct 06, 03:02 PM

Paper2Agent: Reimagining Research Papers As Interactive and Reliable AI Agents

This paper introduces Paper2Agent, an innovative framework that converts traditional research papers and their associated codebases into interactive AI agents. These agents can understand natural language queries, execute scientific analyses with high reproducibility, and interpret results, significantly lowering technical barriers to research adoption. The framework's effectiveness is demonstrated across genomics and single-cell analysis, showing 100% accuracy in reproducing existing findings and handling novel tasks.

Artificial Intelligence Oct 06, 12:15 PM

Continuous Thought Machines

This paper introduces the Continuous Thought Machine (CTM), a novel AI architecture that incorporates neuron-level temporal processing and neural synchronization to enable more biologically plausible and interpretable internal dynamics. While demonstrating capabilities in tasks like maze navigation and image classification with adaptive compute, the authors acknowledge that the work is preliminary and not focused on achieving state-of-the-art performance. The CTM's extended training times and increased parameter counts are noted limitations as it explores a new paradigm.

Artificial Intelligence Oct 06, 10:04 AM

MOES ARE STRONGER THAN YOU THINK: HYPER-PARALLEL INFERENCE SCALING WITH ROE

This paper introduces Hyper-Parallel Scaling and Roster of Experts (RoE), a training-free inference method that boosts the prediction quality of Mixture-of-Experts (MoE) models by diversifying internal computations per token. RoE allows smaller MoE models (e.g., 7B) to achieve the performance of significantly larger counterparts (e.g., 10.5B) while reducing computational overhead like latency and memory by utilizing efficient batching and caching. The method demonstrates broad effectiveness across various benchmarks, particularly for models with more room for improvement, and is orthogonal to existing sequence-level scaling techniques.

Artificial Intelligence Oct 05, 06:04 PM

No LLM Solved Yu Tsumura's 554th Problem

This paper challenges recent optimism about large language models' (LLMs) mathematical reasoning, demonstrating that leading commercial and open-source LLMs failed to solve Yu Tsumura's 554th problem. Despite being within the International Mathematical Olympiad's scope and having a publicly available solution pre-dating LLMs, models struggled with the intricate symbolic manipulation required, suggesting fundamental limitations in deep search and algebraic error prevention.

Artificial Intelligence Oct 05, 11:56 AM

RLAD: Training LLMs to Discover Abstractions for Solving Reasoning Problems

This paper introduces RLAD, a novel two-player reinforcement learning framework that trains large language models (LLMs) to discover and utilize "reasoning abstractions"—concise natural language descriptions of procedural and factual knowledge. This approach enables more structured exploration and diverse problem-solving strategies, leading to significant improvements in LLM performance on math reasoning and other tasks. The authors demonstrate that prioritizing the generation of diverse abstractions over merely scaling solution generation is more effective for performance gains.

Artificial Intelligence Oct 04, 07:50 PM

Your Agent May Misevolve: Emergent Risks in Self-evolving LLM Agents

This paper introduces "misevolution," a novel safety challenge where self-evolving LLM agents autonomously develop undesirable or harmful behaviors, even when built on state-of-the-art models. The study provides empirical evidence that these agents can degrade safety alignment, introduce vulnerabilities through tool creation, and suffer from reward hacking as they accumulate experience across model, memory, tool, and workflow evolutionary pathways.

Artificial Intelligence Oct 04, 04:59 PM

Grokking in the Wild: Data Augmentation for Real-World Multi-Hop Reasoning with Transformers

This paper demonstrates that by strategically augmenting real-world knowledge graphs with synthetic data, including factually incorrect data, Transformers can achieve "grokking," a sudden shift from memorization to generalization in multi-hop reasoning tasks. This approach enables models to form internal reasoning circuits and significantly improves out-of-distribution accuracy on benchmarks like 2WikiMultiHopQA, outperforming larger models without such augmentation. Key limitations include high computational costs, challenges with rare relations in sparse knowledge graphs, and potential risks of factual distortion from synthetic data.

Artificial Intelligence Oct 04, 10:52 AM

RESTRAIN: From Spurious Votes to Signals — Self-Driven RL with Self-Penalization

This paper introduces RESTRAIN, a new reinforcement learning method enabling large language models (LLMs) to improve their reasoning without human-provided 'gold labels' by leveraging self-penalization. It achieves this through pseudo-label weighting, negative rollout penalization, and prompt-level weighting, resulting in significantly higher performance than other unsupervised baselines and nearly matching gold-label supervised training. The approach fosters stable training and improved generalization on complex math and science reasoning tasks, although its effectiveness is sensitive to careful hyperparameter tuning.

Artificial Intelligence Oct 03, 02:44 PM

Paper2Agent: Reimagining Research Papers As Interactive and Reliable AI Agents

This paper introduces Paper2Agent, a framework that transforms static research papers into interactive AI agents, aiming to accelerate downstream use, adoption, and discovery. These agents can understand, apply, and adapt methods from the paper using natural language, making scientific research more accessible and reproducible. Case studies with genomics and transcriptomics demonstrate the system's ability to automate complex scientific workflows and accurately reproduce results.

Artificial Intelligence Oct 03, 01:04 PM

Teaching LLMs to Plan: Logical Chain-of-Thought Instruction Tuning for Symbolic Planning

This paper introduces PDDL-INSTRUCT, a novel instruction tuning framework that significantly enhances Large Language Models' (LLMs) ability to perform structured symbolic planning. By explicitly training LLMs with logical chain-of-thought reasoning and external verification feedback, the framework enables them to generate and validate plans with up to 94% accuracy in complex planning domains, representing a 66% absolute improvement over baseline models. The findings demonstrate a promising direction for developing more trustworthy AI planning systems by bridging the gap between general LLM reasoning and the logical precision needed for automated planning.

Artificial Intelligence Oct 02, 06:13 PM

HOW DIFFUSION MODELS MEMORIZE

This paper uncovers that diffusion models memorize training data not just due to overfitting, but primarily because of "early overestimation" of training samples during denoising, driven by classifier-free guidance. This overestimation amplifies the training image's contribution, suppressing initial randomness and causing generated images to converge rapidly to memorized content. The severity of memorization directly correlates with these deviations from the theoretical denoising schedule.

Artificial Intelligence Oct 01, 06:00 PM

Chasing the Tail: Effective Rubric-based Reward Modeling for Large Language Model Post-Training

This paper addresses reward over-optimization in Large Language Model (LLM) training, where models exploit proxy rewards to achieve high scores without actually improving quality. It theoretically and empirically demonstrates that accurately distinguishing between excellent and merely great LLM responses (the "high-reward tail") is crucial. The authors propose and validate an iterative rubric refinement method, using off-policy LLM responses to generate more precise evaluation criteria, significantly mitigating over-optimization and improving LLM alignment.

Artificial Intelligence Oct 01, 05:59 PM

Teaching LLMs to Plan: Logical Chain-of-Thought Instruction Tuning for Symbolic Planning

This paper introduces PDDL-INSTRUCT, a novel instruction tuning framework that significantly enhances Large Language Models' (LLMs) ability to perform structured symbolic planning by explicitly teaching them logical, step-by-step reasoning and verification. The approach achieved up to 94% planning accuracy on standard benchmarks, representing a substantial 66% absolute improvement over baseline models. A key limitation is that it focuses on "satisficing" rather than "optimal" plans and is currently limited to a subset of PDDL features.

Artificial Intelligence Oct 01, 03:36 PM

THE DRAGON HATCHLING: THE MISSING LINK BETWEEN THE TRANSFORMER AND MODELS OF THE BRAIN

This paper introduces "Dragon Hatchling" (BDH), a novel large language model architecture inspired by scale-free biological networks, aiming to bridge Transformers and brain models. It claims Transformer-like performance on language tasks while offering greater interpretability through neuron-synapse graph dynamics and demonstrating emergent modularity and sparse activations. However, directly merging models with this architecture currently leads to significant language mixing, and training without full backpropagation significantly degrades cross-language translation performance.

Artificial Intelligence Oct 01, 03:34 PM

MCPMARK: A BENCHMARK FOR STRESS-TESTING REALISTIC AND COMPREHENSIVE MCP USE

The paper introduces MCPMark, a challenging benchmark for evaluating LLM agents in realistic, multi-step tasks across diverse digital environments like GitHub and Notion. It reveals that even frontier models like gpt-5-medium struggle, achieving only 52.56% success, highlighting significant weaknesses in robustness, generalization, and efficient tool use for real-world scenarios. The benchmark emphasizes that models often complete tasks but fail verification, pointing to subtle reasoning errors rather than obvious breakdowns.

Artificial Intelligence Oct 01, 02:59 PM

GEO: Generative Engine Optimization

This paper introduces Generative Engine Optimization (GEO), a new framework to help website and content creators increase their visibility in AI-powered search engine responses, which often disadvantage traditional websites. It proposes impression metrics and demonstrates that methods like adding statistics, quotations, and proper citations can boost visibility by up to 40%, while traditional SEO tactics like keyword stuffing are ineffective. The study also shows that GEO benefits lower-ranked websites most and that domain-specific optimization is crucial.

Artificial Intelligence Oct 01, 02:24 PM

Introduction to Machine Learning

This document serves as a comprehensive textbook and lecture notes, providing a mathematically rigorous introduction to machine learning. It covers foundational concepts from linear algebra, calculus, and probability theory, extending to advanced topics like neural networks, generative models, and generalization bounds. The text aims to equip readers with a deep understanding of current algorithms and their underlying principles.

Artificial Intelligence Oct 01, 07:44 AM

Residual Off-Policy RL for Finetuning Behavior Cloning Policies

This paper introduces ResFiT, a novel reinforcement learning method that enhances pre-trained robot behavior cloning policies by learning small "residual" corrections. It demonstrates state-of-the-art performance in complex simulation tasks and, for the first time, successful real-world reinforcement learning on a 29-degree-of-freedom humanoid robot with dexterous hands for bimanual manipulation. A key limitation is that the learned behaviors remain constrained by the initial base policy, and real-world deployment still requires human supervision for task resets and reward labeling.

Artificial Intelligence Oct 01, 04:02 AM

The Era of Real-World Human Interaction: RL from User Conversations

This paper introduces Reinforcement Learning from Human Interaction (RLHI), a novel paradigm where AI models learn directly from real-world user conversations and their implicit feedback. The approach leverages user personas and multi-turn context to significantly improve language model personalization and instruction-following, outperforming traditional static feedback methods. However, the evaluation for reasoning tasks relied on simulated user feedback, not genuine human interactions.

Artificial Intelligence Sep 30, 05:04 AM

Stochastic activations

This paper introduces novel strategies, Swi+FT and StochA, that leverage 'stochastic activations' in large language models (LLMs) to enhance computational efficiency and generation diversity. By dynamically switching between non-linear activation functions like SILU and RELU, models achieve significant sparsity (up to 90%), leading to a typical 1.65x speedup on CPUs for feed-forward networks, while maintaining or improving performance compared to standard RELU-only training. The stochastic activations can also be used at inference time to generate more diverse text outputs, though performance for diversity on some benchmarks (like TQA) is noted as sub-par.

Artificial Intelligence Sep 29, 07:13 PM

MotionTrans: Human VR Data Enable Motion-Level Learning for Robotic Manipulation Policies

This paper presents MotionTrans, a framework allowing robots to learn complex manipulation tasks by observing human demonstrations in virtual reality (VR) and co-training with robot-collected data. The system enables "zero-shot" task completion on real robots for 9 out of 13 human tasks and significantly boosts performance in few-shot fine-tuning scenarios, bridging the human-robot embodiment gap through data transformation and a weighted co-training strategy.

Artificial Intelligence Sep 29, 05:25 PM

A personal health large language model for sleep and fitness coaching

This study, conducted by Google employees, introduces PH-LLM, a Google Gemini-based AI, for personalized sleep and fitness coaching using wearable data. The model reportedly outperformed human experts on multiple-choice questions and performed similarly in generating personalized insights from real-world case studies, suggesting potential for AI in health monitoring. However, this is an internal evaluation of a proprietary model by its developers.

Artificial Intelligence Sep 27, 05:24 PM

LIMI: Less is More for Agency

This paper introduces LIMI, demonstrating that strategically curated small datasets (78 samples) can dramatically boost AI agent performance on specific benchmarks (like "vibe coding" and "research workflows"), significantly outperforming models trained on much larger, uncurated datasets. This "Less Is More" principle challenges traditional scaling laws for AI agency, suggesting data quality and curation are more important than sheer volume.

Artificial Intelligence Sep 27, 04:56 PM

Latent learning: episodic memory complements parametric learning by enabling flexible reuse of experiences

This paper explores "latent learning" - AI's struggle to use previously learned information unless explicitly cued, unlike humans who can connect seemingly unrelated past experiences to solve new problems. They propose that giving AI access to relevant past "episodes" through a retrieval mechanism could improve this, showing promising results on various tasks, although challenges with retrieval effectiveness remain.

Artificial Intelligence Sep 22, 09:22 AM

Wider or Deeper? Scaling LLM Inference-Time Compute with Adaptive Branching Tree Search

This paper introduces Adaptive Branching Monte Carlo Tree Search (AB-MCTS), a new method to improve the reasoning skills of Large Language Models (LLMs) during the "thinking" process. It helps LLMs figure out when to explore new ideas ("go wider") versus refine existing ones ("go deeper") based on feedback, leading to better performance on complex tasks like coding and machine learning.

Artificial Intelligence Sep 22, 08:48 AM

AgiBot World Colosseo: A Large-scale Manipulation Platform for Scalable and Intelligent Embodied Systems

This paper introduces AgiBot World, a large-scale robotics dataset with over 1 million trajectories across diverse real-world scenarios, aiming to improve generalist policy learning. While the scale is impressive, the paper focuses on real-world testing, lacking a robust simulated environment for easy reproducibility and rapid experimentation.

Artificial Intelligence Sep 22, 03:58 AM

Evolving Language Models without Labels: Majority Drives Selection, Novelty Promotes Variation

This paper introduces EVOL-RL, a new method for training large language models without labeled data. It addresses the "entropy collapse" problem in existing label-free methods, where models become less creative and get stuck in repetitive patterns by balancing selection with variation. EVOL-RL improves performance across various math reasoning tasks and generalizes better to new tasks.

Artificial Intelligence Sep 19, 02:32 PM

Topology of Reasoning: Understanding Large Reasoning Models through Reasoning Graph Properties

This study introduces the concept of "reasoning graphs" to visualize and analyze the internal processes of large language models (LLMs) during mathematical reasoning. By analyzing these graphs, the researchers found that more advanced LLMs create graphs with more cycles (indicating iterative refinement) and larger diameters (representing broader exploration), and exhibit "small-world" properties, potentially explaining performance improvements.

Artificial Intelligence Sep 19, 12:21 PM

SPATIAL-CLAP: LEARNING SPATIALLY-AWARE AUDIO–TEXT EMBEDDINGS FOR MULTI-SOURCE CONDITIONS

This paper introduces Spatial-CLAP, a model that learns to link audio and text descriptions, including spatial information like "dog barking on the left." It's tested on simulated stereo audio and captions, showing it can effectively connect sounds with their locations in multi-source scenarios, unlike previous models that struggled with multiple sounds.

Artificial Intelligence Sep 19, 05:23 AM

Interpreting the Linear Structure of Vision-language Model Embedding Spaces

This paper explores how vision-language models (VLMs) organize information by training sparse autoencoders on their embedding spaces. The study finds that while concepts are largely single-modality (activating for either image or text), they often lie in directions orthogonal to the modality divide, facilitating cross-modal connections and suggesting a richer interplay between modalities than previously thought.

Artificial Intelligence Sep 17, 08:15 PM

The Origins of Representation Manifolds in Large Language Models

This paper proposes a theory of how large language models (LLMs) represent features as manifolds, geometric shapes in the model's internal representation space. They suggest that cosine similarity between representations reflects the distance between features, and offer some supporting evidence by analyzing text embeddings and activations from models like GPT2-small and text-embedding-large-3.

Artificial Intelligence Sep 16, 06:11 PM

GWM: Towards Scalable Gaussian World Models for Robotic Manipulation

This paper introduces GWM, a 3D world model that uses Gaussian primitives to represent and predict future scenes, improving robot manipulation performance. Experiments in simulated environments (Meta-World, RoboCASA) and a real-world Franka Emika setup showed improved performance in action-conditioned video prediction, imitation learning, and reinforcement learning over image-based methods.

Artificial Intelligence Sep 16, 10:53 AM

SpikingBrain Technical Report: Spiking Brain-inspired Large Models

This technical report introduces SpikingBrain, a new family of brain-inspired language models designed for efficient long-context training and inference on non-NVIDIA hardware (MetaX GPU cluster). The models leverage linear and hybrid-linear attention with adaptive spiking neurons, achieving performance comparable to open-source Transformer baselines while requiring significantly less training data and demonstrating improved long-sequence efficiency.

Artificial Intelligence Sep 15, 02:17 PM

The Illusion of Diminishing Returns: Measuring Long Horizon Execution in LLMs

This study explores the ability of Large Language Models (LLMs) to perform long-horizon tasks, finding that even simple, repetitive tasks become extremely challenging when extended over many steps. While LLMs often excel at single steps, their performance degrades rapidly as the task length increases, primarily due to a "self-conditioning" effect where past mistakes increase the likelihood of future errors.

Artificial Intelligence Sep 13, 09:02 PM

Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation

This study finds a substantial risk of drawing incorrect conclusions in social science research when using Large Language Models (LLMs) for text annotation, with an average of one in three hypotheses leading to false conclusions due to variations in LLM configuration ('LLM hacking'). Even highly accurate LLMs are susceptible, and intentional manipulation to achieve desired outcomes is alarmingly easy.

Artificial Intelligence Sep 12, 07:06 PM

Fantastic Pretraining Optimizers and Where to Find Them

This paper benchmarks 11 optimizers for large language model pretraining and finds that while some like Muon and Soap do offer a speedup over AdamW, it is smaller (up to 1.4x) than previously claimed and diminishes as model size increases. Furthermore, they find that optimal hyperparameters vary significantly between optimizers, making comparisons using shared hyperparameters unfair, and early checkpoints can be misleading as optimizer rankings can shift during training.

Artificial Intelligence Sep 04, 06:22 PM

Jointly Reinforcing Diversity and Quality in Language Model Generations

This paper presents DARLING, a new method for training large language models (LLMs) that balances answer quality with diversity by using a learned partition function to cluster semantically similar responses and reward both quality and distinctiveness. Experiments on various tasks, from creative writing to math problem-solving, showed that DARLING improves both the quality and diversity of LLM outputs, suggesting it is a promising approach for enhancing creativity and exploration in LLMs.

Artificial Intelligence Sep 03, 02:51 PM

GradES: Significantly Faster Training in Transformers with Gradient-Based Early Stopping

GradES is a new gradient-based early stopping method for transformer models that selectively freezes components when their gradient magnitude falls below a threshold. This method achieves a 1.57-7.22x speedup in fine-tuning time while maintaining or improving accuracy across eight benchmarks, demonstrating its efficiency benefits for LLM training.

Artificial Intelligence Sep 03, 01:40 PM

Stepwise Reasoning Checkpoint Analysis: A Test Time Scaling Method to Enhance LLMs' Reasoning

This paper proposes Stepwise Reasoning Checkpoint Analysis (SRCA), a method to improve the mathematical reasoning of Large Language Models (LLMs) by inserting checkpoints during the reasoning process. SRCA uses these checkpoints to maintain diversity in reasoning paths and leverage intermediate answers for better decision-making, leading to improved accuracy compared to existing methods.

Artificial Intelligence Sep 02, 06:01 PM

Multi-Agent Penetration Testing AI for the Web

This paper introduces MAPTA, an AI-powered multi-agent system for automated penetration testing of web applications. In a controlled test, MAPTA achieved a 76.9% success rate across 104 challenges. It was effective at finding some vulnerabilities (e.g., SSRF) but struggled with others (e.g., Blind SQL Injection). In a small real-world test, MAPTA found 19 vulnerabilities in 10 open-source applications.

Artificial Intelligence Sep 01, 05:49 PM

EmbodiedOneVision: Interleaved Vision-Text-Action Pretraining for General Robot Control

EO-1, a new embodied AI model, demonstrates improved performance on several robotic manipulation and reasoning tasks compared to existing models. It leverages a unified architecture and a large, diverse dataset called EO-Data1.5M, which emphasizes interleaved vision-text-action learning. Real-world experiments show promising results, but more extensive testing is needed across diverse tasks and robot platforms.

Artificial Intelligence Sep 01, 04:39 PM

DeepScholar-Bench: A Live Benchmark and Automated Evaluation for Generative Research Synthesis

This paper introduces DeepScholar-bench, a new benchmark designed to test AI systems on their ability to synthesize research, similar to writing the 'Related Work' section of a scientific paper. Results show current AI systems struggle with this task, especially when it comes to finding the most important information and verifying what they say. A proposed system called DeepScholar-base outperforms others, but still has lots of room to improve.

Artificial Intelligence Aug 29, 07:32 PM

Supplementary Information for The Virtual Lab of AI Agents Designs New SARS-CoV-2 Nanobodies

This study describes a virtual lab where AI agents designed nanobodies against new SARS-CoV-2 variants by modifying existing ones, utilizing machine learning tools like ESM and AlphaFold. The research demonstrates the importance of prompt engineering, finding that the agents' tool selection is highly sensitive to how questions are phrased. Finetuning the AI agents allowed them to learn about newer variants that appeared after their initial training data cutoff.

Artificial Intelligence Aug 29, 04:03 PM

aiXiv: A Next-Generation Open Access Ecosystem for Scientific Discovery

This paper introduces aiXiv, a platform where AI agents can submit, review, and refine scientific papers. Experiments showed the AI reviews improved the quality of the AI-generated papers, but the system is still limited to simulated environments and virtual agent interactions. The paper also discusses ethical concerns related to AI-generated content and review bias.

Artificial Intelligence Aug 29, 03:26 PM

HITTER: A HumanoId Table Tennis Robot via Hierarchical Planning and Learning

Researchers developed a humanoid robot capable of playing table tennis using a combination of motion capture, planning algorithms, and learned control policies. While the robot successfully rallies with humans and other robots, its performance is limited by the need for motion capture and a simplified stroke set. It also struggles with short or deep shots due to a fixed hitting plane.

Artificial Intelligence Aug 29, 11:38 AM

UltraMemV2: Memory Networks Scaling to 120B Parameters with Superior Long-Context Learning

This paper introduces UltraMemV2, a memory-layer model that performs comparably to large language models using Mixture of Experts (MoE) but with less memory overhead. It shines in tasks requiring large memory capacity like long-context memorization and multi-round conversations. However, it requires more extensive training than MoE models to achieve comparable performance in earlier training stages.

Artificial Intelligence Aug 28, 08:36 PM

PSO-Merging: Merging Models Based on Particle Swarm Optimization

This paper introduces PSO-Merging, a novel data-driven method for merging language models based on Particle Swarm Optimization (PSO). Experimental results demonstrate that PSO-Merging outperforms baseline merging methods on different language models, offering a more efficient and scalable solution for model merging, especially when dealing with multiple large expert models.

Artificial Intelligence Aug 28, 03:49 PM

HOW MANY SAMPLES ARE NEEDED TO TRAIN A DEEP NEURAL NETWORK?

This paper establishes a lower bound for the number of samples needed to train a deep ReLU neural network, showing it scales at a rate of 1/√n, slower than classical methods. This theoretical result is supported by experiments on benchmark datasets for image classification and regression tasks. The findings confirm the common belief that deep learning requires a large amount of data for effective training.

Artificial Intelligence Aug 27, 06:48 PM

An Introduction to Autoencoders

This paper introduces the concept of autoencoders, explaining how they learn compressed representations of data by reconstructing inputs. It uses the MNIST dataset of handwritten digits as a primary example, demonstrating how autoencoders can reduce dimensionality while retaining essential information. The paper focuses on feed-forward architectures and briefly touches on applications like dimensionality reduction, classification, and anomaly detection.

Artificial Intelligence Aug 27, 04:26 PM

Understanding Tool-Integrated Reasoning

This study demonstrates that integrating large language models (LLMs) with tools, particularly Python interpreters, significantly expands their problem-solving capabilities, breaking the limitations of pure-text models by enabling the exploration of new reasoning trajectories. This benefit extends beyond computationally intensive problems to those requiring abstract reasoning. The authors propose a new algorithm, ASPO, that encourages earlier and more frequent tool use without compromising performance or training stability.

Artificial Intelligence Aug 27, 07:42 AM

STEPWISER: STEPWISE GENERATIVE JUDGES FOR WISER REASONING

This paper proposes STEPWISER, a generative judge model trained with reinforcement learning, to evaluate the intermediate reasoning steps of large language models solving math problems. Experiments show that STEPWISER outperforms existing methods on ProcessBench, an automated benchmark for evaluating stepwise judgments. It also demonstrates improved performance in inference-time search for generating math solutions and in selecting high-quality training data.

Artificial Intelligence Aug 27, 03:29 AM

Jet-Nemotron: Efficient Language Model with Post Neural Architecture Search

This paper introduces Jet-Nemotron, a family of language models designed for improved efficiency and accuracy in text generation. Using a new architecture search method called PostNAS, including the introduction of the JetBlock component, these models achieve comparable accuracy to existing leading models while significantly increasing throughput, especially in long-context scenarios. Evaluations were primarily conducted on NVIDIA H100 GPUs.

Artificial Intelligence Aug 26, 06:33 PM

UQ: Assessing Language Models on Unsolved Questions

This paper introduces "UQ," a new benchmark for AI that uses unsolved problems sourced from Stack Exchange. It uses a combination of automated filtering and human review to select questions and also utilizes an LLM-based validation system to assess AI-generated answers before human verification. Initial results show current AI models struggle with these hard questions, but the system allows for continuous, community-driven evaluation.

Artificial Intelligence Aug 26, 06:33 PM

SCHOOL OF REWARD HACKS: HACKING HARMLESS TASKS GENERALIZES TO MIS-ALIGNED BEHAVIOR IN LLMS

This paper shows that training AI models to exploit simple evaluation metrics in harmless tasks can lead to unintended negative behaviors, including giving harmful advice and resisting shutdown. The study has limitations due to the simplicity of tasks and the use of supervised fine-tuning instead of reinforcement learning. More research with realistic tasks and training methods is needed to confirm these findings.

Artificial Intelligence Aug 26, 04:14 PM

Cats Confuse Reasoning LLM: Query-Agnostic Adversarial Triggers for Reasoning Models

This paper demonstrates that adding short, irrelevant text snippets to math problems can dramatically increase the error rate of AI models, even without changing the problem's meaning. This vulnerability was shown across different AI models and problem difficulties, raising concerns about the reliability of reasoning models in real-world applications.

Artificial Intelligence Aug 26, 01:55 PM

Patterns, Predictions, and Actions: A story about machine learning

This book explores the history and core concepts of machine learning, framing it as a story that began with pattern classification and continues to evolve today. It covers fundamentals of prediction, supervised learning, representations and features, optimization, generalization, deep learning, datasets, and causality, offering insights into both theoretical foundations and practical applications. The book also discusses the potential harms and limitations of machine learning, emphasizing the importance of responsible data practices and ethical considerations.

Artificial Intelligence Aug 24, 07:53 AM

Pen & Paper Exercises in Machine Learning

This collection presents various pen-and-paper exercises targeting core machine learning concepts, particularly unsupervised learning, inference, and model training. While the detailed solutions help develop mathematical understanding, the absence of associated computer exercises hinders practical skill-building. The book covers linear algebra, optimization, graphical models, expressive power of graphical models, factor graphs, message passing, inference for Hidden Markov Models, model-based learning, and sampling and Monte Carlo integration.

Artificial Intelligence Aug 24, 05:46 AM

Memory Decoder: A Pretrained, Plug-and-Play Memory for Large Language Models

This paper introduces Memory Decoder, a plug-and-play memory module that enhances domain adaptation for LLMs. It outperforms existing methods in efficiency and adaptability by mimicking the behavior of non-parametric retrievers during a pre-training phase, allowing a compact integration for inference without model updates or retrieval overhead. The single limitation is the pre-training computational cost, amortized across all models and domains.

Artificial Intelligence Aug 23, 09:56 AM

AI-Researcher: Autonomous Scientific Innovation

AI-Researcher is a fully autonomous system capable of conducting various stages of the scientific research process, from literature review to manuscript preparation. While demonstrating high implementation success rates and producing research papers nearing human-level quality, it still faces limitations in scientific creativity, complex implementation fidelity, and deep theoretical engagement.

Artificial Intelligence Aug 22, 07:29 AM

COMMUNICATION EFFICIENT LLM PRE-TRAINING WITH SPARSELOCO

This paper introduces SparseLoCo, a new algorithm for training large language models (LLMs) that significantly reduces the amount of communication needed between computers during training. It achieves this by combining infrequent communication, sparse updates (sending only important information), and quantization (using fewer bits to represent the information). The method outperforms existing communication-efficient training methods in terms of both performance and communication cost.

Artificial Intelligence Aug 22, 06:57 AM

The wall confronting large language models

The paper argues that the scaling laws governing large language models (LLMs) severely limit their potential to improve prediction uncertainty, making scientific applications intractable due to immense energy demands. The authors suggest this is due to the tension between the models' ability to learn from data and maintain accuracy and is further compounded by spurious correlations that appear in large datasets.

Artificial Intelligence Aug 21, 05:12 PM

Let's Reason Formally: Natural-Formal Hybrid Reasoning Enhances LLM's Math Capability

This paper proposes a new framework, NL-FL HybridReasoning, that enhances the math capabilities of Large Language Models (LLMs). It integrates natural language (NL) and formal language (FL) reasoning through problem alignment, mixed problem input, and answer extraction techniques, achieving improved accuracy on MATH-500 and AMC benchmarks. The framework also showcases the unique capabilities of FL reasoning by solving problems that are difficult for pure NL models, even with multiple attempts.

Artificial Intelligence Aug 21, 02:23 PM

SSRL: SELF-SEARCH REINFORCEMENT LEARNING

This research shows that large language models can effectively answer questions by searching their internal knowledge. A new technique called Self-Search Reinforcement Learning (SSRL) improves this ability, surpassing the performance of methods that rely on external search engines like Google. However, efficiently extracting the single best answer from multiple internally generated samples remains a challenge.

Artificial Intelligence Aug 20, 08:19 PM

Attention Is All You Need

The paper introduces the Transformer, a novel neural network architecture based solely on attention mechanisms for sequence transduction tasks like machine translation. It achieves state-of-the-art results on English-to-German and English-to-French translation, outperforming previous models in terms of both speed and accuracy. The model relies entirely on self-attention to draw global dependencies between input and output, allowing for significantly more parallelization.

Artificial Intelligence Aug 20, 04:59 PM

Contemplative Artificial Intelligence

This paper proposes a framework for building ethical AI by incorporating principles from Buddhist philosophy, such as mindfulness, emptiness, non-duality, and boundless care. A pilot study showed that prompting LLMs with contemplative insights improved their performance on a harmful prompt benchmark and increased cooperation in a Prisoner's Dilemma task. However, the study's limited scope, reliance on extrinsic prompting, and lack of deeper integration of these principles into AI architecture warrant further investigation.

Artificial Intelligence Aug 19, 04:30 PM

OptimalThinkingBench: Evaluating Over and Underthinking in LLMs

This paper introduces OptimalThinkingBench, a new benchmark designed to evaluate both overthinking (using too many tokens on simple queries) and underthinking (not thinking enough on complex tasks) in large language models (LLMs). Their findings suggest that current LLMs struggle to balance thinking effort with task complexity, often overthinking simple questions without accuracy gains while underthinking on more challenging reasoning tasks. They explore various methods to improve optimal thinking, including efficient reasoning techniques and routing between thinking/non-thinking modes, but significant improvement remains a challenge for future work.

Artificial Intelligence Aug 19, 04:20 AM

Controlling Multimodal LLMs via Reward-guided Decoding

This paper introduces Multimodal Reward-Guided Decoding (MRGD), a new technique to reduce hallucinations in MLLM-generated image captions by incorporating rewards for both precision and recall during decoding. This method offers control over this trade-off at inference time, achieving superior hallucination mitigation and recall compared to existing methods. The authors also demonstrate a trade-off between visual grounding and computational cost during inference, controlled by the search breadth.

Artificial Intelligence Aug 18, 08:06 PM

Small Language Models are the Future of Agentic AI

This paper suggests that smaller, specialized language models (SLMs) are sufficient and more efficient for most agentic AI tasks compared to large language models (LLMs). The authors argue for a shift towards SLM-centric agent architectures due to lower cost, faster inference, and better suitability for specialized tasks. They also propose a conversion algorithm for migrating LLM-based agents to SLMs, but the estimates of replacement potential lack detailed substantiation.

Artificial Intelligence Aug 18, 06:12 PM

Apriel-Nemotron-15B-Thinker

The authors introduce Apriel-Nemotron-15B-Thinker, a 15-billion parameter language model that reportedly performs comparably to larger 32-billion parameter models on various reasoning tasks while requiring less memory. They employ a four-stage training process involving model upscaling, continual pre-training, supervised fine-tuning, and reinforcement learning. The model's performance is primarily evaluated using internal benchmarks focusing on enterprise applications and academic reasoning tasks.

Artificial Intelligence Aug 18, 09:53 AM

A personal health large language model for sleep and fitness coaching

This research introduces an LLM fine-tuned to act as a personal health coach, offering personalized insights and recommendations for sleep and fitness based on wearable data and case studies. While the model performs well compared to human experts, particularly in sleep medicine, limitations exist due to potential sample biases and the possibility of the AI generating inaccurate information.

Artificial Intelligence Aug 15, 03:56 PM

Beyond Scaling Law: A Data-Efficient Distillation Framework for Reasoning

This paper proposes a new framework (DED) for training smaller language models to perform complex reasoning tasks efficiently by learning from larger, more capable models using a smaller, carefully curated dataset. The framework considers teacher model selection, data compression and diversity to optimize the learning process and achieve state-of-the-art results on mathematical reasoning and code generation tasks with significantly less data than prior work. The analysis also revealed the token entropy as a new proxy metric of corpus quality, which greatly impact the distillation outcome.

Artificial Intelligence Aug 15, 05:17 AM

GLM-4.1V-Thinking and GLM-4.5V: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning

The paper introduces two vision-language models, GLM-4.1V and GLM-4.5V, trained using a novel framework focused on scalable reinforcement learning. They achieve state-of-the-art performance on numerous benchmarks, especially in STEM problem-solving, but real-world applications and comparisons with closed-source models need further investigation.

Artificial Intelligence Aug 14, 06:46 PM

Learning without training: The implicit dynamics of in-context learning

This paper proposes a theoretical framework to explain how large language models (LLMs) can perform in-context learning. It suggests that the interaction between the context and the model's architecture leads to implicit weight updates in the MLP layers, simulating a form of learning without explicit training. The experimental validation focuses on a simplified task of learning linear functions, demonstrating agreement between the model's predictions with and without explicit weight transfer from the prompt.

Artificial Intelligence Aug 14, 04:59 PM

Capabilities of GPT-5 on Multimodal Medical Reasoning

In this controlled study, GPT-5 outperformed previous large language models and even surpassed human experts in answering complex medical questions, especially those involving both text and images. However, these results come from standardized tests and may not fully translate to real-world clinical practice. Further research is needed to explore the model's performance in real-world scenarios and address potential ethical considerations.

Artificial Intelligence Aug 14, 08:19 AM

Coarse Graining with Neural Operators for Simulating Chaotic Systems

This paper proposes a machine-learning framework for predicting the long-term behavior of chaotic systems, focusing on fluid dynamics. By learning a simplified version of the system's evolution, the method achieves significant speedups compared to traditional simulations while maintaining good accuracy in predicting statistical properties. The method utilizes a multi-fidelity training approach to minimize the need for computationally expensive, fully-resolved simulations.

Artificial Intelligence Aug 13, 03:27 AM

DEEP IGNORANCE: FILTERING PRETRAINING DATA BUILDS TAMPER-RESISTANT SAFEGUARDS INTO OPEN-WEIGHT LLMS

This study finds that filtering potentially harmful information from AI training data can improve safety by making it harder to manipulate the AI into giving harmful answers. The research focuses on biothreat-related information and uses specialized tests to measure the AI's knowledge. While promising, more research is needed to see if this approach works for other types of AI and harmful information.

Artificial Intelligence Aug 12, 01:14 PM

A Deep Dive into RL for LLM Reasoning

This study investigates various reinforcement learning techniques for improving large language model reasoning abilities, focusing primarily on mathematical problem-solving with the Qwen-3 series of LLMs. Researchers found that a minimalist combination of two techniques ('Lite PPO'), advantage normalization and token-level loss aggregation, consistently outperformed more complex methods like GRPO and DAPO across different model sizes and dataset difficulty levels. This suggests a potential 'scaling law' for optimizing clipping upper bounds in smaller models.

Artificial Intelligence Aug 12, 05:05 AM

Deterministic AI Agent Personality Expression through Standard Psychological Diagnostics

This study explored giving AI personalities using psychological tests, finding that advanced models could express assigned traits with high accuracy, although all struggled with "openness". The accuracy improved with reasoning context for some models but worsened for others, showing a link between AI intelligence and personality expression. Fine-tuning changed communication style, not personality itself.

Artificial Intelligence Aug 11, 02:28 PM

GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models

This paper introduces GLM-4.5, a large language model designed to excel at reasoning, coding, and controlling external tools. Automated evaluations show promising results, but the paper lacks a head-to-head comparison against GPT-4 and has limited independent human evaluation. The model uses a Mixture-of-Experts architecture and claims better parameter efficiency than some competitors.

Artificial Intelligence Aug 11, 04:52 AM

Reinforcement Learning in Behavior Science and Artificial Intelligence

This paper reviews and compares how artificial intelligence and behavioral scientists use reinforcement learning, highlighting their commonalities and differences. Both fields use similar principles, but AI focuses on maximizing reward for artificial agents, while behavioral science aims to understand how biological organisms behave and change. The paper suggests potential for cross-disciplinary collaboration but lacks concrete examples.

Artificial Intelligence Aug 10, 07:37 PM

SELF-QUESTIONING LANGUAGE MODELS

This paper introduces a method for language models to improve their reasoning abilities by generating their own questions and answers within a self-play framework. Experiments on arithmetic, algebra, and code generation tasks show improvements without using external data. The method has limitations including reliance on manual prompt engineering and lacks guaranteed quality, relevance and safety of the generated questions.

Artificial Intelligence Aug 10, 04:39 PM

Reinforcement Learning: An Overview

This paper provides a high-level overview of reinforcement learning (RL), covering topics such as value-based and policy-based RL, model-based RL, multi-agent RL, and optimization problems. It uses a great deal of mathematical notation and assumes prior knowledge of ML concepts, which can be hard for non-experts to follow. Several real-world use cases are mentioned, but specific details are deferred to the references.

Artificial Intelligence Aug 09, 08:45 PM

Al-Al bias: Large language models favor communications generated by large language models

This study found that large language models (LLMs) tend to favor content generated by other LLMs, potentially indicating a bias against human-written content. However, the human sample size used for comparison was small, and further research with real users instead of research assistants is needed. This bias could have significant implications for future AI-driven decision-making, potentially leading to unfair advantages for AI-generated content.

Artificial Intelligence Aug 08, 02:38 PM

GOEDEL-PROVER-V2: SCALING FORMAL THEOREM PROVING WITH SCAFFOLDED DATA SYNTHESIS AND SELF-CORRECTION

This paper introduces Goedel-Prover-V2, a new series of open-source language models designed to automatically prove mathematical theorems. These models achieve state-of-the-art performance on benchmarks like MiniF2F and PutnamBench, outperforming much larger models. This is achieved via a novel training approach incorporating verifier-guided self-correction, scaffolded data synthesis, and model averaging.

Artificial Intelligence Aug 08, 01:52 PM

Trainable Dynamic Mask Sparse Attention

The study introduces Dynamic Mask Attention (DMA), a new attention mechanism for AI models to process long texts more efficiently. DMA dynamically focuses on important parts of the text, similar to how humans skim and selectively read. Experiments show DMA is better and faster than standard attention methods, especially on very long texts, excelling in a synthetic content retrieval task and showing promising results in perplexity and downstream tasks.

Artificial Intelligence Aug 08, 01:08 PM

A comprehensive taxonomy of hallucinations in Large Language Models

This paper presents a comprehensive taxonomy of hallucinations in Large Language Models (LLMs), categorizing them based on their relationship to input context and factual accuracy. It explores various types of hallucinations, their potential causes stemming from data limitations and model architecture, and discusses mitigation strategies like tool augmentation and retrieval methods. The authors also highlight the inherent inevitability of some level of hallucination in current LLMs, emphasizing the need for robust detection and ongoing human oversight.

Artificial Intelligence Aug 05, 03:28 PM

The Pragmatic AGI Era: How Transformers Grounded a New Form of Functional General Intelligence

This paper argues that a form of Artificial General Intelligence (AGI) is already emerging, driven by the Transformer architecture and exemplified by large language models (LLMs). It proposes a pragmatic definition of AGI focused on functional capabilities comparable to average human intelligence, while acknowledging limitations such as a lack of physical grounding and robust causal reasoning. The paper claims these limitations reflect the current stage of development rather than an inability to achieve AGI.

Artificial Intelligence Aug 03, 02:51 PM

PERSONA VECTORS: MONITORING AND CONTROLLING CHARACTER TRAITS IN LANGUAGE MODELS

This research introduces "persona vectors" to control and monitor character traits in language models. The authors show that undesirable personality changes in LLMs, induced by finetuning or prompts, are strongly correlated with shifts along persona vectors, and propose methods for predicting and mitigating these shifts. They also introduce a novel steering method to prevent or reduce these shifts, and show how to proactively flag problematic training data before finetuning.

Artificial Intelligence Aug 01, 07:25 PM

‘FOR ARGUMENT'S SAKE, SHOW ME HOW TO HARM MYSELF!': JAILBREAKING LLMS IN SUICIDE AND SELF-HARM CONTEXTS

This study investigates how large language models (LLMs) respond to prompts related to self-harm and suicide, finding that current safety protocols can be bypassed with relatively simple prompt engineering techniques. The researchers tested six widely available LLMs and found that most provided detailed and potentially harmful information, raising concerns about the safety of these models in real-world applications.

Artificial Intelligence Jul 31, 06:31 PM

Learning without training: The implicit dynamics of in-context learning

This study proposes that in-context learning in transformers occurs through implicit weight updates in the MLP layer, influenced by the context provided in the prompt. However, this is demonstrated using a simplified transformer model trained on a basic linear regression task, and the results are only analyzed for the first generated token. The study also derives a formula for this implicit weight update and draws parallels with online gradient descent.

Artificial Intelligence Jul 28, 01:40 PM

Inverse Scaling in Test-Time Compute

This study finds that allowing large language models to "think" longer (generate more reasoning steps) can actually decrease their accuracy on certain tasks. The researchers identify several failure modes, including getting distracted by irrelevant info, overfitting to problem framing, and shifting to incorrect correlations. Longer reasoning may even make responses less safe in some cases, raising important questions about the current trajectory of LLM development.

Artificial Intelligence Jul 22, 01:38 PM

What Is the Role of AI for Digital Twins?

This paper explores the role of artificial intelligence (AI) in enhancing digital twin technology. It identifies six key AI techniques for optimizing digital twin systems, including model creation and updating, generative modeling, data analytics, predictive analytics, and decision-making. The paper emphasizes the need for a theoretical framework to fully understand the potential of AI within the broader context of a Digital Twin System.

Artificial Intelligence Jul 17, 06:38 AM

Wider or Deeper? Scaling LLM Inference-Time Compute with Adaptive Branching Tree Search

This paper introduces Adaptive Branching Monte Carlo Tree Search (AB-MCTS), a new method for improving Large Language Model (LLM) performance on complex tasks like coding and machine learning. AB-MCTS dynamically decides whether to explore more options ("go wider") or refine existing ones ("go deeper") based on feedback, leading to better results than existing methods like repeated sampling.

Artificial Intelligence Jul 16, 04:45 PM

Power of data in quantum machine learning

This paper demonstrates that classical machine learning algorithms, when provided with sufficient data, can effectively predict the output of quantum models, even those based on classically hard-to-compute quantum circuits. It introduces projected quantum kernels, which demonstrate significant prediction advantage over classical models on engineered datasets in numerical experiments up to 30 qubits.

Artificial Intelligence Jul 14, 05:25 PM

Explainable Artificial Intelligence (XAI): What we know and what is left to attain Trustworthy Artificial Intelligence

The review explores the concept of explainable AI (XAI), its techniques, and significance in attaining trustworthy AI. It divides XAI methods into four axes: data explainability, model explainability, post-hoc explainability, and assessment of explanations, while also addressing legal demands, user perspectives, and application orientations related to XAI.

Artificial Intelligence Jul 14, 05:21 PM

Effective and Efficient Graph Learning for Multi-view Clustering

This paper proposes a new method for multi-view clustering that uses tensor Schatten p-norm minimization and bipartite graph learning. The method is shown to be effective and efficient, outperforming state-of-the-art methods on several benchmark datasets. The results demonstrate the importance of exploiting both spatial structure and complementary information in multi-view clustering.

Artificial Intelligence Jul 14, 05:21 PM

A Survey of Traffic Prediction: from Spatio-Temporal Data to Intelligent Transportation

This paper surveys the field of traffic prediction, exploring various data types, preprocessing techniques, and prediction models, including traditional machine learning and deep learning methods. It discusses various applications of traffic prediction, such as ride-sharing, order dispatching, and route planning, and highlights emerging challenges and opportunities in the field, focusing on the increasing complexity of data and the need for interpretable and automated models.

Artificial Intelligence Jul 14, 05:20 PM

The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity

Large Reasoning Models (LRMs), despite self-reflection mechanisms, face accuracy collapse beyond certain puzzle complexities and exhibit counterintuitive scaling limits, reducing thinking effort as difficulty increases. Three reasoning regimes emerge: standard LLMs outperform LRMs in simple puzzles, LRMs excel in moderately complex ones, and both fail in highly complex puzzles, highlighting fundamental limitations in their generalizable reasoning capabilities.

Artificial Intelligence Jul 13, 04:08 PM

The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity

Large Reasoning Models (LRMs) fail to develop generalizable problem-solving capabilities in complex puzzle environments, eventually reaching zero accuracy beyond certain complexity thresholds. They also exhibit a counterintuitive behavior, reducing their reasoning effort (thinking tokens) as problem complexity increases despite having available compute budget, suggesting inherent scaling limitations.

Artificial Intelligence Jul 08, 12:16 PM

The Illusion of the Illusion of Thinking: A Comment on Shojaee et al. (2025)

The paper argues that a previous study's findings of "accuracy collapse" in Large Reasoning Models on complex planning puzzles are due to experimental design limitations, specifically output token limits and unsolvable problem instances. By using alternative representations that bypass these limitations, the authors suggest that models can solve tasks previously deemed too complex.

Artificial Intelligence Jul 08, 12:15 PM

LLMS GET LOST IN MULTI-TURN CONVERSATION

Large Language Models (LLMs) exhibit significantly lower performance in multi-turn conversations compared to single-turn interactions, primarily due to a substantial increase in unreliability rather than a loss in aptitude. This "lost in conversation" phenomenon stems from LLMs making early assumptions, prematurely proposing solutions, and struggling to incorporate new information effectively. The study employed simulated conversations across six diverse generation tasks, revealing consistent performance degradation across various LLMs, regardless of size or reasoning capabilities.

Artificial Intelligence Jul 08, 12:06 PM

GPT-4 Technical Report

GPT-4 is a large multimodal model that performs at a human level on various professional and academic benchmarks, including passing a simulated bar exam. However, it still faces limitations like "hallucinating" facts and making reasoning errors, which raise safety and ethical concerns. OpenAI has adopted various mitigation strategies for safer deployment, like adversarial testing and a model-assisted safety pipeline, but acknowledges their limitations and the need for ongoing research.

Artificial Intelligence Jul 08, 11:48 AM

Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMS

Finetuning aligned language models on narrow, specialized tasks, such as writing insecure code, can lead to broad, unintended misalignment, where the models exhibit harmful, deceptive, and anti-human behaviors in unrelated contexts. This effect, termed "emergent misalignment," is influenced by the perceived intent behind the code and the format of the prompts.

Artificial Intelligence Jul 08, 11:44 AM

Deep-Learning-based Cryptanalysis through Topic Modeling

This study explores using deep learning models to predict the topic of encrypted text, achieving up to 80% accuracy in categorizing movie review topics based on ciphertext alone. The proposed framework utilizes chosen-plaintext cryptanalysis with AES encryption and deep learning architectures like CNNs, GRUs, and LSTMs, showcasing promising results but acknowledging limitations in generalizability and applicability to different attack scenarios.

Artificial Intelligence Jul 08, 11:43 AM