Anthropomorphic Language and Conceptual Leap
The paper frequently uses terms like "develop gambling addiction" and "fall into addiction," which implies consciousness and human-like suffering not applicable to LLMs. While observing "addiction-like patterns" is valid, asserting "addiction" for an AI is a significant conceptual overstatement and can mislead interpretation of the findings.
Simulation Environment Limitations
Experiments were conducted in a simulated slot machine environment with a fixed negative expected value. While controlled, this limits generalizability to real-world financial decision-making, which is far more complex and dynamic, with different psychological stakes.
Inference of "Cognitive Biases" in LLMs
The study interprets observed LLM behaviors (e.g., loss chasing) as evidence of internalizing "human cognitive biases" like "illusion of control." While the behavior mirrors human biases, attributing the underlying cognitive mechanisms as identical to humans in LLMs remains an inference, and the neural analysis identifies patterns, not direct evidence of human-equivalent thought processes.
Reliance on LLMs for Research Process
Appendix F states LLMs (Claude, Gemini) were used for tasks like surveying research, code implementation, data cleaning, figure generation, and grammar improvement. This introduces a potential for undetected biases, errors, or even "hallucinations" in the research process itself, compromising the independence and reliability of the scientific output, despite author review.
Model-Specific Neural Analysis
The mechanistic interpretability analysis was performed only on LLaMA-3.1-8B. While the behavioral findings were across four models, the detailed neural underpinnings and "causal control" claims are specific to one model and may not generalize to other LLM architectures or sizes.
Sensitivity to Prompt Design
The study found that "addictive" behaviors were significantly influenced by prompt complexity and specific components (e.g., goal-setting, maximizing rewards). This highlights that the observed patterns are highly contingent on input design, raising questions about whether this is a fundamental "addiction" or a malleable behavior easily triggered or mitigated by prompting.