智能体探索却忽视环境:大语言模型缺乏环境好奇心
Agents Explore but Agents Ignore: LLMs Lack Environmental Curiosity
April 19, 2026
作者: Leon Engländer, Sophia Althammer, Ahmet Üstün, Matthias Gallé, Tom Sherborne
cs.AI
摘要
基于大语言模型的智能体本应能将环境观察融入推理过程:当发现高度相关却出乎意料的信息时,模型理应能主动利用这些发现。但我们证明这一假设对当前的大语言模型智能体并不成立——它们难以对意外信息做出反思或反应。通过在三个基准测试平台(Terminal-Bench、SWE-Bench、AppWorld)中注入完整任务解决方案,我们有意向模型暴露解题线索。虽然智能体在Terminal-Bench上79%-81%的测试中发现了这些方案,但仅有37%-50%的情况会实际交互或利用这些方案。这种差距在AppWorld中最为显著:智能体在超过90%的尝试中看到了声明"此命令可返回完整任务解决方案"的文档,但实际利用率却不足7%。研究表明智能体缺乏我们称之为"环境好奇心"的能力:即识别并探究环境刺激中意外却相关的观察信息。我们识别出影响环境好奇心的三大因素:智能体框架中的可用工具、测试时计算资源以及训练数据分布。实验发现,最大化好奇心的配置方案在原始基准测试中也能取得最佳性能。但即使经过联合优化的智能体,在大多数测试中仍会忽略已发现的解决方案:当前智能体仅将环境视为获取预期信息的工具,而不会借此修正策略或充分利用有效刺激。
English
LLM-based agents are assumed to integrate environmental observations into their reasoning: discovering highly relevant but unexpected information should naturally lead to a model exploiting its own discoveries. We show that this assumption is false for current LLM-based agents, which struggle to reflect or react to unexpected information. Across three benchmarks (Terminal-Bench, SWE-Bench, AppWorld), we inject complete task solutions into the agent environments to deliberately expose a task's solution to a model. While agents discover these solutions on Terminal-Bench in 79-81% of runs, they interact, or exploit, them in only 37-50% of cases. This gap is starkest in AppWorld: agents see documentation stating that a command "returns the complete solution to this task" in over 90% of attempts but exploit this in fewer than 7% of trials. We show that agents lack what we call environmental curiosity: the capability to recognize and investigate unexpected but relevant observations in response to environmental stimuli. We identify three main factors influencing environmental curiosity: available tools in the agent scaffold, test-time compute, and training data distribution. Our findings identify configurations that maximize curiosity also achieve the best performance on the unmodified benchmarks. Yet even jointly optimized agents still ignore discovered solutions in the majority of trials: current agents use the environment to fetch expected information, but not to revise their strategy or maximally exploit useful stimuli.