ChatPaper.aiChatPaper

记忆重于存储:以事件为核心构建智能体搜索与推理的逻辑图谱

Memory Matters More: Event-Centric Memory as a Logic Map for Agent Searching and Reasoning

January 8, 2026
作者: Yuyang Hu, Jiongnan Liu, Jiejun Tan, Yutao Zhu, Zhicheng Dou
cs.AI

摘要

大型语言模型(LLMs)正越来越多地被部署为能够推理、规划并与环境交互的智能体。为有效适应长周期场景,此类智能体的关键能力在于具备可存储、组织并调用过往经验以支持下游决策的记忆机制。然而现有方法大多以扁平化方式组织存储记忆,并依赖简单的基于相似度的检索技术。即便引入结构化记忆,现有方法仍难以显式捕捉经验或记忆单元间的逻辑关系。此外,记忆访问机制与构建的结构严重脱节,仍停留于浅层语义检索层面,阻碍了智能体对长周期依赖关系进行逻辑推理。本研究提出以事件为中心的记忆框架CompassMem,其设计灵感源自事件分割理论。该框架通过将经验增量式分割为事件单元,并以显式逻辑关系链接形成事件图谱,使记忆系统成为逻辑地图,让智能体能够超越表层检索进行结构化、目标导向的记忆导航,逐步积累有价值记忆以支持长周期推理。在LoCoMo和NarrativeQA数据集上的实验表明,CompassMem在多种骨干模型中持续提升了检索与推理性能。
English
Large language models (LLMs) are increasingly deployed as intelligent agents that reason, plan, and interact with their environments. To effectively scale to long-horizon scenarios, a key capability for such agents is a memory mechanism that can retain, organize, and retrieve past experiences to support downstream decision-making. However, most existing approaches organize and store memories in a flat manner and rely on simple similarity-based retrieval techniques. Even when structured memory is introduced, existing methods often struggle to explicitly capture the logical relationships among experiences or memory units. Moreover, memory access is largely detached from the constructed structure and still depends on shallow semantic retrieval, preventing agents from reasoning logically over long-horizon dependencies. In this work, we propose CompassMem, an event-centric memory framework inspired by Event Segmentation Theory. CompassMem organizes memory as an Event Graph by incrementally segmenting experiences into events and linking them through explicit logical relations. This graph serves as a logic map, enabling agents to perform structured and goal-directed navigation over memory beyond superficial retrieval, progressively gathering valuable memories to support long-horizon reasoning. Experiments on LoCoMo and NarrativeQA demonstrate that CompassMem consistently improves both retrieval and reasoning performance across multiple backbone models.
PDF32January 13, 2026