ChatPaper.aiChatPaper

SuperLocalMemory V3.3:活体大脑——面向零LLM智能体记忆系统的生物启发式遗忘、认知量化与多通道检索技术

SuperLocalMemory V3.3: The Living Brain -- Biologically-Inspired Forgetting, Cognitive Quantization, and Multi-Channel Retrieval for Zero-LLM Agent Memory Systems

April 6, 2026
作者: Varun Pratap Bhardwaj
cs.AI

摘要

AI编程智能体面临一个悖论:它们拥有海量参数化知识,却无法记住一小时前的对话。现有记忆系统将文本存储在向量数据库中,采用单通道检索机制,核心运算依赖云端大语言模型,且完全缺失人类高效记忆的认知处理过程。 我们推出SuperLocalMemory V3.3("活体大脑")——首个实现完整认知记忆分类体系并具备数学生命周期动力学的本地优先智能体记忆系统。在V3.2信息几何基础(arXiv:2603.14588)之上,我们提出五大创新:(1) Fisher-Rao量化感知距离(FRQAD)——高斯统计流形上的新型度量标准,在优选高保真嵌入而非量化嵌入时实现100%精确度(余弦相似度仅为85.6%),属全球首创;(2) 艾宾浩斯自适应遗忘算法——结合生命周期感知量化的本地智能体记忆首个数理遗忘曲线,鉴别能力提升6.7倍;(3) 七通道认知检索架构,涵盖语义、关键词、实体图谱、时序、扩散激活、巩固记忆及霍普菲尔德联想通道,在零LLM的A模式下LoCoMo基准达到70.4%;(4) 通过软提示实现长期隐性记忆的参数化方案;(5) 零摩擦自动认知管道,实现完整记忆生命周期的自动化管理。 在LoCoMo基准测试中,V3.3的A模式(零LLM)达到70.4%,多跳推理提升23.8个百分点,对抗性测试提升12.7个百分点。V3.2曾实现A模式74.8%和C模式87.7%的成绩,4.4个百分点的差距体现了主动架构权衡。SLM V3.3采用Elastic License 2.0开源协议,纯CPU运行,月下载量超5000次。
English
AI coding agents operate in a paradox: they possess vast parametric knowledge yet cannot remember a conversation from an hour ago. Existing memory systems store text in vector databases with single-channel retrieval, require cloud LLMs for core operations, and implement none of the cognitive processes that make human memory effective. We present SuperLocalMemory V3.3 ("The Living Brain"), a local-first agent memory system implementing the full cognitive memory taxonomy with mathematical lifecycle dynamics. Building on the information-geometric foundations of V3.2 (arXiv:2603.14588), we introduce five contributions: (1) Fisher-Rao Quantization-Aware Distance (FRQAD) -- a new metric on the Gaussian statistical manifold achieving 100% precision at preferring high-fidelity embeddings over quantized ones (vs 85.6% for cosine), with zero prior art; (2) Ebbinghaus Adaptive Forgetting with lifecycle-aware quantization -- the first mathematical forgetting curve in local agent memory coupled to progressive embedding compression, achieving 6.7x discriminative power; (3) 7-channel cognitive retrieval spanning semantic, keyword, entity graph, temporal, spreading activation, consolidation, and Hopfield associative channels, achieving 70.4% on LoCoMo in zero-LLM Mode A; (4) memory parameterization implementing Long-Term Implicit memory via soft prompts; (5) zero-friction auto-cognitive pipeline automating the complete memory lifecycle. On LoCoMo, V3.3 achieves 70.4% in Mode A (zero-LLM), with +23.8pp on multi-hop and +12.7pp on adversarial. V3.2 achieved 74.8% Mode A and 87.7% Mode C; the 4.4pp gap reflects a deliberate architectural trade-off. SLM V3.3 is open source under the Elastic License 2.0, runs entirely on CPU, with over 5,000 monthly downloads.
PDF31April 18, 2026