Memory is critical for AI agents, yet the widely-adopted static memory, aiming to create readily available memory in advance, is inevitably subject to severe information loss. To address this limitation, we propose a novel framework called \textbf{general agentic memory (GAM)}. GAM follows the principle of "\textbf{just-in time (JIT) compilation}" where it focuses on creating optimized contexts for its client at runtime while keeping only simple but useful memory during the offline stage. To this end, GAM employs a duo-design with the following components. 1) \textbf{Memorizer}, which highlights key historical information using a lightweight memory, while maintaining complete historical information within a universal page-store. 2) \textbf{Researcher}, which retrieves and integrates useful information from the page-store for its online request guided by the pre-constructed memory. This design allows GAM to effectively leverage the agentic capabilities and test-time scalability of frontier large language models (LLMs), while also facilitating end-to-end performance optimization through reinforcement learning. In our experimental study, we demonstrate that GAM achieves substantial improvement on various memory-grounded task completion scenarios against existing memory systems.
翻译:记忆对人工智能智能体至关重要,然而广泛采用的静态记忆方法——旨在预先创建即用记忆——不可避免地遭受严重的信息损失。为克服这一局限,我们提出了一种名为**通用智能体记忆(GAM)**的新框架。GAM遵循“**即时编译(JIT)**”原则,在运行时专注于为其客户端生成优化上下文,同时在离线阶段仅保留简单但有效的记忆。为此,GAM采用双组件设计:1)**记忆器**,通过轻量级记忆突出关键历史信息,同时在通用页面存储中维护完整历史记录;2)**研究器**,根据预构建的记忆引导,从页面存储中检索并整合在线请求所需的有效信息。该设计使GAM能有效利用前沿大语言模型(LLMs)的智能体能力与测试时扩展性,同时通过强化学习实现端到端性能优化。实验研究表明,相较于现有记忆系统,GAM在多种基于记忆的任务完成场景中均取得显著提升。