NextFin

China's EverMind Unveils Open-Source Long-Term Memory System for AI Agents

Summarized by NextFin AI
  • EverMemOS is a new long-term memory operating system for AI agents, designed to provide durable and coherent memory, addressing the limitations of fixed context windows in current AI models.
  • It has achieved impressive performance metrics, scoring 92.3% on LoCoMo and 82% on LongMemEval-S, establishing itself as the new state of the art in long-term memory benchmarks.
  • The system's architecture is inspired by human brain mechanisms, incorporating a four-layer design that enhances AI's ability to remember, adapt, and evolve over time.
  • EverMemOS aims to fill the gap in the market for a unified, high-performance memory solution that supports personalized experiences and complex multi-user workflows.

EverMind, an incubation project under the research team of Shanda Investment Group, the global investment firm founded by Chinese billionaire Chen Tianqiao, has launched an open-source long-term memory system intended to serve as foundational data infrastructure for next-generation AI agents.

The system, EverMemOS, draws inspiration from human memory processes, spanning sensory encoding, hippocampal indexing, and cortical long-term storage, with the hippocampus and prefrontal cortex jointly forming and retrieving memories, the team announced. This “brain-like” architecture is central to the system’s design, enabling AI models to think, store experiences, and develop in a manner closer to human cognition.

Long-term memory is emerging as a strategic feature for leading AI platforms such as Claude and ChatGPT, said Chen, who also co-founded the Tianqiao and Chrissy Chen Institute. Memory, he added, will become the core differentiator and inflection point for future AI applications — the key to transforming AI from a reactive tool into an autonomous, evolving intelligent agent.

Chen Tianqiao

At the inaugural Symposium for AI Accelerated Science (AIAS 2025), held by the Tianqiao and Chrissy Chen Institute (TCCI) in San Francisco on Oct. 27–28, a leading researcher outlined five core capabilities of what he described as “discovery-based intelligence,” including long-term memory.

He argued that today’s mainstream AI systems are built on a “spatial-structure paradigm,” one that is instantaneous and static, relying on massive parameter scaling to fit snapshots of the world. In contrast, the human brain operates under a “temporal-structure paradigm” that is continuous and dynamic, aiming to manage and predict information as it unfolds over time. Within this framework, long-term memory serves as the essential bridge linking time and intelligence.

Inspired by this principle, the team introduced EverMemOS, a system designed to give AI continuity across time. The technology enables models to remember, adapt, and evolve within a temporal stream, representing what the institute views as a foundational step toward more human-like, time-aware artificial intelligence.

Against this backdrop, the EverMind team launched EverMemOS, a memory system that has achieved key breakthroughs in both application scenarios and technological performance.

  • In terms of application scenarios: It is the industry's first memory system truly capable of supporting both one-on-one conversations and complex multi-user collaboration. The innovative AI Native product Tanka has already adopted it as an early adopter.

Chen Tianqiao

  • In terms of technological performance: Leveraging an innovative bio-inspired “engram” heuristic for memory retrieval and application, EverMemOS achieved impressively high scores on the most prominent long-term memory evaluation datasets: 92.3% on LoCoMo and 82% on LongMemEval-S. Both scores significantly surpass the SOTA (State-of-the-Art) benchmarks, setting new industry standards.

EverMemOS Four-Layer Architecture

Inspired by the memory mechanisms of the human brain, EverMemOS introduces an innovative four-layer architecture, drawing analogies to the brain’s key functional areas:

  • Agentic Layer — Responsible for task understanding, decomposition, and generation. This can be compared to the role of the "prefrontal cortex" in attention, planning, and executive control.
  • Memory Layer — Manages the extraction and structured storage of long-term memory, corresponding to the function of the "cerebral cortical network" in long-term consolidation and storage.
  • Index Layer — Connects and efficiently retrieves memories using embeddings, key-value pairs, and knowledge graphs. This resembles the "hippocampus," which is responsible for associating and rapidly indexing memories.
  • API/MCP Interface Layer — Seamlessly integrates with enterprise applications, serving as the AI's sensory interface with the external environment.

Chen Tianqiao

Three Core Features of EverMemOS:

Feature 1: From a “Memory Database” to a “Memory Processor”. EverMemOS’s foremost innovation is that it functions not just as a repository for memory, but as a processor for memory applications. It addresses the core limitation of existing solutions that “just find data without actually using it.” Through its unique reasoning and integration mechanisms, EverMemOS enables memories to actively and instantly shape the model’s thought processes and responses, ensuring that every statement made by the AI is grounded in a long-term understanding of the user. This delivers truly coherent and personalized interactions.

Feature 2: Innovative Hierarchical Memory Extraction and Dynamic Organization. At the heart of EverMemOS lies its novel approach to “hierarchical memory extraction.” Instead of treating memory as a chaotic collection of text fragments, it extracts continuous semantic blocks as contextual memory units and dynamically organizes them into structured memories. This layered memory organization links related memories together, overcoming the limitations of pure text-based similarity searches that struggle to capture implicit context, and lays a solid foundation for advanced memory utilization.

Feature 3: The Industry’s First Scalable Modular Memory Framework. In real-world applications, memory requirements can vary greatly in different scenarios. To address this, EverMemOS has introduced an innovative, scenario-based scalable memory framework. It flexibly supports various types of memory needs—whether high-precision, structured information is required for work scenarios, or empathy and understanding of implicit emotions are needed for companionship scenarios, EverMemOS can intelligently provide the optimal strategy for organizing and utilizing memory. This effectively overcomes the limitations of traditional single-form memory solutions, meeting the ever-changing needs of diverse environments.

Currently, EverMind has released the open-source version of EverMemOS on GitHub for developers and AI teams to deploy and try out. You can access at: https://github.com/EverMind-AI/EverMemOS/. The team plans to launch a cloud service version later this year, offering enterprise users more robust technical support, persistent data storage, and a scalable experience. Interested developers or businesses can leave their email on the official website (http://everm.com) for a chance to be among the first to try the new service.

Explore more exclusive insights at nextfin.ai.

Insights

What are the key features of EverMemOS that differentiate it from traditional memory systems?

How does the four-layer architecture of EverMemOS relate to human brain functions?

What performance benchmarks has EverMemOS achieved compared to previous systems?

What are the main limitations of current large language models that EverMemOS aims to address?

How does EverMemOS integrate neuroscience principles into its design?

What are the potential applications of EverMemOS in enterprise environments?

What challenges does the AI industry face in adopting long-term memory systems?

How does EverMemOS improve the personalization and coherence of AI interactions?

What recent developments or announcements have been made regarding EverMemOS?

How does the introduction of EverMemOS impact the competitive landscape of AI memory systems?

What role does long-term memory play in the evolution of AI towards higher intelligence?

How does EverMemOS's memory processor function differ from traditional memory databases?

What feedback have early adopters of EverMemOS provided?

How does EverMemOS compare to existing memory solutions like RAG pipelines?

What are the industry trends regarding memory integration in AI systems?

How can developers leverage the open-source version of EverMemOS for their projects?

What are the implications of memory systems like EverMemOS for the future of AI agents?

What does the term 'discovery intelligence' mean in the context of AI development?

How might EverMemOS evolve in response to changing user needs and technological advancements?

What controversies or debates exist regarding the integration of memory in AI systems?

Search
NextFinNextFin
NextFin.Al
No Noise, only Signal.
Open App