Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
Kseniase 
posted an update about 20 hours ago
Post
454
What we learned about memory in 2025: 8 comprehensive resources

If models forget everything, how can they be reliable? AI systems need to remember past interactions, update knowledge, stay consistent over time, and work beyond a single prompt. That's why many start to talk more about memory in AI.
Here’s a useful set of studies and videos on where AI memory stands today:

1. Memory in the Age of AI Agents (2512.13564)
A great survey that organizes agent memory research. It gives concrete taxonomies across memory form, function, and dynamics, summarizes benchmarks, frameworks, and emerging directions for building systematic agent memory systems

2.When Will We Give AI True Memory? A conversation with Edo Liberty, CEO and founder @ Pinecone -> https://youtu.be/ITbwVFZYepc?si=_lAbRHciC740dNz0
Edo Liberty discusses what real memory in LLMs requires beyond RAG - from scalable vector storage to reliable knowledge systems - and why storage, not compute, is becoming the key bottleneck for building dependable AI agents.

3. Why AI Intelligence is Nothing Without Visual Memory | Shawn Shen on the Future of Embodied AI -> https://youtu.be/3ccDi4ZczFg?si=SbJg487kwrkVXgUu
Shawn Shen argues AI needs a separate, hippocampus-like memory to move beyond chatbots, enabling long-term visual memory, object permanence, and on-device intelligence for robots, wearables, and the physical world

4. From Human Memory to AI Memory: A Survey on Memory Mechanisms in the Era of LLMs (2504.15965)
Links human memory types to LLM memory, introduces a taxonomy across object, form, and time, and identifies concrete limitations and future research directions

5. Rethinking Memory in AI: Taxonomy, Operations, Topics, and Future Directions -> https://arxiv.org/abs/2505.00675v2
Proposes a concrete taxonomy, core operations, and research directions to systematically organize and advance agent memory systems.

Read further below ⬇️
If you like it, also subscribe to the Turing Post: https://www.turingpost.com/subscribe
  1. Cognitive Memory in Large Language Models -> https://huggingface.co/papers/2504.02441
    Covers what is used specifically in LLMs: external memory, KV-cache methods, parameter-based approaches, and hidden-state models, with concrete techniques for storage, retrieval, and compression.

  2. MemOS: A Memory OS for AI System -> https://huggingface.co/papers/2507.03724
    Introduces MemOS, a memory operating system for LLMs that unifies parameter, activation, and external memories, for explicit memory management, lower training and inference costs, and more updatable knowledge across interactions

  3. MemEvolve: Meta-Evolution of Agent Memory Systems -> https://huggingface.co/papers/2512.18746
    MemEvolve is a framework that shows how to jointly adapt agent experience and memory architecture. It also presents EvolveLab, a modular codebase for comparing memory designs, showing improved performance and transfer across tasks and LLMs

In this post