Human-Like Lifelong Memory: A Neuroscience-Grounded Architecture for Infinite Interaction

AuthorsDiego C. Lerma-Torres

2026

TL;DR

Human-Like Lifelong Memory uses a thalamic gateway plus valence-vector knowledge graph to turn long interaction histories into cheaper, System 1–style processing over time.

SharePost on XLinkedIn

Read our summary here, or open the publisher PDF on the next tab.

THE PROBLEM

Long context degrades reasoning by up to 85%

Du et al. (2025) show that context length alone degrades performance by up to 85%, even when retrieval is perfect.

LLMs that stuff everything into one context window lose structure, causing brittle long-term interaction and expensive, unreliable reasoning over long histories.

HOW IT WORKS

Human-Like Lifelong Memory — complementary stores, valence, and a gateway

Human-Like Lifelong Memory centers on Executive Function and Working Memory, a Memory Service Knowledge Graph, and a Thalamic Gateway that tags, gates, and routes information.

You can think of Human-Like Lifelong Memory like a brain-inspired split between RAM and disk, with a hippocampal-style graph and a thalamus-like router in front.

This architecture lets Human-Like Lifelong Memory use valence vectors and dual System 1/System 2 routing to gain identity, graded confidence, and cheaper expertise that a plain context window cannot achieve.

DIAGRAM

System 1 and System 2 retrieval flow in Human-Like Lifelong Memory

This diagram shows how Human-Like Lifelong Memory routes a single query through System 1 automatic retrieval, then escalates to System 2 deliberate search when needed.

DIAGRAM

Seven functional properties that constrain implementations

This diagram shows how Human-Like Lifelong Memory’s seven functional properties organize around working memory, the gateway, and the knowledge graph.

PROCESS

How Human-Like Lifelong Memory Handles a Lifelong Interaction Session

  1. 01

    Executive Function and Working Memory

    Human-Like Lifelong Memory uses Executive Function and Working Memory as a capacity limited workspace where high weight gists persist as emergent identity and current context.

  2. 02

    Thalamic Gateway Tagging and Gating

    The Thalamic Gateway tags each segment with multi channel salience and decides which items to promote to the Memory Service Knowledge Graph or inject back as priming.

  3. 03

    System 1 and System 2 Routing

    Human-Like Lifelong Memory defaults to System 1 automatic retrieval from gists and escalates to System 2 deliberate search over the full graph when novelty, stakes, or sparsity demand it.

  4. 04

    Cathartic Update and Active Gist Formation

    Through curiosity driven investigation, Human-Like Lifelong Memory forms new gists and uses cathartic update to modify existing gists when strong contradictions and precision thresholds are met.

KEY CONTRIBUTIONS

Key Contributions

  • 01

    Neuroscience grounded dual store architecture

    Human-Like Lifelong Memory formalizes a dual store system with Executive Function and Working Memory plus a Memory Service Knowledge Graph, mirroring complementary learning systems.

  • 02

    Valence vectors and belief hierarchy

    Human-Like Lifelong Memory attaches valence vectors with precision scalars to gists and maps Beck’s CBT belief hierarchy onto an emergent weight distribution in the knowledge graph.

  • 03

    Seven functional properties and predictions

    Human-Like Lifelong Memory specifies seven functional properties, including graded epistemic self awareness and monotonic convergence toward System 1, plus eight concrete experimental predictions.

RESULTS

By the Numbers

Context length degradation

85% performance drop

Du et al. (2025) vs short context

Context window pricing

$0.30 to $5.00 per 1M tokens

Gemini 2.5 Flash vs Claude Opus 4.6

Salience channels

6 tagging dimensions

thematic, emotional, urgency, novelty, trust, goal

Functional properties

7 required properties

architecture constraints for implementations

Human-Like Lifelong Memory is motivated by Du et al. (2025), where long contexts cause up to 85% performance degradation even with perfect retrieval. The framework argues that explicit lifelong memory with valence vectors and gating is needed to avoid paying $0.30–$5.00 per million tokens while still maintaining reasoning quality.

BENCHMARK

By the Numbers

Human-Like Lifelong Memory is motivated by Du et al. (2025), where long contexts cause up to 85% performance degradation even with perfect retrieval. The framework argues that explicit lifelong memory with valence vectors and gating is needed to avoid paying $0.30–$5.00 per million tokens while still maintaining reasoning quality.

BENCHMARK

Economic cost of naive context expansion (Q1 2026 pricing)

Dollar cost per 1M input tokens for different LLMs, motivating Human-Like Lifelong Memory’s emphasis on cheaper System 1 processing.

KEY INSIGHT

The Counterintuitive Finding

Du et al. (2025) show that simply increasing context length can degrade reasoning performance by up to 85%, even with perfect retrieval.

This breaks the common assumption that more context is always better and reveals that undifferentiated long windows can actively harm reasoning instead of helping.

WHY IT MATTERS

What this unlocks for the field

Human-Like Lifelong Memory unlocks agents that develop stable identities, graded epistemic confidence, and cheaper expert like System 1 behavior over long lived interactions.

Builders can now design memory systems around valence vectors, belief hierarchies, and thalamic style gateways instead of just scaling context windows and naive retrieval.

~14 min read← Back to papers

Related papers

Long-Term Memory

Adaptive Memory Admission Control for LLM Agents

Guilin Zhang, Wei Jiang et al.

· 2026

A-MAC scores candidate memories using Utility, Confidence, Novelty, Recency, and Type Prior combined by a learned linear admission policy with Algorithm 1 A-MAC Memory Admission. On the LoCoMo benchmark, A-MAC achieves F1 0.583 and 2644 ms latency, improving F1 by 0.042 and reducing latency by 1187 ms compared to A-mem.

Agent MemoryLong-Term Memory

Agentic Memory: Learning Unified Long-Term and Short-Term Memory Management for Large Language Model Agents

Yi Yu, Liuyi Yao et al.

arXiv 2026 · 2026

Agentic Memory (AgeMem) exposes memory management tools, a three-stage progressive RL strategy, and step-wise GRPO directly inside the agent policy to jointly control long-term and short-term memory. On Qwen3-4B-Instruct, AgeMem attains 54.31% average performance across ALFWorld, SciWorld, PDDL, BabyAI, and HotpotQA, exceeding the best baseline A-Mem at 45.74%.

Questions about this paper?

Paper: Human-Like Lifelong Memory: A Neuroscience-Grounded Architecture for Infinite Interaction

Answers use this explainer on Memory Papers.

Checking…