Adaptive Memory Admission Control for LLM Agents

AuthorsGuilin Zhang, Wei Jiang, Xiejiashan Wang et al.

2026

TL;DR

A-MAC uses five interpretable admission signals with a learned linear policy to reach F1 0.583 on LoCoMo while cutting latency 31% vs A-mem.

SharePost on XLinkedIn

Read our summary here, or open the publisher PDF on the next tab.

THE PROBLEM

LLM agents lack controllable long term memory admission

LLM-based agents either accumulate large volumes of conversational content, including hallucinated or obsolete facts, or depend on opaque, fully LLM-driven memory policies.

This makes long-term memory a weakly controlled component, so agents like MemGPT and A-mem risk memory bloat, hallucination propagation, and degraded retrieval latency over time.

HOW IT WORKS

Adaptive Memory Admission Control A-MAC

A-MAC treats memory admission as a decision problem, combining Utility, Confidence, Novelty, Recency, and Type Prior via a learned linear score and Algorithm 1 A-MAC Memory Admission.

Think of A-MAC like an OS page scheduler plus a librarian: it checks usefulness, trustworthiness, redundancy, freshness, and category before shelving any memory.

This explicit admission control lets A-MAC keep long-term memory compact, reliable, and auditable in ways a plain context window or fully LLM-native policy cannot.

DIAGRAM

End to end admission flow for a single memory candidate

This diagram shows how A-MAC processes a candidate memory from extraction through feature computation to admit, update, or reject decisions.

DIAGRAM

LoCoMo evaluation and ablation pipeline

This diagram shows how A-MAC is trained and evaluated on LoCoMo with cross validated weight learning and feature ablations.

PROCESS

How A-MAC Handles a Multi Turn Conversation Session

  1. 01

    Memory admission as a decision problem

    A-MAC takes conversation history H and existing memory store M, extracts atomic candidate memories m, and frames admission as scoring with S(m).

  2. 02

    Interpretable memory value signals

    A-MAC computes Utility U with a single LLM call and derives Confidence C, Novelty N, Recency R, and Type Prior T using rule based features.

  3. 03

    Policy learning and admission rule

    A-MAC learns weights ω and threshold θ via 5 fold cross validation and grid search to maximize F1 on labeled admission decisions.

  4. 04

    Algorithm 1 A-MAC Memory Admission

    A-MAC applies Algorithm 1 to aggregate features, compare S(m) to θ, detect conflicts in M, and either admit, merge, or reject each candidate.

KEY CONTRIBUTIONS

Key Contributions

  • 01

    Adaptive Memory Admission Control A-MAC

    A-MAC formalizes memory admission as a structured decision problem with a linear score over Utility, Confidence, Novelty, Recency, and Type Prior, implemented in Algorithm 1 A-MAC Memory Admission.

  • 02

    Interpretable five factor value signals

    A-MAC decomposes memory value into Utility U, Confidence C, Novelty N, Recency R, and Type Prior T, enabling explicit control over reliability, redundancy, and persistence.

  • 03

    Efficient hybrid design and empirical gains

    A-MAC reaches F1 0.583 with 2644 ms latency on LoCoMo, improving F1 by 0.042 and cutting latency by 31% compared to A-mem while identifying Type Prior as the dominant feature.

RESULTS

By the Numbers

F1

0.583

+0.042 over A-mem

Prec.

0.417

+0.046 over A-mem

Recall

0.972

-0.028 vs A-mem

Lat.(ms)

2644

-1187 ms vs A-mem

On the LoCoMo benchmark with 225 test samples, A-MAC improves F1 from 0.541 to 0.583 over A-mem while reducing latency from 3831 ms to 2644 ms. This shows A-MAC can be both more accurate and more efficient than fully LLM native memory systems for long term agent memory admission.

BENCHMARK

By the Numbers

On the LoCoMo benchmark with 225 test samples, A-MAC improves F1 from 0.541 to 0.583 over A-mem while reducing latency from 3831 ms to 2644 ms. This shows A-MAC can be both more accurate and more efficient than fully LLM native memory systems for long term agent memory admission.

BENCHMARK

Performance comparison on LoCoMo test set

F1 on LoCoMo for A-MAC and baseline memory admission policies.

BENCHMARK

Ablation study showing performance impact of removing each feature

F1 on LoCoMo when dropping each A-MAC feature from the admission score.

KEY INSIGHT

The Counterintuitive Finding

Removing Type Prior drops A-MAC’s F1 from 0.583 to 0.476, a 0.107 decrease that is larger than removing any other feature.

This is surprising because many builders assume semantic Utility or Novelty dominate, but A-MAC shows simple content category priors can be the strongest admission signal.

WHY IT MATTERS

What this unlocks for the field

A-MAC unlocks controllable, interpretable long term memory admission where every stored fact is scored along Utility, Confidence, Novelty, Recency, and Type Prior.

Builders can now tune and audit memory policies per domain, trading precision, recall, and latency explicitly instead of relying on opaque LLM heuristics or unbounded memory growth.

~11 min read← Back to papers

Related papers

Agent MemoryLong-Term Memory

Agentic Memory: Learning Unified Long-Term and Short-Term Memory Management for Large Language Model Agents

Yi Yu, Liuyi Yao et al.

arXiv 2026 · 2026

Agentic Memory (AgeMem) exposes memory management tools, a three-stage progressive RL strategy, and step-wise GRPO directly inside the agent policy to jointly control long-term and short-term memory. On Qwen3-4B-Instruct, AgeMem attains 54.31% average performance across ALFWorld, SciWorld, PDDL, BabyAI, and HotpotQA, exceeding the best baseline A-Mem at 45.74%.

Long-Term Memory

AlpsBench: An LLM Personalization Benchmark for Real-Dialogue Memorization and Preference Alignment

Jianfei Xiao, Xiang Yu et al.

· 2026

AlpsBench combines Personalized Information Extraction, Personalized Information Update, Personalized Information Retrieval, and Personalized Information Utilization over 2,500 WildChat dialogues with human-verified structured memories. AlpsBench shows, for example, that Gemini-3 Flash scores 51.67 on Task 1 Extraction while DeepSeek Reasoner reaches 0.9569 retrieval recall with 100 distractors on AlpsBench.

Questions about this paper?

Paper: Adaptive Memory Admission Control for LLM Agents

Answers use this explainer on Memory Papers.

Checking…