Cooperative Memory Paging with Keyword Bookmarks for Long-Horizon LLM Conversations

📰 ArXiv cs.AI

arXiv:2604.12376v1 Announce Type: cross Abstract: When LLM conversations grow beyond the context window, old content must be evicted -- but how does the model recover it when needed? We propose cooperative paging: evicted segments are replaced with minimal keyword bookmarks ([pN:keywords], ~8-24 tokens each), and the model is given a recall() tool to retrieve full content on demand. On the LoCoMo benchmark (10 real multi-session conversations, 300+ turns), cooperative paging achieves the highest

Published 15 Apr 2026
Read full paper → ← Back to Reads