UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Human-inspired Episodic Memory for Infinite Context LLMs

Fountas, Z; Benfeghoul, MA; Oomerjee, A; Christopoulou, F; Lampouras, G; Bou-Ammar, H; Wang, J; (2025) Human-inspired Episodic Memory for Infinite Context LLMs. In: 13th International Conference on Learning Representations ICLR 2025. ICLR: Singapore. Green open access

[thumbnail of 12669_Human_inspired_Episodic_.pdf]
Preview
PDF
12669_Human_inspired_Episodic_.pdf - Accepted Version

Download (1MB) | Preview

Abstract

Large language models (LLMs) have shown remarkable capabilities, but still struggle with processing extensive contexts, limiting their ability to maintain coherence and accuracy over long sequences. In contrast, the human brain excels at organising and retrieving episodic experiences across vast temporal scales, spanning a lifetime. In this work, we introduce EM-LLM, a novel approach that integrates key aspects of human episodic memory and event cognition into LLMs with no fine-tuning, enabling them to handle practically infinite context lengths while maintaining computational efficiency. EM-LLM organises sequences of tokens into coherent episodic events using a combination of Bayesian surprise and graph-theoretic boundary refinement in an online fashion. When needed, these events are retrieved through a two-stage memory process, combining similarity-based and temporally contiguous retrieval for efficient, human-inspired access to relevant information. Experiments on the LongBench and ∞-Bench benchmarks demonstrate EM-LLM's superior performance, consistently outperforming the state-of-the-art retrieval model InfLLM across various baseline LLMs. In addition, EM-LLM outperforms its popular counterpart, RAG, in a wide range of tasks, while requiring similar resources. Notably, EM-LLM's performance even surpasses full-context models in most tasks, while successfully performing retrieval across 10 million tokens - a scale computationally infeasible for such models. Finally, our analysis reveals strong correlations between EM-LLM's event segmentation and human-perceived events, suggesting parallels between this artificial system and its biological counterpart, thereby offering a novel computational framework for exploring human memory mechanisms.

Type: Proceedings paper
Title: Human-inspired Episodic Memory for Infinite Context LLMs
Event: ICLR 2025
Open access status: An open access version is available from UCL Discovery
Publisher version: https://openreview.net/forum?id=BI2int5SAC
Language: English
Additional information: This version is the version of record. For information on re-use, please refer to the publisher’s terms and conditions.
Keywords: large language models, long context, retrieval, episodic memory, event cognition, training-free
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL BEAMS
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Computer Science
URI: https://discovery.ucl.ac.uk/id/eprint/10212510
Downloads since deposit
4Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item