In trying to address the problem of limited contexts for large language models (LLMs), researchers have tried to emulate human episodic memory with EM-LLM.
The paper, titled “Human-like Episodic Memory for Infinite Context LLMs”, was released by researchers from Huawei and University College London.
EM-LLM (wherein EM stands for episodic memory) integrates important aspects of episodic memory and event cognition, used by the human brain, into LLMs. Through this, researchers have suggested that LLMs using EM-LLM (wherein EM stands for episodic memory) can potentially have infinite context lengths while maintaining their regular functioning.
This is particularly interesting as the method can be scalable without increasing the amount of computing power needed for the LLM to function. “By bridging insights from cognitive science with machine learning, our approach not only enhances the performance of LLMs on long-context tasks but also provides a scalable computational framework for testing hypotheses about human memory,” the researchers said.
EM-LLM goes about recreating human episodic memory by organising tokens into episodic events using Bayesian surprise and graph-theoretic boundary refinement. Following this, they employ a two-stage retrieval process, based on time and similarity, to allow for human-like access and retrieval of information.
Interestingly, when compared to similar proposals to improve context windows for LLMs, like InfLLM, EM-LLM excelled, with an overall improvement of 4.3%. Under the PassageRetrieval task, EM-LLM had a 33% improvement over InfLLM.
“Our analysis reveals strong correlations between EM-LLM’s event segmentation and human-perceived events, suggesting a bridge between this artificial system and its biological counterpart,” the researchers stated.
This is yet another step towards tackling the problem of context lengths when it comes to interacting with LLMs. Research has been ongoing to increase the context lengths, with major companies like Google and Meta releasing their own papers on pursuing infinite context lengths.