Command Palette
Search for a command to run...
LMEB : Benchmark d'incorporation de mémoire à long horizon
LMEB : Benchmark d'incorporation de mémoire à long horizon
Résumé
Les représentations vectorielles (embeddings) de la mémoire sont essentielles pour les systèmes enrichis par la mémoire, tels qu'OpenClaw. Pourtant, leur évaluation reste largement sous-exploitée dans les benchmarks actuels d'embeddings textuels, lesquels se concentrent de manière étroite sur la récupération traditionnelle de passages et échouent à évaluer la capacité des modèles à gérer des tâches de récupération de mémoire à long horizon impliquant des informations fragmentées, dépendantes du contexte et temporellement distantes. Pour remédier à cette lacune, nous introduisons le Benchmark d'Embedding de Mémoire à Long Horizon (LMEB), un cadre complet évaluant les capacités des modèles d'embedding face à des tâches complexes de récupération de mémoire à long horizon. LMEB couvre 22 jeux de données et 193 tâches de récupération en zéro-shot réparties sur quatre types de mémoire : épisodique, dialogique, sémantique et procédurale, avec des données à la fois générées par l'IA et annotées par des humains. Ces types de mémoire diffèrent par leur niveau d'abstraction et leur dépendance temporelle, capturant ainsi des aspects distincts de la récupération de mémoire qui reflètent la diversité des défis du monde réel. Nous avons évalué 15 modèles d'embedding largement utilisés, dont la taille varie de quelques centaines de millions à dix milliards de paramètres. Les résultats révèlent que : (1) LMEB présente un niveau de difficulté raisonnable ; (2) les modèles plus grands ne performant pas systématiquement mieux ; (3) LMEB et MTEB sont orthogonaux. Cela suggère que le domaine n'a pas encore convergé vers un modèle universel capable d'exceller dans toutes les tâches de récupération de mémoire, et que les performances en récupération traditionnelle de passages ne se généralisent pas nécessairement à la récupération de mémoire à long horizon. En résumé, en offrant un cadre d'évaluation standardisé et reproductible, LMEB comble une lacune cruciale dans l'évaluation des embeddings de mémoire, favorisant ainsi des avancées supplémentaires dans le domaine des embeddings textuels pour la gestion de la récupération de mémoire à long terme et dépendante du contexte. LMEB est disponible à l'adresse suivante : https://github.com/KaLM-Embedding/LMEB.
One-sentence Summary
Researchers from Harbin Institute of Technology and Shenzhen Loop Area Institute introduce LMEB, a comprehensive benchmark evaluating memory embeddings across diverse long-horizon retrieval tasks. Unlike traditional benchmarks, LMEB reveals that larger models do not always excel, highlighting a critical gap in current text embedding capabilities for complex, context-dependent memory scenarios.
Key Contributions
- Current text embedding benchmarks fail to assess long-horizon memory retrieval involving fragmented, context-dependent, and temporally distant information, creating a gap in evaluating memory-augmented systems.
- The authors introduce the Long-horizon Memory Embedding Benchmark (LMEB), a comprehensive framework spanning 22 datasets and 193 zero-shot tasks across episodic, dialogue, semantic, and procedural memory types.
- Evaluation of 15 embedding models reveals that larger models do not consistently outperform smaller ones and that LMEB performance is orthogonal to traditional benchmarks like MTEB, indicating a lack of universal models for this domain.
Introduction
Text embedding models are critical for enabling efficient similarity search and powering downstream applications like retrieval and classification, yet current evaluation standards fall short in assessing their ability to handle long-horizon memory tasks. Existing benchmarks primarily focus on sentence-level similarity or standard retrieval across domains but rarely test scenarios requiring the synthesis of fragmented, context-dependent, and temporally distant evidence. To address this gap, the authors introduce LMEB, a specialized benchmark designed to evaluate how well embedding models support complex memory-centric retrieval that traditional metrics overlook.
Dataset
LMEB Dataset Overview
The authors introduce the Long-horizon Memory Embedding Benchmark (LMEB), a comprehensive framework designed to evaluate embedding models on complex, long-term memory retrieval tasks that traditional benchmarks like MTEB fail to address.
-
Dataset Composition and Sources
- The benchmark consolidates 22 English datasets into a unified schema, covering 193 zero-shot retrieval tasks.
- Data sources include a mix of AI-generated content and human-annotated material derived from crowdsourcing, academic papers, novels, and real-world logs.
- The datasets are categorized into four distinct memory types: Episodic, Dialogue, Semantic, and Procedural.
-
Key Details by Subset
- Episodic Memory: Focuses on recalling past events with temporal and spatial cues using datasets like EPBench (synthetic events) and KnowMeBench (autobiographical narratives).
- Dialogue Memory: Targets multi-turn context retention and user preferences using long-form conversations from LoCoMo, LongMemEval, REALTALK, and TMD.
- Semantic Memory: Assesses retrieval of stable, general knowledge from scientific papers (QASPER, SciFact), novels (NovelQA), and reports (ESG-Reports).
- Procedural Memory: Evaluates the retrieval of skills and action sequences using API documentation (Gorilla, ToolBench) and task trajectories (ReMe, DeepPlanning).
-
Usage in Model Evaluation
- The authors utilize the entire collection for zero-shot evaluation, meaning models are tested on their pre-trained capabilities without task-specific fine-tuning.
- The benchmark serves as a diagnostic tool to measure performance across varying levels of abstraction and temporal dependency.
- Results indicate that LMEB and MTEB are orthogonal, suggesting that high performance in traditional passage retrieval does not guarantee success in long-horizon memory tasks.
-
Processing and Construction Details
- Unified Schema: All resources are converted into a standard Information Retrieval format containing
queries.jsonl,corpus.jsonl,qrels.csv, and an optionalcandidates.jsonl. - Temporal Anchoring: For queries with relative time expressions, the authors append explicit time anchors (e.g., "Current time: 11:17 AM on Sunday") to disambiguate references.
- Metadata Encoding: Timestamps and hierarchical structures (such as session or turn levels in dialogues) are preserved in the title or text fields to support time-sensitive and scope-specific queries.
- Text Segmentation: Long documents in semantic tasks, such as novels and research papers, are segmented into passages or sentences using tools like
semchunkwith a chunk size of 256 tokens. - Candidate Constraints: An optional candidates file restricts retrieval to specific memory scopes, such as a single conversation history, rather than the full corpus.
- Unified Schema: All resources are converted into a standard Information Retrieval format containing
Method
The authors introduce the Long-horizon Memory Embedding Benchmark (LMEB) to systematically evaluate memory capabilities across diverse scenarios. The framework categorizes tasks into four distinct memory types: Episodic, Dialogue, Semantic, and Procedural. Each type involves a query retrieving relevant information from a specific memory store.

To structure these evaluations, the authors define a taxonomy based on abstraction and temporal dependency. As shown in the figure below, Episodic and Dialogue memories exhibit strong temporal dependencies, while Semantic and Procedural memories vary in abstraction levels.

To ensure embeddings follow instructions during downstream tasks, the authors prepend specific task instructions to the queries. The instructed query is formulated as follows:
qinst=Instruct: {task instruction} \nQuery: q
where q denotes the original query and qinst is the instructed query. This mechanism allows the model to adapt its retrieval and processing based on the specific requirements of the task.
Experiment
- The LMEB benchmark evaluation validates a unified pipeline for testing embedding models across episodic, dialogue, semantic, and procedural memory tasks, demonstrating that the benchmark offers a balanced difficulty level that effectively challenges current models.
- Experiments comparing models of varying scales reveal that larger parameter counts do not guarantee superior performance, as smaller models often achieve comparable or better results depending on architecture and training data.
- Analysis of task instructions shows that their impact on retrieval performance is model-dependent, with some models benefiting from instructions while others perform better without them or remain unaffected.
- Correlation studies confirm that LMEB evaluates capabilities orthogonal to traditional benchmarks like MTEB, particularly showing that strong performance in standard passage retrieval does not generalize well to complex episodic or dialogue memory scenarios.