δ-mem Provides Efficient Online Memory for LLMs
- •Researchers introduced δ-mem, a lightweight associative memory mechanism for large language models on May 12, 2026.
- •The system uses an 8x8 online memory state to improve performance 1.10x over frozen backbones.
- •Benchmarks show 1.31x gains on MemoryAgentBench and 1.20x on LoCoMo without requiring full model fine-tuning.
Researchers introduced δ-mem, a lightweight memory mechanism for large language models, on May 12, 2026. The system enhances performance by adding a compact associative memory state to a frozen attention backbone, providing low-rank corrections to attention computations. This approach compresses historical information into a fixed-size state matrix, which is updated using delta-rule learning. By utilizing an 8x8 online memory state, δ-mem improves the average model score to 1.10x compared to a frozen backbone and 1.15x over the strongest non-δ-mem memory baseline.
The mechanism demonstrates significant gains on memory-heavy benchmarks, achieving 1.31x on MemoryAgentBench and 1.20x on LoCoMo. These improvements occur without requiring full fine-tuning, backbone replacement, or explicit context window extensions. The implementation allows for efficient accumulation and reuse of information in long-term assistants and agent systems while preserving the general capabilities of the original model.