

AI agent framework featuring Observational Memory that achieves 95% on LongMemEval with 5-40x compression and stable, reproducible context windows.
Observational Memory (OM) is a memory system developed by Mastra that achieves 94.87% on LongMemEval with GPT-5-mini (the highest score ever recorded on this benchmark) and 84.23% with GPT-4o, beating the previous state-of-the-art.
The architecture uses two background agents (Observer and Reflector) that watch conversations and maintain a dense text-only observation log that replaces raw message history as it grows. The context window is broken into two blocks: the first is the list of observations, and the second is raw messages that haven't yet been compressed.
When messages hit 30k tokens (configurable threshold), a separate "observer agent" compresses messages into new observations that are appended to the first block. When observations hit 40k tokens (also configurable), a separate "reflector agent" garbage collects observations that don't matter.
The system was announced in February 2026 and is available for use with Mastra's agent framework.
Open-source framework.
Loading more......