Memory Systems
Long-term memory, vector stores, and context persistence
Build a playbook about Memory Systems
Save articles from this feed, then generate a personalized implementation guide
OpenClaw v2026.4.12 is a comprehensive quality release emphasizing plugin loading reliability, memory system improvements, and expanded local model support. Key additions include an Active Memory sub-agent plugin, local MLX speech provider for Talk Mode, LM Studio provider for self-hosted models, and enhanced Feishu setup workflows. The release also addresses critical security fixes, dreaming/narrative stability, and improves gateway lifecycle management and Matrix client streaming support.
★★★★★OpenClaw v2026.4.12-beta.1 introduces significant improvements to plugin loading, memory recall, and dreaming functionality. Key changes include narrowing plugin activation to manifest-declared needs, improving QMD recall search telemetry, and fixing critical issues with memory retrieval, dreaming narrative cleanup, and gateway authentication. The release also expands documentation for providers and memory-wiki setups while addressing numerous edge cases in CLI updates, Unicode handling, and WebSocket keepalive mechanisms.
★★★★★OpenClaw v2026.4.11-beta.1 introduces significant enhancements to memory/dreaming features with ChatGPT import ingestion, improved media rendering in chat UIs, expanded video generation capabilities, and better integration with Feishu and Microsoft Teams. The release includes numerous fixes for OAuth flows, audio transcription, TTS persistence, and configuration handling, plus improved plugin manifest descriptors and model provider diagnostics.
★★★★★crewAI 1.14.0 introduces checkpoint management with CLI commands and SqliteProvider storage, implements runtime state checkpointing with an event system, and refactors the executor. Security enhancements include SSRF/path traversal protections and RAG tool validation. The release also optimizes memory serialization by excluding embedding vectors and addresses critical CVE-2026-35030 via litellm upgrade.
★★★★★The Claude Agent SDK provides built-in session management functions to build conversation history sidebars without writing custom parsers. This cookbook demonstrates how to list, read, rename, tag, and fork sessions stored as JSONL transcripts on disk. By leveraging SDK primitives like list_sessions(), get_session_messages(), and rename_session(), developers can implement session browsers for desktop apps, IDE extensions, and chatbots without managing file I/O directly.
★★★★★crewAI 1.14.0a4 introduces checkpoint storage capabilities with SqliteProvider and CheckpointConfig, implements runtime state checkpointing with an event system, and refactors the executor. Key improvements include guardrail tracing enhancements, embedding vector exclusion from memory serialization to reduce token usage, and a security bump to litellm. The release also deprecates CodeInterpreterTool and code execution parameters while improving documentation and performance with JSONB storage.
★★★★★CrewAI v1.13.0 introduces significant architectural improvements including RuntimeState RootModel for unified serialization, enhanced event telemetry for skills and memory, and A2UI extension support. The release includes critical bug fixes for GPT-5.x model compatibility, RBAC permission alignment, and Windows lancedb compatibility. Performance optimizations reduce framework overhead through lazy event bus implementation and conditional tracing.
★★★★★Claude Code v2.1.89 is a major release introducing deferred permission decisions for headless sessions, flicker-free rendering, improved MCP connection handling, and numerous bug fixes across tool execution, file handling, and UI rendering. Key improvements include named subagents, enhanced auto-mode permission workflows, and fixes for critical issues like memory leaks, LSP crashes, and prompt cache misses in long sessions.
★★★★★This cookbook teaches context engineering strategies for long-running AI agents, focusing on three key techniques: compaction (summarizing context), tool-result clearing (removing re-fetchable tool outputs), and memory (persistent external storage). The guide addresses context rot—the degradation of model performance as context windows grow—and provides practical implementations using Claude's API. Through a research agent example, it demonstrates how to combine these strategies to manage token growth, maintain conversation continuity, and persist knowledge across sessions.
★★★★★This guide teaches how to build knowledge graphs from unstructured text using Claude's structured outputs for entity extraction, relation mining, and entity resolution. Rather than training traditional NER and relation classifiers, Claude handles each stage via prompts, enabling multi-hop graph reasoning without a database. The approach uses Haiku for high-volume extraction and Sonnet for entity resolution, with techniques transferable to production databases like Neo4j or PostgreSQL.
★★★★★CrewAI 1.12.0 introduces significant enhancements including Qdrant Edge storage backend, native OpenAI-compatible provider support (OpenRouter, DeepSeek, Ollama, vLLM, Cerebras, Dashscope), agent skills implementation, and automatic root_scope for hierarchical memory isolation. The release adds Arabic language support, a new docs-check CLI command, logout functionality, and resolves multiple bugs across type checking, HITL flow system, and memory management. Documentation improvements include a new CONTRIBUTING.md guide and instructions for using CrewAI without LiteLLM.
★★★★★CrewAI 1.12.0a1 introduces native OpenAI-compatible provider support (OpenRouter, DeepSeek, Ollama, vLLM, Cerebras, Dashscope), agent skills, and automatic hierarchical memory isolation via root_scope. The release includes Arabic language support for documentation and tooling, a new docs-check command, and significant refactoring to decouple from LiteLLM while resolving mypy type-checking errors across all packages.
★★★★★OpenAI Agents Python v0.13.0 is a minor release that updates the default Realtime websocket model to gpt-realtime-1.5, expands MCP server capabilities with resource management and session tracking, and introduces opt-in reasoning content replay for chat completions. The release includes critical runtime stability fixes for concurrent writes, tool removal edge cases, and batch executor races, with no breaking changes.
★★★★★This tutorial introduces Projects in Claude.ai, a feature that enables users to organize their work by maintaining persistent context and custom instructions across conversations. Projects allow you to group related conversations, documents, and settings in one place, making it easier to manage complex workflows and maintain consistency across multiple interactions with Claude.
★★★★★OpenAI Agents Python v0.12.4 release includes critical bug fixes for MCP (Model Context Protocol) tool invocations, retry logic improvements, and SQLite session memory handling. The update normalizes cancelled MCP invocations into tool errors, adds resilience for transient HTTP failures, fixes custom table name support in AdvancedSQLiteSession, and caps jittered retry delays. Documentation improvements and CI hardening were also included.
★★★★★OpenClaw v2026.3.13-1 is a recovery release addressing a GitHub immutability constraint where the original v2026.3.13 tag could not be reused. This patch release includes 50+ bug fixes and improvements across compaction, session management, platform integrations (Telegram, Discord, Slack, Android, iOS), security hardening, and UI refinements. Key fixes address token counting in compaction, SSRF vulnerabilities, gateway metadata failures, and memory regressions in the build system.
★★★★★This cookbook demonstrates how to build AI agents with persistent memory using Claude's memory tool and context editing capabilities. It addresses challenges of long-running agents losing learned patterns between sessions and context window overflow by implementing cross-conversation learning and automatic context management. The guide covers practical implementations for use cases like code review assistants, research assistants, and customer support bots, with setup instructions and best practices for memory security and organization.
★★★★★This cookbook article introduces the Claude Agent SDK for building multi-agent systems, using a Chief of Staff agent for a startup as the primary example. It demonstrates key features including persistent memory via CLAUDE.md files, bash tool execution for Python scripts, and coordination of specialized subagents. The article progressively builds a comprehensive agent system that aggregates insights from multiple domains to provide executive summaries and actionable recommendations for a CEO managing a $10M Series A startup.
★★★★★This cookbook teaches developers how to manage long-running Claude conversations by implementing session memory compaction using background threading and prompt caching. Rather than waiting for context limits to be exceeded (reactive approach), the pattern enables instant compaction by proactively building summaries in the background. The guide covers writing effective session memory prompts, implementing background threading for zero-latency compaction, and applying prompt caching to reduce costs by ~80%. It includes Python code examples demonstrating both traditional (slow) and instant (fast) compaction strategies for conversational applications.
★★★★★This cookbook demonstrates automatic context compaction for managing token limits in long-running agentic workflows. It shows how the Claude Agent Python SDK can automatically compress conversation history when token usage exceeds a threshold, enabling tasks to continue beyond the 200k token context limit. The example uses a customer service agent processing support tickets, where each ticket requires multiple tool calls that accumulate in conversation history. By implementing context compaction with the compaction_control parameter, agents can maintain focus and efficiency across many iterations without manual context management.
★★★★★