Ai Memory AI Agent Skills
Browse AI agent skills tagged "Ai Memory". Find and install skills, MCP servers, and plugins for your AI coding assistant.
6 listings
Claude Mem
PluginPersistent memory compression system built for Claude Code. Gives your AI agent long-term memory across sessions with automatic context compression and retrieval.
Claude Memory Skill
Claude Code forgets everything between sessions. Built-in auto-memory exists but: - It's opaque (Claude decides what's "meaningful") - Limited to 200 lines loaded at startup - Not tightly integrated into the agentic loop - No hierarchical organization (scales poorly) A skill-based memory protocol with: - Hierarchical storage: core.md summaries → topics/ .md details - Background agents: Memory ops
Optional: Enable auto-save hooks (new in v2.5.0)
MEMCORD v3.0.2 (mcp server) This privacy-first, self-hosted MCP server helps you organize chat history, summarize messages, search across past chats with AI — and keeps everything secure and fully under your control. Never Lose Context Again Transform your Claude conversations into a searchable, organized knowledge base that grows with you - Core Benefits - Prerequisites - Quick Start - IDE Config
Mini Claude
Persistent memory for Claude Code. Also: loop detection, scope guards, code analysis, and a local LLM for second opinions. Claude Code forgets everything between sessions and after context compaction. Mini Claude provides: - Checkpoints - Save task state, auto-restore after compaction - Mistake memory - Log errors, get warned next time you touch that file - Decision memory - Log WHY you chose some
Phloem
Your AI forgets everything when you close the tab. Phloem fixes that. Phloem is a local MCP server that gives your AI persistent memory across sessions. It works with any tool that supports the Model Context Protocol — an open standard. Today that includes Claude Code, VS Code, Cursor, Windsurf, Zed, Neovim, Cline, Warp, Continue, JetBrains, and more arriving every week. You install it once. Every
Start your Cognee API server on the host
cognee‑mcp - Run cognee’s memory engine as a Model Context Protocol server Join Discord Join r/AIMemory Build memory for Agents and query from any client that speaks MCP – in your terminal or IDE. - Multiple transports – choose Streamable HTTP --transport http (recommended for web deployments), SSE --transport sse (real‑time streaming), or stdio (classic pipe, default) - API Mode – connect to an a