mdskills
← All tags

Ai Memory AI Agent Skills

Browse AI agent skills tagged "Ai Memory". Find and install skills, MCP servers, and plugins for your AI coding assistant.

6 listings

Claude Mem

Plugin

Persistent memory compression system built for Claude Code. Gives your AI agent long-term memory across sessions with automatic context compression and retrieval.

8.0thedotmack/claude-mem

Claude Memory Skill

Claude Code forgets everything between sessions. Built-in auto-memory exists but: - It's opaque (Claude decides what's "meaningful") - Limited to 200 lines loaded at startup - Not tightly integrated into the agentic loop - No hierarchical organization (scales poorly) A skill-based memory protocol with: - Hierarchical storage: core.md summaries → topics/ .md details - Background agents: Memory ops

6.0hanfang/claude-memory-skill

Optional: Enable auto-save hooks (new in v2.5.0)

MEMCORD v3.0.2 (mcp server) This privacy-first, self-hosted MCP server helps you organize chat history, summarize messages, search across past chats with AI — and keeps everything secure and fully under your control. Never Lose Context Again Transform your Claude conversations into a searchable, organized knowledge base that grows with you - Core Benefits - Prerequisites - Quick Start - IDE Config

8.0ukkit/memcord

Mini Claude

Persistent memory for Claude Code. Also: loop detection, scope guards, code analysis, and a local LLM for second opinions. Claude Code forgets everything between sessions and after context compaction. Mini Claude provides: - Checkpoints - Save task state, auto-restore after compaction - Mistake memory - Log errors, get warned next time you touch that file - Decision memory - Log WHY you chose some

8.020alexl/mini_claude

Phloem

Your AI forgets everything when you close the tab. Phloem fixes that. Phloem is a local MCP server that gives your AI persistent memory across sessions. It works with any tool that supports the Model Context Protocol — an open standard. Today that includes Claude Code, VS Code, Cursor, Windsurf, Zed, Neovim, Cline, Warp, Continue, JetBrains, and more arriving every week. You install it once. Every

7.0CanopyHQ/phloem

Start your Cognee API server on the host

cognee‑mcp - Run cognee’s memory engine as a Model Context Protocol server Join Discord Join r/AIMemory Build memory for Agents and query from any client that speaks MCP – in your terminal or IDE. - Multiple transports – choose Streamable HTTP --transport http (recommended for web deployments), SSE --transport sse (real‑time streaming), or stdio (classic pipe, default) - API Mode – connect to an a

8.0topoteretes/cognee