SKILL.md files package domain expertise into something any AI agent can use. Drop one into your project and your agent learns how to process PDFs, design interfaces, write tests, or whatever the skill teaches.
112 skills
Archive session learnings, debugging solutions, and deployment logs to .archive/yyyy-mm-dd/ as indexed markdown with searchable tags. Use when completing a significant task, resolving a tricky bug, deploying, or when the user says \"archive this\". Maintains .archive/MEMORY.md index for cross-session knowledge reuse.
Capture conversations and decisions into structured Notion pages; use when turning chats/notes into wiki entries, how-tos, decisions, or FAQs with proper linking.
Use when the user asks to generate or edit images via the OpenAI Image API (for example: generate image, edit/inpaint/mask, background removal or replacement, transparent background, product shots, concept art, covers, or batch variants); run the bundled CLI (`scripts/image_gen.py`) and require `OPENAI_API_KEY` for live calls.
Deploy web projects to Netlify using the Netlify CLI (`npx netlify`). Use when the user asks to deploy, host, publish, or link a site/repo on Netlify, including preview and production deploys.
Prepare meeting materials with Notion context and Codex research; use when gathering context, drafting agendas/pre-reads, and tailoring materials to attendees.
Deploy applications to Render by analyzing codebases, generating render.yaml Blueprints, and providing Dashboard deeplinks. Use when the user wants to deploy, host, publish, or set up their application on Render's cloud platform.
Use when the user asks for text-to-speech narration or voiceover, accessibility reads, audio prompts, or batch speech generation via the OpenAI Audio API; run the bundled CLI (`scripts/text_to_speech.py`) with built-in voices and require `OPENAI_API_KEY` for live calls. Custom voice creation is out of scope.
Token usage tracker for OpenAI, Claude, and Gemini APIs with MCP (Model Context Protocol) support. Pass accurate API costs to your users. - 🎯 Simple Integration - One line to wrap your API client - 📊 Automatic Tracking - No manual token counting - 💰 Accurate Pricing - Up-to-date pricing for all models (2025) - 🔄 Multiple Providers - OpenAI, Claude, and Gemini support - 📈 User Management - Tra
🚀 Try ApeRAG Live Demo - Experience the full platform capabilities with our hosted demo ApeRAG is a production-ready RAG (Retrieval-Augmented Generation) platform that combines Graph RAG, vector search, and full-text search with advanced AI agents. Build sophisticated AI applications with hybrid retrieval, multimodal document processing, intelligent agents, and enterprise-grade management feature
A Model Context Protocol (MCP) server that provides access to all the curated awesome lists and their items. It can provide the best resources for your agent from sections of the 8500+ awesome lists on github and more then 1mn+ (growing) awesome row items. What are Awesome Lists? Awesome lists are community-curated collections of the best tools, libraries, and resources on any topic - from machine
A vibe-coded memory management system with RAG capabilities for Claude via the Model Context Protocol (MCP). SimpleMem is an MCP server that provides persistent memory storage and retrieval for Claude and other MCP clients. It combines traditional file-based storage with modern RAG (Retrieval-Augmented Generation) capabilities, including semantic search and automatic relationship discovery. Think
Multi-agent AI consultation framework for Claude Code via MCP. Get a second (and third) opinion from other LLMs when Claude Code alone isn't enough. Claude Code is powerful, but one brain can miss bugs, overlook edge cases, or get stuck in a local optimum. Critical decisions benefit from diverse perspectives. Concilium runs parallel consultations with multiple LLMs through standard MCP protocol. E
>
MEMCORD v3.0.2 (mcp server) This privacy-first, self-hosted MCP server helps you organize chat history, summarize messages, search across past chats with AI — and keeps everything secure and fully under your control. Never Lose Context Again Transform your Claude conversations into a searchable, organized knowledge base that grows with you - Core Benefits - Prerequisites - Quick Start - IDE Config
Persistent memory for Claude Code. Also: loop detection, scope guards, code analysis, and a local LLM for second opinions. Claude Code forgets everything between sessions and after context compaction. Mini Claude provides: - Checkpoints - Save task state, auto-restore after compaction - Mistake memory - Log errors, get warned next time you touch that file - Decision memory - Log WHY you chose some
A powerful Ghidra extension that provides an MCP (Model Context Protocol) server, enabling AI assistants and other tools to interact with Ghidra's reverse engineering capabilities through a standardized API. GhidrAssistMCP bridges the gap between AI-powered analysis tools and Ghidra's comprehensive reverse engineering platform. By implementing the Model Context Protocol, this extension allows exte
This project implements a sophisticated multi-agent research system based on the concepts described in Anthropic's article: Building a Multi-Agent Research System. The system leverages the power of coordinated AI agents to handle deep and complex research queries that require multiple perspectives and iterative investigation. The system is built around a Lead Research Agent that orchestrates the e
Voice agents represent the frontier of AI interaction - humans speaking naturally with AI systems. The challenge isn't just speech recognition and synthesis, it's achieving natural conversation flow with sub-800ms latency while handling interruptions, background noise, and emotional nuance. This skill covers two architectures: speech-to-speech (OpenAI Realtime API, lowest latency, most natural) and pipeline (STT→LLM→TTS, more control, easier to debug). Key insight: latency is the constraint. Hu
Jotdown is a Model Context Protocol (MCP) server that allows large language models (LLMs) to interact with Notion and also generate Markdown Books. It provides two primary tools for LLMs: - 👉 Notion Integration: Create or update pages in Notion with content generated by the LLM. - 👉 Mdbook Generation: Generate a mdbook from content and manage the structure. Jotdown enables LLMs to seamlessly int
Practitioner-focused workflows for running, validating, and packaging reproducible open AI governance studies. This repository contains execution pipelines, source artifacts, and validation gates for reproducible governance reports. Every headline finding is expected to map to: - a versioned artifact - a deterministic query in the claims ledger - a reproducible execution path in pipelines/ Use thi
RAGMap is a lightweight MCP Registry-compatible subregistry + MCP server focused on RAG-related MCP servers. - Ingests the official MCP Registry, enriches records for RAG use-cases, and serves a subregistry API. - Exposes an MCP server (remote Streamable HTTP + local stdio) so agents can search/filter RAG MCP servers. MapRag is a discovery + routing layer for retrieval. It helps agents and humans
You are an AI assistant development expert specializing in creating intelligent conversational interfaces, chatbots, and AI-powered applications. Design comprehensive AI assistant solutions with natur
Design LLM applications using the LangChain framework with agents, memory, and tool integration patterns. Use when building LangChain applications, implementing AI agents, or creating complex LLM workflows.
aka Vibe Debugging This is an MCP Server and VS Code extension which enables claude to interactively debug and evaluate expressions. That means it should also work with other models / clients etc. but I only demonstrate it with Claude Desktop and Continue here. It's language-agnostic, assuming debugger console support and valid launch.json for debugging in VSCode. 1. Download the extension from re