SKILL.md files package domain expertise into something any AI agent can use. Drop one into your project and your agent learns how to process PDFs, design interfaces, write tests, or whatever the skill teaches.
112 skills
A powerful Ghidra extension that provides an MCP (Model Context Protocol) server, enabling AI assistants and other tools to interact with Ghidra's reverse engineering capabilities through a standardized API. GhidrAssistMCP bridges the gap between AI-powered analysis tools and Ghidra's comprehensive reverse engineering platform. By implementing the Model Context Protocol, this extension allows exte
This project implements a sophisticated multi-agent research system based on the concepts described in Anthropic's article: Building a Multi-Agent Research System. The system leverages the power of coordinated AI agents to handle deep and complex research queries that require multiple perspectives and iterative investigation. The system is built around a Lead Research Agent that orchestrates the e
Free Local RAG for Claude Code - Save Tokens & Time 日本語版はこちら | Japanese Version DevRag is a lightweight RAG (Retrieval-Augmented Generation) system designed specifically for developers using Claude Code. Stop wasting tokens by reading entire documents - let vector search find exactly what you need. When using Claude Code, reading documents with the Read tool consumes massive amounts of tokens: - ❌
Voice agents represent the frontier of AI interaction - humans speaking naturally with AI systems. The challenge isn't just speech recognition and synthesis, it's achieving natural conversation flow with sub-800ms latency while handling interruptions, background noise, and emotional nuance. This skill covers two architectures: speech-to-speech (OpenAI Realtime API, lowest latency, most natural) and pipeline (STT→LLM→TTS, more control, easier to debug). Key insight: latency is the constraint. Hu
Jotdown is a Model Context Protocol (MCP) server that allows large language models (LLMs) to interact with Notion and also generate Markdown Books. It provides two primary tools for LLMs: - 👉 Notion Integration: Create or update pages in Notion with content generated by the LLM. - 👉 Mdbook Generation: Generate a mdbook from content and manage the structure. Jotdown enables LLMs to seamlessly int
Memory is the cornerstone of intelligent agents. Without it, every interaction starts from zero. This skill covers the architecture of agent memory: short-term (context window), long-term (vector stores), and the cognitive architectures that organize them. Key insight: Memory isn't just storage - it's retrieval. A million stored facts mean nothing if you can't find the right one. Chunking, embedding, and retrieval strategies determine whether your agent remembers or forgets. The field is fragm
This skill should be used when users want to train or fine-tune language models using TRL (Transformer Reinforcement Learning) on Hugging Face Jobs infrastructure. Covers SFT, DPO, GRPO and reward modeling training methods, plus GGUF conversion for local deployment. Includes guidance on the TRL Jobs package, UV scripts with PEP 723 format, dataset preparation and validation, hardware selection, cost estimation, Trackio monitoring, Hub authentication, and model persistence.
English | 简体中文 🧠 The data layer for AI systems. Skill Seekers turns any documentation, GitHub repo, or PDF into structured knowledge assets—ready to power AI Skills (Claude, Gemini, OpenAI), RAG pipelines (LangChain, LlamaIndex, Pinecone), and AI coding assistants (Cursor, Windsurf, Cline) in minutes, not hours. Skill Seekers is the universal preprocessing layer that sits between raw documentatio
Analyzes web performance using Chrome DevTools MCP. Measures Core Web Vitals (FCP, LCP, TBT, CLS, Speed Index), identifies render-blocking resources, network dependency chains, layout shifts, caching issues, and accessibility gaps. Use when asked to audit, profile, debug, or optimize page load performance, Lighthouse scores, or site speed.
Practitioner-focused workflows for running, validating, and packaging reproducible open AI governance studies. This repository contains execution pipelines, source artifacts, and validation gates for reproducible governance reports. Every headline finding is expected to map to: - a versioned artifact - a deterministic query in the claims ledger - a reproducible execution path in pipelines/ Use thi
Expert in building voice AI applications - from real-time voice agents to voice-enabled apps. Covers OpenAI Realtime API, Vapi for voice agents, Deepgram for transcription, ElevenLabs for synthesis, LiveKit for real-time infrastructure, and WebRTC fundamentals. Knows how to build low-latency, production-ready voice experiences. Use when: voice ai, voice agent, speech to text, text to speech, realtime voice.
RAGMap is a lightweight MCP Registry-compatible subregistry + MCP server focused on RAG-related MCP servers. - Ingests the official MCP Registry, enriches records for RAG use-cases, and serves a subregistry API. - Exposes an MCP server (remote Streamable HTTP + local stdio) so agents can search/filter RAG MCP servers. MapRag is a discovery + routing layer for retrieval. It helps agents and humans
You are an AI assistant development expert specializing in creating intelligent conversational interfaces, chatbots, and AI-powered applications. Design comprehensive AI assistant solutions with natur
Design LLM applications using the LangChain framework with agents, memory, and tool integration patterns. Use when building LangChain applications, implementing AI agents, or creating complex LLM workflows.
aka Vibe Debugging This is an MCP Server and VS Code extension which enables claude to interactively debug and evaluate expressions. That means it should also work with other models / clients etc. but I only demonstrate it with Claude Desktop and Continue here. It's language-agnostic, assuming debugger console support and valid launch.json for debugging in VSCode. 1. Download the extension from re
Please migrate to our new KnowledgeBase MCP Server, which provides enhanced capabilities and improved accuracy. Model Context Protocol (MCP) is a standardized protocol designed to manage context between large language models (LLMs) and external systems. The Chargebee MCP Server offers a robust set of tools to improve developer efficiency. It integrates with AI-powered code editors like Cursor, Win
You are an expert LangChain agent developer specializing in production-grade AI systems using LangChain 0.1+ and LangGraph.
CentralMind Gateway: Create API or MCP Server in Minutes 🚀 Interactive Demo avialable here: https://centralmind.ai Simple way to expose your database to AI-Agent via MCP or OpenAPI 3.1 protocols. This will run for you an API: Which you can use inside your AI Agent: Gateway will generate AI optimized API. AI agents and LLM-powered applications need fas
Semantic search, similar content discovery, and structured research using Exa API
- MCP Server: code-to-tree - Using code-to-tree - Configure MCP Clients - Building (Windows) - Building (macOS) The code-to-tree server's goals are: 1. Give LLMs the capability of accurately converting source code into AST(Abstract Syntax Tree), regardless of language. 2. One standalone binary should be everything the MCP client needs. These goals imply: 1. The underlying syntax parser should be v
Execute Hugging Face Hub operations using the `hf` CLI. Use when the user needs to download models/datasets/spaces, upload files to Hub repositories, create repos, manage local cache, or run compute jobs on HF infrastructure. Covers authentication, file transfers, repository creation, cache operations, and cloud compute.
Stop copy-pasting between AI models. Roundtable AI is a local MCP server that lets your primary AI assistant delegate tasks to specialized models like Gemini, Claude, Codex, and Cursor. Solve complex engineering problems in parallel, directly from your IDE. Key Features: - Context Continuity: Shared project context across all sub-agents - Parallel Execution: All agents work simultaneously - Model
A Mattermost integration that connects to Model Context Protocol (MCP) servers, leveraging a LangGraph-based AI agent to provide an intelligent interface for interacting with users and executing tools directly within Mattermost. - 🤖 Langgraph Agent Integration: Uses a LangGraph agent to understand user requests and orchestrate responses. - 🔌 MCP Server Integration: Connects to multiple MCP serve
A tool similar to cloc, sloccount and tokei. For counting the lines of code, blank lines, comment lines, and physical lines of source code in many programming languages. Goal is to be the fastest code counter possible, but also perform COCOMO calculation like sloccount, LOCOMO estimation for LLM-based development costs, estimate code complexity similar to cyclomatic complexity calculators and prod