LLM & AI Agent Skills
AI agent skills for working with large language models. Prompt engineering, API integration, and AI workflow patterns.
223 listings
Terminator
Give AI assistants (Claude, Cursor, VS Code, etc.) the ability to control your desktop and automate tasks across any application. Claude Code (one-liner): Other clients (Cursor, VS Code, Windsurf, etc.): Add to your MCP config file: See the MCP Agent README for detailed setup instructions. - Uses your browser session - no need to relogin, keeps all your cookies and auth - Doesn't take
Piston MCP Server
MCP ServerPiston MCP Server piston-mcp is an MCP server that allows LLMs to connect to and execute code using Piston . You can try out piston-mcp locally without cloning it. To try out piston-mcp you'll need to install uv: You will also need to download an MCP client to connect to piston-mcp, such as Claude Desktop. Update the MCP client's configuration with the following configuration to connect to piston-
AniList MCP Server
MCP ServerA Model Context Protocol (MCP) server that interfaces with the AniList API, allowing LLM clients to access and interact with anime, manga, character, staff, and user data from AniList. - Search for anime, manga, characters, staff, and studios - Get detailed information about specific anime, manga, characters, and staff members - Access user profiles and lists - Support for advanced filtering optio
MCP OpenAPI Schema Explorer
MCP ServerAn MCP (Model Context Protocol) server that provides token-efficient access to OpenAPI (v3.0) and Swagger (v2.0) specifications via MCP Resource Templates. The primary goal of this project is to allow MCP clients (like Cline or Claude Desktop) to explore the structure and details of large OpenAPI specifications without needing to load the entire file into an LLM's context window. It achieves this
Reapi/mcp Openapi
MCP ServerA Model Context Protocol (MCP) server that loads and serves multiple OpenAPI specifications to enable LLM-powered IDE integrations. This server acts as a bridge between your OpenAPI specifications and LLM-powered development tools like Cursor and other code editors. - Loads multiple OpenAPI specifications from a directory - Exposes API operations and schemas through MCP protocol - Enables LLMs to
Mermaid MCP Server
MCP ServerThe Mermaid MCP Server enables AI assistants like GitHub Copilot, Claude, and custom LLM applications to generate professional architecture diagrams, flowcharts, sequence diagrams, and more using natural language. It provides a Model Context Protocol interface for seamless integration with AI coding assistants. - 🤖 AI-Powered Generation: Create diagrams using natural language with GitHub Copilot
MCP Transcribe
MCP ServerTranscribe MCP Automate your transcriptions with AI. Transcribe MCP instantly connects your account to assistants like Claude, Windsurf, Cursor, and more so they can automate tasks on your behalf. The Local Server can add local files for transcription and return result to your Assistant in seconds. - ⚡ Fast, lightweight and LLM-friendly. No special ASR models needed, no setup and fighting python p
YouTube Summarize
MCP server that fetches YouTube video transcripts and optionally summarizes them. - Fetch transcripts in multiple formats (text, JSON, SRT, WebVTT, pretty-print) - Summarize videos — returns transcript with instructions for the LLM to produce a summary - List available languages for any video's transcripts - Flexible URL parsing — accepts full YouTube URLs (youtube.com/watch?v=, youtu.be/, youtube
Urlbox MCP Server
MCP ServerMCP server for the Urlbox Screenshot API. Enables your client to take screenshots, generate PDFs, extract HTML/markdown, and more from websites. Visit Urlbox for more information, and have a read of our docs or chat with your LLM post install to get a good understanding of its options and capabilities. 1. Install dependencies and build: 2. Get Urlbox API credentials: - Sign up at urlbox.com - Get
MCP Server Webcrawl
MCP ServerAdvanced search and retrieval for web crawler data. With mcp-server-webcrawl, your AI client filters and analyzes web content under your direction or autonomously. The server includes a fulltext search interface with boolean support, and resource filtering by type, HTTP status, and more. mcp-server-webcrawl provides the LLM a complete menu with which to search, and works with a variety of web craw
Finding Duplicate Functions
Use when auditing a codebase for semantic duplication - functions that do the same thing but have different names or implementations. Especially useful for LLM-generated codebases where new functions are often created rather than reusing existing ones.
Slack MCP Server
MCP ServerSlack MCP Server Give Claude the same Slack access you have. DMs, threads, history—no admin approval. I built this because I was working with someone to help me manage a complex workload, and we kept hitting walls. They needed context from my messages—"what did X say about Y?"—but Slack's API blocks access to DMs without admin approval. Screenshotting messages is not a workflow. This server bridge
ScreenPilot
MCP server to let LLM take full control on your device by providing screen automation toolkit for controlling and interacting with graphical user interfaces. Good for automation, education and having fun. - 📷 Screen capture and analysis - 🖱️ Mouse control (clicking, positioning) - ⌨️ Keyboard input (typing, key presses, hotkeys) 0. Install python 3.12 1. Clone the repository: 2. create virtiual
MCP Ragchat
MCP Servermcp-ragchat An MCP server that adds RAG-powered AI chat to any website. One command from Claude Code. Tell Claude Code "add AI chat to mysite.com" and it will crawl your content, build a local vector store, spin up a chat server, and hand you an embed snippet. No cloud infra. No database. Just one API key. 1. Clone and build 2. Configure Claude Code (~/.claude/mcp.json) Open Claude Code and say: C
Fal.ai MCP Server
MCP ServerA Model Context Protocol (MCP) server that enables Claude Desktop (and other MCP clients) to generate images, videos, music, and audio using Fal.ai models. - Native Async API - Uses falclient.runasync() for optimal performance - Queue Support - Long-running tasks (video/music) use queue API with progress updates - Non-blocking - All operations are truly asynchronous - STDIO - Traditional Model Con
Prometheus MCP Server
MCP ServerGive AI assistants the power to query your Prometheus metrics. A [Model Context Protocol][mcp] (MCP) server that provides access to your Prometheus metrics and queries through standardized MCP interfaces, allowing AI assistants to execute PromQL queries and analyze your metrics data. [mcp]: https://modelcontextprotocol.io - Prometheus server accessible from your environment - MCP-compatible client
Ntfy Me MCP
MCP Serverntfy-me-mcp provides AI assistants with the ability to send real-time notifications to your devices through the ntfy service (either public or selfhosted with token support). Get notified when your AI completes tasks, encounters errors, or reaches important milestones - all without constant monitoring. The server includes intelligent features like automatic URL detection for creating view actions
Dialog MCP Server
MCP Servermcp-name: io.github.king-of-the-grackles/reddit-research-mcp AI-Powered Reddit Intelligence for Market Research & Competitive Analysis Version: 1.0.0 Turn Reddit's chaos into actionable insights. Dialog is your all-in-one Reddit intelligence platform for competitive analysis, market research, and customer discovery across 20,000+ active subreddits. This is the official MCP server powering Dialog -
Conversation Memory
Persistent memory systems for LLM conversations including short-term, long-term, and entity-based memory Use when: conversation memory, remember, memory persistence, long-term memory, chat history.
OSV MCP Server
MCP ServerAn MCP (Model Context Protocol) server that provides access to the OSV (Open Source Vulnerabilities) database. This project implements an SSE-based MCP server that allows LLM-powered applications to query the OSV database for vulnerability information. The server provides tools for: 1. Querying vulnerabilities for a specific package version or commit 2. Batch querying vulnerabilities for multiple
Dingo
English · 简体中文 · 日本語 👋 join us on Discord and WeChat If you like Dingo, please give us a ⭐ on GitHub! Dingo is A Comprehensive AI Data, Model and Application Quality Evaluation Tool, designed for ML practitioners, data engineers, and AI researchers. It helps you systematically assess and improve the quality of training data, fine-tuning datasets, and production AI systems. 🎯 Production-Grade Qua
Mendeley MCP Server
MCP ServerAn MCP (Model Context Protocol) server that connects your Mendeley reference library to LLM applications like Claude Desktop, Cursor, and other MCP-compatible clients. - Search your library - Find papers by title, author, abstract, or notes - Browse folders - Navigate your collection structure - Get full metadata - Retrieve complete document details including abstracts - Search global catalog - Ac
MongoDB Lens
MongoDB Lens is a local Model Context Protocol (MCP) server with full featured access to MongoDB databases using natural language via LLMs to perform queries, run aggregations, optimize performance, and more. - Quick Start - Installation - Configuration - Client Setup - Data Protection - Test Suite - Disclaimer - Install MongoDB Lens - Configure MongoDB Lens - Set up your MCP Client (e.g. Claude D
Prompt Caching
Caching strategies for LLM prompts including Anthropic prompt caching, response caching, and CAG (Cache Augmented Generation) Use when: prompt caching, cache prompt, response cache, cag, cache augmented.