LLM & AI Agent Skills
AI agent skills for working with large language models. Prompt engineering, API integration, and AI workflow patterns.
208 listings
Continue
ExtensionOpen-source VS Code and JetBrains extension that connects any LLM for autocomplete, chat, and inline editing with full model and context control.
LLM Application Dev Langchain Agent
You are an expert LangChain agent developer specializing in production-grade AI systems using LangChain 0.1+ and LangGraph.
RAG Engineer
Expert in building Retrieval-Augmented Generation systems. Masters embedding models, vector databases, chunking strategies, and retrieval optimization for LLM applications. Use when: building RAG, vector search, embeddings, semantic search, document retrieval.
AI Engineer
Build production-ready LLM applications, advanced RAG systems, and
Context7 MCP - Up-to-date Code Docs For Any Prompt
LLMs rely on outdated or generic information about the libraries you use. You get: - ❌ Code examples are outdated and based on year-old training data - ❌ Hallucinated APIs that don't even exist - ❌ Generic answers for old package versions Context7 MCP pulls up-to-date, version-specific documentation and code examples straight from the source — and places them directly into your prompt. Add use con
Ntfy Me MCP
MCP Serverntfy-me-mcp provides AI assistants with the ability to send real-time notifications to your devices through the ntfy service (either public or selfhosted with token support). Get notified when your AI completes tasks, encounters errors, or reaches important milestones - all without constant monitoring. The server includes intelligent features like automatic URL detection for creating view actions
Voice AI Engine Development
Build real-time conversational AI voice engines using async worker pipelines, streaming transcription, LLM agents, and TTS synthesis with interrupt handling and multi-provider support
Audio Transcriber
Transform audio recordings into professional Markdown documentation with intelligent summaries using LLM integration
Graphical Apps Development
RulesProject Synopsis
LLM Application Dev AI Assistant
You are an AI assistant development expert specializing in creating intelligent conversational interfaces, chatbots, and AI-powered applications. Design comprehensive AI assistant solutions with natur
HashiCorp Agent Skills
PluginA collection of Agent skills and Claude Code plugins for HashiCorp products. Install Agent Skills in GitHub Copilot, Claude Code, Opencode, Cursor, and more: First, add the marketplace, then install plugins: Or use the interactive interface: Each product folder contains plugins, and each plugin contains skills:
Deep Research
Execute autonomous multi-step research using Google Gemini Deep Research Agent. Use for: market analysis, competitive landscaping, literature reviews, technical research, due diligence. Takes 2-10 minutes but produces detailed, cited reports. Costs $2-5 per task.
Mermaid MCP Server
MCP ServerThe Mermaid MCP Server enables AI assistants like GitHub Copilot, Claude, and custom LLM applications to generate professional architecture diagrams, flowcharts, sequence diagrams, and more using natural language. It provides a Model Context Protocol interface for seamless integration with AI coding assistants. - 🤖 AI-Powered Generation: Create diagrams using natural language with GitHub Copilot
AI Product
Every product will be AI-powered. The question is whether you'll build it right or ship a demo that falls apart in production. This skill covers LLM integration patterns, RAG architecture, prompt engineering that scales, AI UX that users trust, and cost optimization that doesn't bankrupt you. Use when: keywords, file_patterns, code_patterns.
Prompt Engineering Patterns
Master advanced prompt engineering techniques to maximize LLM performance, reliability, and controllability in production. Use when optimizing prompts, improving LLM outputs, or designing production prompt templates.
Context Window Management
Strategies for managing LLM context windows including summarization, trimming, routing, and avoiding context rot Use when: context window, token limit, context management, context engineering, long context.
Next.js TypeScript
RulesHolistic understanding of requirements & stack
Notion MCP Server
MCP ServerMCP Server for the Notion API, enabling LLM to interact with Notion workspaces. Additionally, it employs Markdown conversion to reduce context size when communicating with LLMs, optimizing token usage and making interactions more efficient. Here is a detailed explanation of the steps mentioned above in the following articles: - English Version: https://dev.to/suekou/operating-notion-via-claude-des
Stripe AI
Guide for upgrading Stripe API versions and SDKs
Vuln Nist MCP Server
MCP ServerA Model Context Protocol (MCP) server for querying NIST National Vulnerability Database (NVD) API endpoints. This MCP server exposes tools to query the NVD/CVE REST API and return formatted text results suitable for LLM consumption via the MCP protocol. It includes automatic query chunking for large date ranges and parallel processing for improved performance. Base API docs: https://nvd.nist.gov/d
Model Context Protocol Server for Home Assistant
MCP ServerThe server uses the MCP protocol to share access to a local Home Assistant instance with an LLM application. A powerful bridge between your Home Assistant instance and Language Learning Models (LLMs), enabling natural language control and monitoring of your smart home devices through the Model Context Protocol (MCP). This server provides a comprehensive API for managing your entire Home Assistant
Code To Tree
- MCP Server: code-to-tree - Using code-to-tree - Configure MCP Clients - Building (Windows) - Building (macOS) The code-to-tree server's goals are: 1. Give LLMs the capability of accurately converting source code into AST(Abstract Syntax Tree), regardless of language. 2. One standalone binary should be everything the MCP client needs. These goals imply: 1. The underlying syntax parser should be v
DroidMind 🧠
Control Android devices with AI through the Model Context Protocol DroidMind is a powerful bridge between AI assistants and Android devices, enabling control, debugging, and system analysis through natural language. By implementing the Model Context Protocol (MCP), DroidMind allows AI models to directly interact with Android devices via ADB in a secure, structured way. When used as part of an agen
Prompt Caching
Caching strategies for LLM prompts including Anthropic prompt caching, response caching, and CAG (Cache Augmented Generation) Use when: prompt caching, cache prompt, response cache, cag, cache augmented.