RAG AI Agent Skills
AI agent skills for retrieval-augmented generation. Embedding pipelines, vector search, and knowledge base workflows.
38 listings
Vectara MCP Server
MCP ServerVectara-MCP provides any agentic application with access to fast, reliable RAG with reduced hallucination, powered by Vectara's Trusted RAG platform, through the MCP protocol. You can install the package directly from PyPI: - Security: Built-in authentication via bearer tokens - Encryption: HTTPS ready - Rate Limiting: 100 requests/minute by default - CORS Protection: Configurable origin validatio
Context CLI
Lint any URL for LLM readiness. Get a 0-100 score for token efficiency, RAG readiness, agent compatibility, and LLM extraction quality. Context CLI is an LLM Readiness Linter that checks how well a URL is structured for AI consumption. As LLM-powered search engines, RAG pipelines, and AI agents become primary consumers of web content, your pages need to be optimized for token efficiency, structure
MCP Local RAG
MCP ServerProvides score interpretation (< 0.3 good, > 0.5 skip), query optimization, and source naming for query_documents, ingest_file, ingest_data tools. Use this skill when working with RAG, searching documents, ingesting files, saving web content, or handling PDF, HTML, DOCX, TXT, Markdown.
DevRag
Free Local RAG for Claude Code - Save Tokens & Time 日本語版はこちら | Japanese Version DevRag is a lightweight RAG (Retrieval-Augmented Generation) system designed specifically for developers using Claude Code. Stop wasting tokens by reading entire documents - let vector search find exactly what you need. When using Claude Code, reading documents with the Read tool consumes massive amounts of tokens: - ❌
Local RAG Search
MCP ServerEfficiently perform web searches using the mcp-local-rag server with semantic similarity ranking. Use this skill when you need to search the web for current information, research topics across multiple sources, or gather context from the internet without using external APIs. This skill teaches effective use of RAG-based web search with DuckDuckGo, Google, and multi-engine deep research capabilities.
Skill Seekers
English | 简体中文 🧠 The data layer for AI systems. Skill Seekers turns any documentation, GitHub repo, or PDF into structured knowledge assets—ready to power AI Skills (Claude, Gemini, OpenAI), RAG pipelines (LangChain, LlamaIndex, Pinecone), and AI coding assistants (Cursor, Windsurf, Cline) in minutes, not hours. Skill Seekers is the universal preprocessing layer that sits between raw documentatio
Skill Depot
MCP Serverskill-depot replaces the "dump all skill frontmatter into context" approach with selective, semantic retrieval. Agent skills are stored as Markdown files and indexed with vector embeddings — only the relevant skills are loaded when needed, keeping context lean. - Semantic Search — Find skills by meaning, not just keywords, using embedded vector search - Fully Local — No API keys, no cloud. U
Knowledge-to-Action MCP
MCP Serverknowledge-to-action-mcp is an MCP server for people whose real project context lives in notes, decisions, roadmaps, and meeting docs, not just code. Most Obsidian MCP servers stop at "read a note" or "search a vault." This one goes further: That means an MCP client can move from: If you work out of Obsidian, your important context is usually spread across: - roadmap notes - meeting notes - decisio
RAGMap (RAG MCP Registry Finder)
RAGMap is a lightweight MCP Registry-compatible subregistry + MCP server focused on RAG-related MCP servers. - Ingests the official MCP Registry, enriches records for RAG use-cases, and serves a subregistry API. - Exposes an MCP server (remote Streamable HTTP + local stdio) so agents can search/filter RAG MCP servers. MapRag is a discovery + routing layer for retrieval. It helps agents and humans
Pinecone Model Context Protocol Server for Claude Desktop.
MCP ServerRead and write to a Pinecone index. The server implements the ability to read and write to a Pinecone index. - semantic-search: Search for records in the Pinecone index. - read-document: Read a document from the Pinecone index. - list-documents: List all documents in the Pinecone index. - pinecone-stats: Get stats about the Pinecone index, including the number of records, dimensions, and namespace
Similarity Search Patterns
Implement efficient similarity search with vector databases. Use when building semantic search, implementing nearest neighbor queries, or optimizing retrieval performance.
RAG Documentation MCP Server
MCP ServerAn MCP server implementation that provides tools for retrieving and processing documentation through vector search, enabling AI assistants to augment their responses with relevant documentation context. - Vector-based documentation search and retrieval - Support for multiple documentation sources - Semantic search capabilities - Automated documentation processing - Real-time context augmentation f
Local FAISS MCP Server
MCP ServerA Model Context Protocol (MCP) server that provides local vector database functionality using FAISS for Retrieval-Augmented Generation (RAG) applications. - Local Vector Storage: Uses FAISS for efficient similarity search without external dependencies - Document Ingestion: Automatically chunks and embeds documents for storage - Semantic Search: Query documents using natural language with sentence
RAG Implementation
PluginComprehensive guide to implementing RAG systems including vector database selection, chunking strategies, embedding models, and retrieval optimization. Use when building RAG systems, implementing semantic search, optimizing retrieval quality, or debugging RAG performance issues.