LLM & AI Agent Skills
AI agent skills for working with large language models. Prompt engineering, API integration, and AI workflow patterns.
223 listings
Winx - High-Performance Rust MCP Server ✨
🚀 1:1 Optimized Rust Implementation of WCGW (What Could Go Wrong) 🚀 Winx is a specialized Model Context Protocol (MCP) server that provides high-performance tools for LLM code agents. It implements the core functionality of WCGW in pure Rust for maximum efficiency and stability. Benchmarks on i9-13900K + RTX 4090 (WSL2) - Rust 1.75+ - Linux / macOS / WSL2 Add to ~/.config/Claude/claudedesktopcon
MaxMSP-MCP Server
MCP ServerThis project uses the Model Context Protocol (MCP) to let LLMs directly understand and generate Max patches. Video link. Acknowledgement: the patch being explained is downloaded from here. Text comments in the original file are deleted. Check out the full video where you can listen to the synthesised sounds. The LLM agent has access to the official documentation of each object, as well as objects
MetaTrader MCP Server
MCP ServerMetaTrader MCP Server Let AI assistants trade for you using natural language Features • Quick Start • Documentation • Examples • Support MetaTrader MCP Server is a bridge that connects AI assistants (like Claude, ChatGPT) to the MetaTrader 5 trading platform. Instead of clicking buttons, you can simply tell your AI assistant what to do: The AI understands your request and executes it on MetaTrader
MCP Server Webcrawl
MCP ServerAdvanced search and retrieval for web crawler data. With mcp-server-webcrawl, your AI client filters and analyzes web content under your direction or autonomously. The server includes a fulltext search interface with boolean support, and resource filtering by type, HTTP status, and more. mcp-server-webcrawl provides the LLM a complete menu with which to search, and works with a variety of web craw
MCP Transcribe
MCP ServerTranscribe MCP Automate your transcriptions with AI. Transcribe MCP instantly connects your account to assistants like Claude, Windsurf, Cursor, and more so they can automate tasks on your behalf. The Local Server can add local files for transcription and return result to your Assistant in seconds. - ⚡ Fast, lightweight and LLM-friendly. No special ASR models needed, no setup and fighting python p
Code Assistant
An AI coding assistant built in Rust that provides both command-line and graphical interfaces for autonomous code analysis and modification. Multi-Modal Tool Execution: Adapts to different LLM capabilities with pluggable tool invocation modes - native function calling, XML-style tags, and triple-caret blocks - ensuring compatibility across various AI providers. Real-Time Streaming Interface: Advan
RAG Engineer
Expert in building Retrieval-Augmented Generation systems. Masters embedding models, vector databases, chunking strategies, and retrieval optimization for LLM applications. Use when: building RAG, vector search, embeddings, semantic search, document retrieval.