SKILL.md files package domain expertise into something any AI agent can use. Drop one into your project and your agent learns how to process PDFs, design interfaces, write tests, or whatever the skill teaches.
123 skills
A unified Model Context Protocol server implementation that aggregates multiple MCP servers into one. 1MCP (One MCP) is designed to simplify the way you work with AI assistants. Instead of configuring multiple MCP servers for different clients (Claude Desktop, Cherry Studio, Cursor, Roo Code, Claude, etc.), 1MCP provides a single, unified server. - 🔄 Unified Interface: Aggregates multiple MCP ser
ALWAYS use this skill when creating pull requests — never create a PR directly without it. Follows Sentry conventions for PR titles, descriptions, and issue references. Trigger on any create PR, open PR, submit PR, make PR, push and create PR, or prepare changes for review task.
ALWAYS use this skill when committing code changes — never commit directly without it. Creates commits following Sentry conventions with proper conventional commit format and issue references. Trigger on any commit, git commit, save changes, or commit message task.
Guide for creating Claude Code skills following Anthropic's official best practices. Use when user wants to create a new skill, build a skill, write SKILL.md, update an existing skill, or needs skill creation guidelines. Provides structure, frontmatter fields, naming conventions, and new features like dynamic context injection and subagent execution.
Screenpipe brand style guide. Reference this when designing UI components, writing copy, or making visual decisions.
Automatically hunt for high-impact OSS contribution opportunities in trending repositories.
Secure environment variable management ensuring secrets are never exposed in Claude sessions, terminals, logs, or git commits
AI-powered PPT generation with document analysis and styled images
Hard-edged writing style contract for timeless, forceful English prose without AI tics
Publish articles to X/Twitter
Text-to-speech and speech-to-text using fal.ai audio models
One-off HLS encoding (local video -> HLS renditions + playlist) using Transloadit via the `transloadit` CLI. Prefer builtin templates (`builtin/encode-hls-video@latest`) and download outputs locally via `-o`.
AI-powered image editing with style transfer and object removal
Add Uppy Dashboard + Transloadit uploads to a Next.js (App Router) app, with server-side signature generation and optional /s3/store export.
Analyze Rails apps and provide upgrade assessments
Use when auditing a codebase for semantic duplication - functions that do the same thing but have different names or implementations. Especially useful for LLM-generated codebases where new functions are often created rather than reusing existing ones.
MemoryMesh is a knowledge graph server designed for AI models, with a focus on text-based RPGs and interactive storytelling. It helps AI maintain consistent, structured memory across conversations, enabling richer and more dynamic interactions. The project is based on the Knowledge Graph Memory Server from the MCP servers repository and retains its core functionality. v0.3.0 Update: The MCP SDK ha
gomcp is an MCP server written in It exposes tools to interact with Lightpanda Browser via CDP protocol. gomcp requires you to have already installed Lightpanda You need to install Go to build from source. Once you have cloned the repository, build the binary with go build. By default, gocmp starts a local instance of Lightpanda browser. On the first run, you need to download the binary with the c
Educational MCP server with 17 tools, persistent workspace, and cloud hosting. Built with FastMCP and the official Model Context Protocol Python SDK. Available on: - Official MCP Registry - io.github.clouatre-labs/math-mcp-learning-server - PyPI - math-mcp-learning-server Requires an MCP client: - Claude Desktop - Anthropic's desktop app - Claude Code - Command-line MCP client - Goose - Open-sourc
Execute Hugging Face Hub operations using the `hf` CLI. Use when the user needs to download models/datasets/spaces, upload files to Hub repositories, create repos, manage local cache, or run compute jobs on HF infrastructure. Covers authentication, file transfers, repository creation, cache operations, and cloud compute.
An AI coding assistant built in Rust that provides both command-line and graphical interfaces for autonomous code analysis and modification. Multi-Modal Tool Execution: Adapts to different LLM capabilities with pluggable tool invocation modes - native function calling, XML-style tags, and triple-caret blocks - ensuring compatibility across various AI providers. Real-Time Streaming Interface: Advan
Opinionated, evolving constraints to guide agents when building interfaces
Start an OpenAI Codex CLI session for deep reasoning and code analysis
Upgrade Expo SDK versions