MCPs tagged with large language models
-
MCP CLI
A powerful CLI for seamless interaction with Model Context Protocol servers and advanced LLMs.
MCP CLI is a modular command-line interface designed for interacting with Model Context Protocol (MCP) servers and managing conversations with large language models. It integrates with the CHUK Tool Processor and CHUK-LLM to provide real-time chat, interactive command shells, and automation capabilities. The system supports a wide array of AI providers and models, advanced tool usage, context management, and performance metrics. Rich output formatting, concurrent tool execution, and flexible configuration make it suitable for both end-users and developers.
- ⭐ 1,755
- MCP
- chrishayuk/mcp-cli
-
Enrichr MCP Server
Gene set enrichment analysis server for LLMs via the Model Context Protocol
Enrichr MCP Server provides gene set enrichment analysis using the Enrichr API, supporting all available gene set libraries. It is designed to integrate with LLM tools through the Model Context Protocol and returns only statistically significant results. The tool allows queries across multiple biological, disease, tissue, drug, and pathway gene set libraries, with customizable configuration for popular or specific libraries. Installation and integration is streamlined for platforms like Claude Desktop, Cursor, and VS Code.
- ⭐ 7
- MCP
- tianqitang1/enrichr-mcp-server
-
OpenZIM MCP Server
Transforms ZIM archives into intelligent, structured knowledge engines for LLMs.
OpenZIM MCP Server provides structured, intelligent access to ZIM-format knowledge bases, enabling large language models to efficiently search, navigate, and understand content in offline archives. Dual operation modes allow support for both advanced and simple LLM integrations. It features smart navigation by namespace, context-aware discovery, intelligent search, and relationship mapping to optimize knowledge extraction and utilization.
- ⭐ 8
- MCP
- cameronrye/openzim-mcp
-
FinData MCP Server
Open-source MCP server for LLM access to professional financial data.
FinData MCP Server offers standardized financial data access to large language models by implementing the Model Context Protocol. It supports multiple data providers such as Tushare and Wind, and enables retrieval of both market and macroeconomic data via Stdio or SSE transports. The server is configurable and integrates into AI systems, providing tools for market, fundamental, and financial data queries.
- ⭐ 45
- MCP
- zlinzzzz/finData-mcp-server
-
LLM Context
Reduce friction when providing context to LLMs with smart file selection and rule-based filtering.
LLM Context streamlines the process of sharing relevant project files and context with large language models. It employs smart file selection and customizable rule-based filtering to ensure only the most pertinent information is provided. The tool supports Model Context Protocol (MCP), allowing AI models to access additional files seamlessly through standardized commands. Integration with MCP enables instant project context sharing during AI conversations, enhancing productivity and collaboration.
- ⭐ 283
- MCP
- cyberchitta/llm-context.py
-
Supabase MCP Server
Connect Supabase projects to AI assistants using the Model Context Protocol.
Supabase MCP Server enables direct, secure integration between Supabase projects and AI assistants such as Cursor, Claude, and Windsurf. Leveraging the Model Context Protocol, it provides standardized endpoints for external LLMs to perform tasks like managing tables, fetching configurations, and querying data on Supabase. The server supports OAuth 2.1 Dynamic Client Registration and offers easy setup with feature groups and popular client installers for local, cloud, and self-hosted environments.
- ⭐ 2,263
- MCP
- supabase-community/supabase-mcp
-
Couchbase MCP Server
Enable LLMs to interact directly with Couchbase clusters via the Model Context Protocol.
Couchbase MCP Server provides an MCP-compliant server for connecting Large Language Models to Couchbase clusters. It supports various database operations such as bucket and collection listing, document retrieval, upsert, and deletion, as well as running SQL++ queries and retrieving index information. Designed for easy integration with MCP clients like Claude Desktop, it includes features for secure authentication and query mode configuration. The server can be deployed using a prebuilt PyPI package or directly from source.
- ⭐ 24
- MCP
- Couchbase-Ecosystem/mcp-server-couchbase
-
Open Data Model Context Protocol
Easily connect open data providers to LLMs using a Model Context Protocol server and CLI.
Open Data Model Context Protocol enables seamless integration of open public datasets into Large Language Model (LLM) applications, starting with support for Claude. Through a CLI tool and server, users can access and query public data providers within their LLM clients. It also offers tools and templates for contributors to publish and distribute new open datasets, making data discoverable and actionable for LLM queries.
- ⭐ 140
- MCP
- OpenDataMCP/OpenDataMCP
-
iMessage Query MCP Server
Securely query and analyze your iMessage history with LLMs using the Model Context Protocol.
iMessage Query MCP Server is an MCP-compliant server that provides safe access to the macOS iMessage database for Large Language Models (LLMs). Built with the FastMCP framework and imessagedb library, it enables querying, validation, and analysis of iMessage conversations, complete with proper phone number handling and permissions management. The tool integrates seamlessly with MCP clients such as Claude Desktop and CLine for VSCode, allowing for controlled retrieval of chat histories via standardized protocols. Enhanced security features and automated dependency management ensure a smooth and safe user experience.
- ⭐ 71
- MCP
- hannesrudolph/imessage-query-fastmcp-mcp-server
-
MaxMSP-MCP Server
Bridge LLMs with Max patches via Model Context Protocol
MaxMSP-MCP Server enables large language models to understand, explain, and generate Max patches by leveraging the Model Context Protocol. It connects LLM agents with MaxMSP environments, providing access to documentation and patch objects for detailed interaction. Installation includes both a Python server and Max environment integration, facilitating seamless Python-Max communication. The tool supports explaining patches, debugging, and synthesizer creation directly through LLM interfaces.
- ⭐ 106
- MCP
- tiianhk/MaxMSP-MCP-Server