MCPs tagged with context sharing
-
LLM Context
Reduce friction when providing context to LLMs with smart file selection and rule-based filtering.
LLM Context streamlines the process of sharing relevant project files and context with large language models. It employs smart file selection and customizable rule-based filtering to ensure only the most pertinent information is provided. The tool supports Model Context Protocol (MCP), allowing AI models to access additional files seamlessly through standardized commands. Integration with MCP enables instant project context sharing during AI conversations, enhancing productivity and collaboration.
- ⭐ 283
- MCP
- cyberchitta/llm-context.py
-
QA Sphere MCP Server
Model Context Protocol server enabling LLMs to interact with QA Sphere test cases
QA Sphere MCP Server provides a Model Context Protocol (MCP) integration for QA Sphere, allowing Large Language Models to interact with, discover, and summarize test cases within the QA Sphere test management system. It enables AI-powered IDEs and MCP clients to reference and manipulate QA Sphere test case data within development workflows. The solution supports quick integration into clients like Claude, Cursor, and 5ire, facilitating seamless collaboration and context sharing for AI-assisted development.
- ⭐ 15
- MCP
- Hypersequent/qasphere-mcp