Simratsingh2002/MCP-Server-Latest_Doc-
If you are the rightful owner of MCP-Server-Latest_Doc- and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to henry@mcphub.com.
A minimal Model Context Protocol (MCP) server that provides documentation-search tools for AI agents.
get_docs
Searches and fetches documentation based on a query and specified library.
MCP Server
A minimal Model Context Protocol (MCP) server that exposes documentation-search tools for any AI agent. Use this server to fetch and search docs for popular libraries (LangChain, LlamaIndex, OpenAI) and wire them seamlessly into your agents.
Features
- Doc Search Tool: Provides a
get_docs
tool to search and fetch documentation based on a query and library. - Web Scraping: Uses Serper API for search results and BeautifulSoup for fetching page text.
- Easy Integration: Decorate functions with
@mcp.tool()
and run the server over stdio transport.
Prerequisites
- Python 3.8 or higher
- A Serper API key (set in
.env
asSERPER_API_KEY
)
Installation
-
Clone this repo:
git clone https://github.com/<your-username>/MCP-Server-Latest_Doc-.git cd MCP-Server-Latest_Doc-
-
Create a virtual environment & install dependencies:
python3 -m venv .venv source .venv/bin/activate pip install -r requirements.txt
-
Add your Serper API key in a
.env
file:echo "SERPER_API_KEY=your_api_key_here" > .env
Usage
Running the MCP Server
Execute the main script to start the MCP server on stdio:
python main.py
Using the get_docs
Tool
Within any AI agent that supports MCP, the get_docs
tool will automatically register. It expects two arguments:
query
: The search term (e.g.,"Chroma DB"
).library
: One oflangchain
,llama-index
, oropenai
.
The tool returns the concatenated text of the top search results from the specified library docs.
Integrating into an AI Agent
Any agent framework that implements the MCP transport can import and run this server. For example, in Python:
from mcp.server.fastmcp import FastMCP
from dotenv import load_dotenv
import os
load_dotenv()
mcp = FastMCP("docs")
# ... define @mcp.tool() functions here ...
if __name__ == "__main__":
mcp.run(transport="stdio")
After starting this script, your agent can call the get_docs
tool over the stdio channel.
Example: Using with Claude Code CLI
First, install the Claude Code CLI:
npm install -g @anthropic-ai/claude-code
Add the MCP server to Claude:
claude mcp add
Specify the uv
binary path (find with where uv
):
claude mcp configure --binary-path /full/path/to/uv
Set the working directory of the server:
claude mcp configure --workdir /path/to/MCP-Server-Latest_Doc-
Run the MCP server:
python main.py
Confirm registration:
claude mcp list
For debugging, you can use the Inspector:
npx @modelcontextprotocol/inspector uv run main.py
License
MIT ©