yet-market/yet-sparql-mcp-server
If you are the rightful owner of yet-sparql-mcp-server and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to henry@mcphub.com.
MCP SPARQL Server is a high-performance, configurable server that connects to any SPARQL endpoint and provides enhanced functionality including result formatting and caching.
MCP SPARQL Server
š Overview
MCP SPARQL Server is a high-performance, configurable server that connects to any SPARQL endpoint and provides enhanced functionality including result formatting and caching. It's built on top of the FastMCP framework implementing the Model Context Protocol (MCP) to provide a seamless interface for AI assistants to query semantic data.
⨠Features
- Universal Endpoint Support: Connect to any SPARQL-compliant endpoint
- Full SPARQL Support: Execute any valid SPARQL query (SELECT, ASK, CONSTRUCT, DESCRIBE)
- Intelligent Result Formatting:
- Standard JSON (compatible with standard SPARQL clients)
- Simplified JSON (easier to work with in applications)
- Tabular format (ready for display in UI tables)
- High-Performance Caching:
- Multiple cache strategies (LRU, LFU, FIFO)
- Configurable TTL (time-to-live)
- Cache management tools
- Flexible Deployment Options:
- Run in foreground mode with stdio or HTTP transport
- Run as a background daemon
- Deploy as a systemd service
- HTTP server mode for nginx reverse proxy integration
- Comprehensive Configuration:
- Command-line arguments
- Environment variables
- No hardcoded values
š Requirements
- Python 3.8 or newer
SPARQLWrapper
libraryfastmcp
frameworkpydantic
for configurationpython-daemon
for background execution
š Installation
From Source
# Clone the repository
git clone https://github.com/yet-market/yet-sparql-mcp-server.git
cd yet-sparql-mcp-server
# Install dependencies
pip install -r requirements.txt
# Install the package
pip install -e .
From PyPI
pip install mcp-server-sparql
Using the Installation Script
For a full installation with systemd service setup:
# Download the repository
git clone https://github.com/yet-market/yet-sparql-mcp-server.git
cd yet-sparql-mcp-server
# Run the installation script (as root for systemd service)
sudo ./install.sh
The installer automatically:
- Sets up virtual environment and dependencies
- Creates and starts systemd services
- Provides management scripts for easy control
- Configures the HTTP service for nginx integration
Quick Management
After installation, use the provided scripts:
# Check status
./status.sh
# Start/stop services
./start.sh [stdio|http] # Start service (http is default)
./stop.sh [stdio|http] # Stop service
./logs.sh [stdio|http] # View logs
# Examples
./start.sh http # Start HTTP service for nginx
./logs.sh http # View HTTP service logs
./test.sh # Test server functionality
./stop.sh # Stop all services
Testing Your Installation
Test your server with the built-in test suite:
# Run comprehensive tests
./test.sh
# Run specific tests
./test.sh endpoint # Test HTTP connectivity
./test.sh service # Test service status
./test.sh config # Test configuration
./test.sh mcp # Test MCP protocol
./test.sh client # Test with FastMCP client
./test.sh info # Show connection info
The test script validates:
- ā Service status and health
- ā HTTP endpoint connectivity
- ā MCP protocol compliance
- ā Configuration validity
- ā SPARQL endpoint reachability
- ā FastMCP client integration
š Usage
Basic Usage (stdio transport)
Start the server by specifying a SPARQL endpoint:
python server.py --endpoint https://dbpedia.org/sparql
HTTP Server Mode (for nginx/web integration)
Run the server as an HTTP server:
# Basic HTTP server
python server.py --transport http --endpoint https://dbpedia.org/sparql
# Custom host and port
python server.py --transport http --host 0.0.0.0 --port 8080 --endpoint https://dbpedia.org/sparql
# Using environment variables
export MCP_TRANSPORT=http
export MCP_HOST=0.0.0.0
export MCP_PORT=8000
export SPARQL_ENDPOINT=https://dbpedia.org/sparql
python server.py
Running as a Daemon
To run the server as a background process (stdio transport):
python server.py --endpoint https://dbpedia.org/sparql --daemon \
--log-file /var/log/mcp-sparql.log \
--pid-file /var/run/mcp-sparql.pid
To run the HTTP server as a daemon:
python server.py --transport http --host 0.0.0.0 --port 8000 \
--endpoint https://dbpedia.org/sparql --daemon \
--log-file /var/log/mcp-sparql.log \
--pid-file /var/run/mcp-sparql.pid
Using with Systemd
If installed with systemd support:
-
Configure your endpoint in the environment file:
sudo nano /etc/mcp-sparql/env
-
Start the service:
sudo systemctl start sparql-server
-
Enable on boot:
sudo systemctl enable sparql-server
Client Query Examples
After starting the server, you can use it with any MCP-compatible client or through the FastMCP client:
Using FastMCP Client with stdio transport (Python)
import asyncio
from fastmcp.client import Client, PythonStdioTransport
async def query_server():
# Connect to the server
transport = PythonStdioTransport(
script_path="server.py",
args=["--endpoint", "https://dbpedia.org/sparql"]
)
async with Client(transport) as client:
# Execute a SPARQL query
result = await client.call_tool("query", {
"query_string": "SELECT * WHERE { ?s ?p ?o } LIMIT 5",
"format": "simplified"
})
print(result[0].text)
asyncio.run(query_server())
Using FastMCP Client with HTTP transport (Python)
import asyncio
from fastmcp.client import Client, HttpTransport
async def query_server():
# Connect to HTTP server
transport = HttpTransport("http://localhost:8000")
async with Client(transport) as client:
# Execute a SPARQL query
result = await client.call_tool("query", {
"query_string": "SELECT * WHERE { ?s ?p ?o } LIMIT 5",
"format": "simplified"
})
print(result[0].text)
asyncio.run(query_server())
Query with Different Formats
# JSON format (default)
result = await client.call_tool("query", {
"query_string": "SELECT * WHERE { ?s ?p ?o } LIMIT 5",
"format": "json"
})
# Tabular format
result = await client.call_tool("query", {
"query_string": "SELECT * WHERE { ?s ?p ?o } LIMIT 5",
"format": "tabular"
})
Cache Management
# Get cache statistics
cache_stats = await client.call_tool("cache", {"action": "stats"})
# Clear the cache
cache_clear = await client.call_tool("cache", {"action": "clear"})
āļø Configuration
Command-line Arguments
Argument | Description | Default |
---|---|---|
--endpoint URL | SPARQL endpoint URL | Required |
--timeout SECONDS | Request timeout in seconds | 30 |
--format FORMAT | Result format (json, simplified, tabular) | json |
--cache-enabled BOOL | Enable result caching | true |
--cache-ttl SECONDS | Cache time-to-live in seconds | 300 |
--cache-max-size SIZE | Maximum cache size | 100 |
--cache-strategy STRATEGY | Cache replacement strategy (lru, lfu, fifo) | lru |
--pretty-print | Pretty print JSON output | false |
--include-metadata BOOL | Include query metadata in results | true |
--daemon | Run as a background daemon | false |
--log-file FILE | Log file location when running as a daemon | /var/log/mcp-sparql-server.log |
--pid-file FILE | PID file location when running as a daemon | /var/run/mcp-sparql-server.pid |
--transport TRANSPORT | Transport type (stdio or http) | stdio |
--host HOST | Host to bind HTTP server to | localhost |
--port PORT | Port to bind HTTP server to | 8000 |
Environment Variables
Variable | Description | Default |
---|---|---|
SPARQL_ENDPOINT | SPARQL endpoint URL | None (required) |
SPARQL_TIMEOUT | Request timeout in seconds | 30 |
SPARQL_FORMAT | Default result format | json |
SPARQL_CACHE_ENABLED | Enable caching | true |
SPARQL_CACHE_TTL | Cache time-to-live in seconds | 300 |
SPARQL_CACHE_MAX_SIZE | Maximum cache size | 100 |
SPARQL_CACHE_STRATEGY | Cache replacement strategy | lru |
SPARQL_PRETTY_PRINT | Pretty print JSON output | false |
SPARQL_INCLUDE_METADATA | Include query metadata in results | true |
MCP_TRANSPORT | Transport type (stdio or http) | stdio |
MCP_HOST | Host to bind HTTP server to | localhost |
MCP_PORT | Port to bind HTTP server to | 8000 |
š Nginx Integration
When using HTTP transport, you can integrate the server with nginx as a reverse proxy:
1. Start the server in HTTP mode
python server.py --transport http --host localhost --port 8000 --endpoint https://your-sparql-endpoint.com/sparql
2. Configure nginx
Add this configuration to your nginx server block:
upstream mcp_sparql {
server localhost:8000;
# Add more servers for load balancing if needed
# server localhost:8001;
# server localhost:8002;
}
server {
listen 80;
server_name your-domain.com;
location /api/sparql {
proxy_pass http://mcp_sparql;
proxy_set_header Host $host;
proxy_set_header X-Real-IP $remote_addr;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Forwarded-Proto $scheme;
# Optional: Add CORS headers
add_header Access-Control-Allow-Origin *;
add_header Access-Control-Allow-Methods "GET, POST, OPTIONS";
add_header Access-Control-Allow-Headers "Content-Type, Authorization";
}
}
3. Production deployment with systemd
Create a systemd service for HTTP mode:
# /etc/systemd/system/mcp-sparql-http.service
[Unit]
Description=MCP SPARQL Server (HTTP)
After=network.target
[Service]
Type=simple
User=mcp-sparql
Group=mcp-sparql
WorkingDirectory=/opt/mcp-sparql
Environment=MCP_TRANSPORT=http
Environment=MCP_HOST=localhost
Environment=MCP_PORT=8000
Environment=SPARQL_ENDPOINT=https://your-sparql-endpoint.com/sparql
ExecStart=/opt/mcp-sparql/venv/bin/python server.py
Restart=always
RestartSec=3
[Install]
WantedBy=multi-user.target
Enable and start the service:
sudo systemctl enable mcp-sparql-http
sudo systemctl start mcp-sparql-http
š Result Formats
The server supports three different output formats:
1. JSON Format (default)
Returns the standard SPARQL JSON results format with optional metadata.
{
"head": {
"vars": ["s", "p", "o"]
},
"results": {
"bindings": [
{
"s": { "type": "uri", "value": "http://example.org/resource" },
"p": { "type": "uri", "value": "http://example.org/property" },
"o": { "type": "literal", "value": "Example Value" }
}
]
},
"metadata": {
"variables": ["s", "p", "o"],
"count": 1,
"query": "SELECT * WHERE { ?s ?p ?o } LIMIT 1"
}
}
2. Simplified Format
Returns a simplified JSON structure that's easier to work with, converting variable bindings into simple key-value objects.
{
"type": "SELECT",
"results": [
{
"s": "http://example.org/resource",
"p": "http://example.org/property",
"o": "Example Value"
}
],
"metadata": {
"variables": ["s", "p", "o"],
"count": 1,
"query": "SELECT * WHERE { ?s ?p ?o } LIMIT 1"
}
}
3. Tabular Format
Returns results in a tabular format with columns and rows, suitable for table display.
{
"type": "SELECT",
"columns": [
{ "name": "s", "label": "s" },
{ "name": "p", "label": "p" },
{ "name": "o", "label": "o" }
],
"rows": [
[
"http://example.org/resource",
"http://example.org/property",
"Example Value"
]
],
"metadata": {
"variables": ["s", "p", "o"],
"count": 1,
"query": "SELECT * WHERE { ?s ?p ?o } LIMIT 1"
}
}
š Cache Strategies
The server supports three cache replacement strategies:
1. LRU (Least Recently Used)
Evicts the least recently accessed items first. This is the default strategy and works well for most scenarios, as it prioritizes keeping recently accessed items in the cache.
2. LFU (Least Frequently Used)
Evicts the least frequently accessed items first. This strategy is good for scenarios where some queries are much more common than others, as it prioritizes keeping frequently accessed items in the cache.
3. FIFO (First In First Out)
Evicts the oldest items first, regardless of access patterns. This strategy is simpler and can be useful when you want a purely time-based caching approach.
š Advanced SPARQL Examples
The server supports all SPARQL features. Here are some example queries you can try:
Basic Triple Pattern
SELECT * WHERE { ?s ?p ?o } LIMIT 10
Filtering by Property Type
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
SELECT ?subject ?label
WHERE {
?subject rdf:type rdfs:Class ;
rdfs:label ?label .
}
LIMIT 10
Using Regular Expressions
PREFIX foaf: <http://xmlns.com/foaf/0.1/>
SELECT ?person ?name
WHERE {
?person foaf:name ?name .
FILTER(REGEX(?name, "Smith", "i"))
}
LIMIT 10
Complex Query with Multiple Patterns
PREFIX dbo: <http://dbpedia.org/ontology/>
PREFIX dbp: <http://dbpedia.org/property/>
PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
SELECT ?city ?name ?population ?country ?countryName
WHERE {
?city a dbo:City ;
rdfs:label ?name ;
dbo:population ?population ;
dbo:country ?country .
?country rdfs:label ?countryName .
FILTER(?population > 1000000)
FILTER(LANG(?name) = 'en')
FILTER(LANG(?countryName) = 'en')
}
ORDER BY DESC(?population)
LIMIT 10
ā ļø Troubleshooting
Common Issues
- Connection refused: Check that the SPARQL endpoint URL is correct and accessible
- Query timeout: Increase the timeout value with
--timeout
option - Memory issues with large result sets: Add LIMIT clause to your queries or reduce cache size
- Permission denied for log/pid files: Check directory permissions or run with appropriate privileges
Logging
When running in foreground mode, logs are output to the console. When running as a daemon, logs are written to the specified log file (default: /var/log/mcp-sparql-server.log
).
To increase verbosity, you can set the Python logging level in the source code.
š ļø Development
Project Structure
mcp-server-sparql/
āāā sparql_server/ # Main package
ā āāā core/ # Core functionality
ā ā āāā __init__.py # Package exports
ā ā āāā config.py # Configuration management
ā ā āāā server.py # Main SPARQL server
ā āāā formatters/ # Result formatters
ā ā āāā __init__.py # Package exports
ā ā āāā formatter.py # Base formatter class
ā ā āāā json_formatter.py # JSON formatter
ā ā āāā simplified_formatter.py # Simplified JSON formatter
ā ā āāā tabular_formatter.py # Tabular formatter
ā āāā cache/ # Caching implementation
ā ā āāā __init__.py # Package exports
ā ā āāā query_cache.py # Base cache interface
ā ā āāā lru_cache.py # LRU cache implementation
ā ā āāā lfu_cache.py # LFU cache implementation
ā ā āāā fifo_cache.py # FIFO cache implementation
ā āāā __init__.py # Package exports
āāā server.py # Main entry point
āāā setup.py # Package setup
āāā install.sh # Installation script
āāā requirements.txt # Python dependencies
āāā sparql-server.service # Systemd service file
āāā README.md # This file
āāā LICENSE # License file
Running Tests
# Test stdio transport
python test_sparql_server.py
# Test HTTP transport (start server first)
# Terminal 1:
python server.py --transport http --endpoint https://dbpedia.org/sparql
# Terminal 2:
# Run HTTP-specific tests (if available)
š Security Considerations
- The server doesn't implement authentication or authorization - it relies on the security of the underlying SPARQL endpoint
- For production use, consider deploying behind a secure proxy
- Be careful with untrusted queries as they could potentially be resource-intensive
š License
This project is licensed under a dual-license model:
- Open Source: GNU Affero General Public License v3.0 (AGPL-3.0) for open source use
- Commercial: Proprietary commercial license available for commercial or proprietary use
See the file for complete details.
This software was imagined and developed by Temkit Sid-Ali for Yet.lu with AI assistance from Claude (Anthropic), GitHub Copilot/Codex (OpenAI), and GPT-o3 (OpenAI).
š Roadmap
We have exciting plans for the future! Check out our to see what's coming next, including:
- š Enhanced security and authentication
- š Web interface for query exploration
- š¤ AI-powered natural language to SPARQL conversion
- š Advanced data visualization and analytics
- š¢ Enterprise features and scalability improvements
š¤ Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
- Fork the repository
- Create your feature branch (
git checkout -b feature/amazing-feature
) - Commit your changes (
git commit -m 'Add some amazing feature'
) - Push to the branch (
git push origin feature/amazing-feature
) - Open a Pull Request
See our for detailed development setup and guidelines.
š Acknowledgments
This project is built on top of several excellent open source projects:
- FastMCP - High-level Python framework for building MCP servers and clients by Joel Lowin
- SPARQLWrapper - SPARQL endpoint interface for Python by the RDFLib team
- Pydantic - Data validation using Python type hints by Samuel Colvin and the Pydantic team
- python-daemon - Library for making Unix daemon processes
Special thanks to:
- The Model Context Protocol (MCP) community for developing the protocol specification
- Anthropic for their work on AI assistants and the MCP ecosystem
- The semantic web and RDF communities for their foundational work
- W3C for the SPARQL specification and standards
- AI Development Partners:
- Claude (Anthropic) - Primary development assistant for architecture, code implementation, and documentation
- GitHub Copilot/Codex (OpenAI) - Code completion and development acceleration
- GPT-o3 (OpenAI) - Advanced reasoning and problem-solving assistance
š¬ Contact
- GitHub Issues: https://github.com/yet-market/yet-sparql-mcp-server/issues
- Website: https://yet.lu
- Development:
- Commercial licensing: