AdamGustavsson/ClaimsMCP
If you are the rightful owner of ClaimsMCP and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to dayong@mcphub.com.
Claimify is a local Model Context Protocol (MCP) server for extracting factual claims from text using a research-based methodology.
Claimify: Research-Based Claim Extraction via MCP
An implementation of the "Claimify" methodology for factual claim extraction, delivered as a local Model Context Protocol (MCP) server. This tool implements the multi-stage claim extraction approach detailed in the academic paper "Towards Effective Extraction and Evaluation of Factual Claims" by Metropolitansky & Larson (2025).
Promps from the paper have been modified for use with Structured Outputs. THIS IS NOT AN OFFICIAL IMPLEMENTATION.
Overview
Claimify extracts verifiable, decontextualized factual claims from text using a sophisticated four-stage pipeline:
- Sentence Splitting: Breaks text into individual sentences with surrounding context
- Selection: Filters for sentences containing verifiable propositions, excluding opinions and speculation
- Disambiguation: Resolves ambiguities or discards sentences that cannot be clarified
- Decomposition: Breaks down sentences into atomic, self-contained factual claims
The tool uses OpenAI's structured outputs feature exclusively for improved reliability and exposes its functionality through the Model Context Protocol, making it available to MCP-compatible clients like Cursor and Claude Desktop.
Features
- Research-based methodology: Implements the peer-reviewed Claimify approach
- Structured outputs: Uses OpenAI's structured outputs for reliable, type-safe responses
- MCP integration: Seamlessly integrates with development environments
- Robust parsing: Handles various text formats including lists and paragraphs
- Context-aware: Uses surrounding sentences to resolve ambiguities
- Multi-language support: Preserves original language while extracting claims
- Resource storage: Automatically stores extracted claims as MCP resources for easy retrieval
- Comprehensive logging: Detailed logging of all LLM calls, responses, and pipeline stages
- Production-ready: Includes error handling, monitoring, and configuration management
Requirements
- OpenAI API: Requires an OpenAI API key (if your MCP host does not support sampling (Github Copilot in vsCode does))
- Compatible Model: Must use a model that supports structured outputs:
gpt-4o(recommended)gpt-4o-mini(faster and cheaper)
- Python 3.10+: For proper type hints and Pydantic support
Quick Start
1. Installation
# Clone the repository
git clone <repository-url>
cd ClaimsMCP
# Create and activate a virtual environment
python -m venv claimify-env
source claimify-env/bin/activate # On Windows: claimify-env\Scripts\activate
# Install dependencies
pip install -r requirements.txt
# Download required NLTK data (done automatically on first run)
python -c "import nltk; nltk.download('punkt_tab')"
2. Configuration
Create a .env file in the project root:
# Copy the example file
cp env.example .env
Edit .env and add your API key:
# API Keys
OPENAI_API_KEY="your-openai-api-key-here"
# LLM Configuration
LLM_MODEL="gpt-4o-2024-08-06" # Model that supports structured outputs
# Logging Configuration
LOG_LLM_CALLS="true" # Set to "false" to disable logging
LOG_OUTPUT="stderr" # "stderr" or "file" - where to send logs
LOG_FILE="claimify_llm.log" # Used only if LOG_OUTPUT="file"
MCP Client Configuration
For Cursor
- Open Cursor and navigate to Settings > MCP
- Click "Add a New Global MCP Server"
- Add the following configuration to your MCP settings file (usually
~/.cursor/mcp.json):
{
"mcpServers": {
"claimify-local": {
"command": "/path/to/your/claimify-env/bin/python",
"args": [
"/path/to/your/project/claimify_server.py"
]
}
}
}
- Replace the paths with the absolute paths to your Python executable and server script
The "Claimify Extraction Server" should now appear as a connected tool in your MCP-enabled chat.
Usage Examples
Once configured, you can use the tool in your MCP client:
Using the Extract Claims Tool
Using the Prompts
The server exposes two prompts to help verify and document extracted claims:
1. Verify Single Claim (verify_claim)
Provides a pre-built prompt that instructs the LLM to verify a single factual claim against external sources.
Arguments:
claim_text(required): The decontextualized factual claim to check.
Behavior:
- LLM is instructed to search authoritative sources (scholarly publications, reputable news outlets, official organizations).
- Returns one of three statuses:
- VERIFIED: Claim is clearly supported by reliable sources (provides at least 3 references with URLs + justification)
- UNCERTAIN: Claim may be correct but lacks precision, has ambiguity, or has limited/conflicting evidence
- DISPUTED: Claim is demonstrably false or contradicted by reliable sources
- Sources must not be fabricated; preference for primary references.
Example retrieval (conceptual – actual call depends on client API):
get_prompt(name="verify_claim", arguments={"claim_text": "Python was first publicly released in 1991."})
Example expected LLM response format:
**Claim:** Python was first publicly released in 1991.
**Status:** IN_PROGRESS
[After research...]
**Claim:** Python was first publicly released in 1991.
**Status:** VERIFIED
**Evidence:**
- Source 1: [Python.org Release History](https://www.python.org/doc/versions/) - Official Python release history page confirms the initial public release year
- Source 2: [Computer History Museum](https://www.computerhistory.org/collections/catalog/102726710) - Archive referencing Python's early development
- Source 3: [Wikipedia - Python](https://en.wikipedia.org/wiki/Python_(programming_language)) - Encyclopedia entry citing original release year
If uncertain:
**Claim:** Stockholm has 800,000 inhabitants.
**Status:** UNCERTAIN
**Evidence:**
- Source 1: [Statistics Sweden](https://www.scb.se/en/) - Reports varying population figures depending on whether measuring city proper, municipality, or metropolitan area
**Analysis:**
The claim lacks specificity about which definition of "Stockholm" is being referenced (city proper ~975k, municipality ~975k, or urban area ~1.6M as of 2023). The figure of 800,000 may have been accurate for certain definitions at specific time periods, but without temporal and geographic context, full verification is not possible.
If disputed:
**Claim:** The Earth is flat.
**Status:** DISPUTED
**Analysis:**
This claim contradicts overwhelming scientific evidence. The Earth's spherical shape has been confirmed by satellite imagery, space missions, and centuries of astronomical observations. Reliable sources universally reject this claim.
2. Create Claims Report (create_claims_report)
Generates an initial CLAIMS.md file with all claims marked as TODO. Claims can then be verified incrementally, updating their status through the workflow: TODO → IN_PROGRESS → VERIFIED/UNCERTAIN/DISPUTED.
Workflow:
- Initial Creation: All claims start with status TODO
- During Verification: Update individual claims to IN_PROGRESS
- After Verification: Update to VERIFIED, UNCERTAIN, or DISPUTED with evidence
Arguments:
- None (attach the extraction resource to the context in VS Code)
Behavior:
- Parses all claims from the extraction resource attached in the context
- Creates CLAIMS.md with all claims marked as TODO
- Provides a template structure ready for incremental verification
Example usage: When viewing an extraction resource in VS Code, attach it to the prompt context. The prompt will generate an initial CLAIMS.md file ready for verification.
Initial CLAIMS.md structure:
# Claims Report
**Extraction ID:** extraction_1_1730678400
**Generated:** 2025-11-03
**Total Claims:** 5
**TODO:** 5
**In Progress:** 0
**Verified:** 0
**Uncertain:** 0
**Disputed:** 0
---
## Claims
### Claim 1
**Text:** Apple Inc. was founded in 1976.
**Status:** TODO
---
### Claim 2
**Text:** Steve Jobs co-founded Apple Inc.
**Status:** TODO
---
...
After verification updates:
### Claim 1
**Text:** Apple Inc. was founded in 1976.
**Status:** VERIFIED
**Evidence:**
- Source 1: [Wikipedia - Apple Inc.](https://en.wikipedia.org/wiki/Apple_Inc.) - States company was founded in 1976
- Source 2: [Apple Official](https://www.apple.com/about/) - Corporate history confirms 1976 founding
- Source 3: [Britannica](https://www.britannica.com/topic/Apple-Inc) - Encyclopedia entry validates founding year
---
### Claim 2
**Text:** Stockholm has 800,000 inhabitants.
**Status:** UNCERTAIN
**Evidence:**
- Source 1: [Statistics Sweden](https://www.scb.se/en/) - Reports varying figures depending on definition
**Analysis:**
The claim lacks specificity about which geographic definition and time period. Population varies significantly between city proper (~975k), municipality (~975k), and metropolitan area (~1.6M) as of 2023. The 800k figure may have been accurate historically for certain definitions.
---
### Claim 3
**Text:** The company invented smartphones.
**Status:** DISPUTED
**Analysis:**
While Apple popularized smartphones with the iPhone in 2007, they did not invent smartphones. Earlier devices like the IBM Simon (1994) and BlackBerry devices (early 2000s) preceded the iPhone. The claim conflates innovation/popularization with invention.
---
...
Note: The server only supplies the prompts; external searching depends on the client/model capabilities.
Example 1: Simple Factual Text
Input: "The American flag contains 50 stars and 13 stripes."
Output: [
"The American flag contains 50 stars [representing the 50 states] and 13 stripes [representing the original 13 colonies].",
"The American flag was designed in 1777",
"The American flag has been modified 27 times"
]
Example 2: Mixed Fact and Opinion
Input: "Apple Inc. was founded in 1976 by Steve Jobs, Steve Wozniak, and Ronald Wayne. The company is incredibly innovative and has the best products in the world."
Output: [
"Apple Inc. was founded in 1976 by Steve Jobs, Steve Wozniak, and Ronald Wayne."
]
(Note: The subjective content about being "incredibly innovative" and having "the best products" is filtered out)
Example 3: Multi-language Support
Input: "String-systemet är en prisbelönt ikon som kombinerar elegant och minimalistisk design med ett brett utbud av färger och storlekar. Nisse Strinning skapade första hyllan redan 1949."
Output: [
"String-systemet [ett hyllsystem] är en prisbelönt ikon [inom design]",
"String-systemet kombinerar elegant och minimalistisk design med ett brett utbud av färger och storlekar",
"Nisse Strinning skapade den första String-hyllan [String-systemet] 1949"
]
(Note: Content preserved in original Swedish, with contextual clarifications added in brackets)
Accessing Extracted Claims as Resources
Each extraction generates two kinds of resources:
- Aggregate Extraction Resource (
claim://extraction_<n>_<timestamp>)- Contains metadata (timestamp, preview, question) and the full list of claims
- Returns JSON format
- Individual Claim Resources (
claim://<slug>)- Each claim is accessible via a unique slug (URL-safe identifier derived from claim text)
- Returns plain text (the claim itself)
Aggregate Extraction JSON Example
{
"id": "extraction_1_1730678400",
"timestamp": "2025-11-03T14:30:00.123456",
"question": "What is the history of Apple?",
"text_preview": "Apple Inc. was founded in 1976 by Steve Jobs...",
"claims": [
"Apple Inc. was founded in 1976 by Steve Jobs, Steve Wozniak, and Ronald Wayne."
],
"claim_count": 1
}
URI: claim://apple-inc-was-founded-in-1976-by-steve-jobs
Individual Claim Resource Examples
URI pattern: claim://<slug>
Examples:
claim://apple-inc-was-founded-in-1976-by-steve-jobs-steve-wozniak-and-ronaldclaim://stockholm-is-the-capital-of-swedenclaim://python-was-first-publicly-released-in-1991
Content: Plain text of the claim (no JSON wrapper)
Benefits of per-claim resources:
- Direct access: Retrieve any claim by its slug
- Simple format: Plain text, no parsing needed
- Unique identifiers: Each claim has a stable, readable URI
- Easy citation: Link directly to individual claims
Project Structure
ClaimsMCP/
├── README.md # This file
├── requirements.txt # Python dependencies
├── env.example # Environment configuration template
├── claimify_server.py # Main MCP server script
├── llm_client.py # LLM client with structured outputs support
├── pipeline.py # Core claim extraction pipeline
├── structured_models.py # Pydantic models for structured outputs
├── structured_prompts.py # Optimized prompts for structured outputs
├── setup.py # Package setup configuration
├── test_claimify.py # Test suite for the claim extraction pipeline
└── LICENSE # Apache 2.0 license
Architecture
The system follows a modular architecture with structured outputs:
- MCP Server: Exposes the claim extraction as a tool via the Model Context Protocol
- ClaimifyPipeline: Orchestrates the multi-stage extraction process using structured outputs
- LLMClient: Handles communication with OpenAI API using structured outputs and Pydantic models
- Structured Models: Pydantic models that define the expected response format for each stage
- Stage Functions: Individual functions for Selection, Disambiguation, and Decomposition
- Prompt Management: Simplified prompts optimized for structured outputs
Structured Outputs Benefits
The implementation uses OpenAI's structured outputs feature, which provides:
- Type Safety: Responses are automatically validated against Pydantic models
- Reliability: No more regex parsing failures or malformed JSON
- Explicit Refusals: Safety-based refusals are programmatically detectable
- Consistency: Guaranteed adherence to the expected response schema
- Performance: Reduced need for retry logic and error handling
Configuration Options
| Environment Variable | Description | Default | Options |
|---|---|---|---|
LLM_MODEL | Specific model to use | gpt-4o-2024-08-06 | Models supporting structured outputs |
OPENAI_API_KEY | OpenAI API key | None | Your API key |
LOG_LLM_CALLS | Enable detailed logging of all LLM interactions | true | true, false |
LOG_OUTPUT | Where to send log output | stderr | stderr, file |
LOG_FILE | Log file name (used when LOG_OUTPUT=file) | claimify_llm.log | Any filename |
Troubleshooting
Common Issues
-
"Model does not support structured outputs" error
- Ensure you're using a compatible model:
gpt-4o-2024-08-06,gpt-4o-mini, orgpt-4o - Update your
.envfile:LLM_MODEL=gpt-4o-2024-08-06
- Ensure you're using a compatible model:
-
"API key not set" error
- Ensure your
.envfile exists and contains the correct OpenAI API key - Check that the key starts with
sk-
- Ensure your
-
"NLTK punkt tokenizer not found"
- Run:
python -c "import nltk; nltk.download('punkt_tab')"orpython -c "import nltk; nltk.download('punkt')"
- Run:
-
MCP client can't connect
- Check that the paths in your MCP configuration are absolute and correct
- Ensure your Python virtual environment is activated
- Verify the server script is executable:
chmod +x claimify_server.py
-
No claims extracted
- Check the logs for detailed information about each pipeline stage
- Ensure the input text contains verifiable factual statements
- Try with simpler, more direct factual sentences first
Development
To extend or modify the system:
- Adding new response fields: Update the Pydantic models in
structured_models.py - Modifying prompts: Edit the prompts in
structured_prompts.py - Adding new stages: Create new functions in
pipeline.pyfollowing the existing pattern - Testing: Use the built-in logging to debug pipeline behavior
The structured outputs approach makes the system much more reliable and easier to debug compared to traditional text parsing methods.
License
This project is licensed under the Apache License 2.0 - see the file for details.
References
Metropolitansky & Larson (2025). "Towards Effective Extraction and Evaluation of Factual Claims"
Support
For issues related to:
- Setup and configuration: Check this README and the troubleshooting section
- MCP integration: Refer to the Model Context Protocol documentation
- Research methodology: Consult the original Claimify paper