gaiaaiagent/regen-registry-review-mcp
If you are the rightful owner of regen-registry-review-mcp and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to dayong@mcphub.com.
The Registry Review MCP Server automates the manual document review process for carbon credit project registration, significantly reducing the time required for review while maintaining a complete audit trail and structured outputs.
Registry Review MCP Server
MCP server that automates carbon credit project registration reviews through an eight-stage workflow.
Overview
The Registry Review MCP Server transforms a 6-8 hour manual document review into a guided workflow where AI handles document organization, data extraction, and consistency checking while humans provide expertise, judgment, and final approval.
Core Capabilities:
- Document discovery and intelligent classification
- Requirement mapping with semantic matching
- Evidence extraction with page citations
- Cross-document validation (dates, land tenure, project IDs)
- Structured report generation (Markdown + JSON)
Quick Start
# Install dependencies
uv sync
# Run the MCP server
uv run python -m registry_review_mcp.server
# Run tests (expensive tests excluded by default)
uv run pytest
Claude Desktop Integration
Add to your claude_desktop_config.json:
{
"mcpServers": {
"registry-review": {
"command": "uv",
"args": [
"--directory", "/path/to/regen-registry-review-mcp",
"run", "python", "-m", "registry_review_mcp.server"
]
}
}
}
The Eight-Stage Workflow
Each stage produces artifacts for human verification before proceeding. The workflow follows a collaboration model where AI handles tedious document processing while humans provide expertise and final judgment.
Stage A: Initialize
Create a review session with project metadata and load the checklist template.
/A-initialize Botany Farm 2022-2023, /path/to/documents
Output: Session ID, project metadata, loaded checklist (23 requirements for Soil Carbon v1.2.2)
Stage B: Document Discovery
Scan the documents directory, extract file metadata, and classify each document by type.
/B-document-discovery
Agent Actions:
- Recursively scan for PDFs, shapefiles, GeoJSON, spreadsheets
- Classify documents (project plan, baseline report, monitoring report, land tenure, etc.)
- Generate document inventory with confidence scores
Human Actions: Review classifications, mark documents as in-scope/ignored/pinned
Output: Document inventory with normalized names, types, and source references
Stage C: Requirement Mapping
Connect discovered documents to specific checklist requirements using semantic matching.
/C-requirement-mapping
Agent Actions:
- Parse checklist into structured requirements with expected evidence types
- Analyze documents and suggest requirement → document mappings
- Flag requirements with no plausible matches
Human Actions: Confirm/reject suggested mappings, manually add missing mappings
Output: Mapping matrix linking each requirement to 0+ documents with confidence scores
Stage D: Evidence Extraction
Extract key data points and text snippets from mapped documents.
/D-evidence-extraction
Agent Actions:
- Parse document content (PDF text, tables, metadata)
- Extract 0-3 evidence snippets per requirement with page citations
- Extract structured data: dates, locations, ownership info, numerical values
Human Actions: Review snippets, delete irrelevant ones, add manual notes
Output: Evidence database with snippets, citations, and structured data points
Stage E: Cross-Validation
Verify consistency, completeness, and compliance across all extracted evidence.
/E-cross-validation
Validation Checks:
- Date Alignment: Sampling dates within ±120 days of imagery dates
- Land Tenure: Owner names consistent across documents (fuzzy matching)
- Project ID: Consistent project identifiers across all documents
- Completeness: Each requirement has mapped documents with sufficient evidence
Output: Validation results with pass/warning/fail flags and coverage statistics
Stage F: Report Generation
Produce structured, auditable Registry Review Report.
/F-report-generation
Output Formats:
- Markdown: Human-readable report with executive summary, per-requirement findings, citations
- JSON: Machine-readable for audit trails and downstream systems
Report Contents: Project metadata, coverage statistics, requirement findings with evidence snippets, validation results, items requiring human review
Stage G: Human Review
Expert validation, annotation, and revision handling.
/G-human-review
Human Actions:
- Review flagged items requiring judgment
- Override agent assessments where expert knowledge differs
- Request revisions from proponent if gaps identified
- Make final determination: Approve / Conditional / Reject / On Hold
Output: Finalized report with human annotations and approval decision
Stage H: Completion
Finalize and archive the review.
/H-completion
Agent Actions:
- Lock finalized report
- Generate archive package with audit trail
- Prepare data for on-chain registration (if approved)
Output: Locked report, complete audit trail, archived session
Quick Example
/A-initialize Botany Farm 2022-2023, /home/user/projects/botany-farm
/B-document-discovery
/C-requirement-mapping
/D-evidence-extraction
/E-cross-validation
/F-report-generation
Each stage auto-selects the most recent session, so you can run them in sequence without specifying session IDs.
Available Tools
Session Management:
create_session- Create new review sessionload_session/list_sessions/delete_session- Session lifecyclestart_review- Quick-start: create session + discover documentslist_example_projects- List available test projects
File Upload:
create_session_from_uploads- Create session from uploaded filesupload_additional_files- Add files to existing sessionstart_review_from_uploads- Full workflow from uploads
Document Processing:
discover_documents- Scan and classify project documentsadd_documents- Add document sources to sessionextract_pdf_text- Extract text from PDFsextract_gis_metadata- Extract GIS shapefile metadata
Requirement Mapping:
map_all_requirements- Semantic mapping to documentsconfirm_mapping/remove_mapping- Manual mapping adjustmentsget_mapping_status- View mapping statistics
Evidence & Validation:
extract_evidence- Extract evidence for all requirementsmap_requirement- Map and extract for single requirement
Configuration
Copy .env.example to .env and configure:
# Required for LLM-powered extraction
REGISTRY_REVIEW_ANTHROPIC_API_KEY=sk-ant-api03-...
REGISTRY_REVIEW_LLM_EXTRACTION_ENABLED=true
# Optional
REGISTRY_REVIEW_LLM_MODEL=claude-sonnet-4-5-20250929
REGISTRY_REVIEW_LOG_LEVEL=INFO
See .env.example for all configuration options including chunking, image processing, cost management, and validation settings.
Project Structure
regen-registry-review-mcp/
├── src/registry_review_mcp/
│ ├── server.py # MCP entry point
│ ├── config/ # Settings management
│ ├── extractors/ # PDF and LLM extraction
│ ├── models/ # Pydantic models
│ ├── prompts/ # A-H workflow prompts
│ ├── services/ # Document processing
│ ├── tools/ # MCP tool implementations
│ └── utils/ # State, cache, helpers
├── data/
│ ├── checklists/ # Methodology requirements (JSON)
│ ├── sessions/ # Active sessions (gitignored)
│ └── cache/ # Cached extractions (gitignored)
├── tests/ # Test suite
├── docs/
│ └── specs/ # Workflow specifications
└── examples/ # Test data (Botany Farm)
Development
# Run tests (fast tests only - expensive tests excluded)
uv run pytest
# Format and lint
uv run black src/ tests/
uv run ruff check src/ tests/
Test Markers:
smoke- Critical path tests (<1s)expensive- Tests with API costs (excluded by default)marker- PDF extraction tests (slow, 8GB+ RAM)accuracy- Ground truth validation tests
See pytest.ini for marker configuration.
Requirements
- Python >= 3.10
- uv package manager
- 4GB RAM minimum (8GB recommended for large PDFs)
License
Copyright 2025 Regen Network Development, Inc.