code-reasoning

code-reasoning

3.6

If you are the rightful owner of code-reasoning and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to henry@mcphub.com.

A Model Context Protocol (MCP) server that enhances Claude's ability to solve complex programming tasks through structured, step-by-step thinking.

Code Reasoning MCP Server

A Model Context Protocol (MCP) server that enhances Claude's ability to solve complex programming tasks through structured, step-by-step thinking.

Quick Installation

  1. Configure Claude Desktop by editing:

    • macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
    • Windows: %APPDATA%\Claude\claude_desktop_config.json
    • Linux: ~/.config/Claude/claude_desktop_config.json
    {
      "mcpServers": {
        "code-reasoning": {
          "command": "npx",
          "args": ["-y", "@mettamatt/code-reasoning"]
        }
      }
    }
    
  2. Configure VS Code:

{
  "mcp": {
    "servers": {
      "code-reasoning": {
        "command": "npx",
        "args": ["-y", "@mettamatt/code-reasoning"]
      }
    }
  }
}

Usage

  1. To trigger this MCP, append this to your chat messages:

    Use sequential thinking to reason about this.
    
  2. Use ready-to-go prompts that trigger Code-Reasoning:

  • Click the "+" icon in the Claude Desktop chat window, or in Claude Code type /help to see the specific commands.
  • Select "Add from Code Reasoning" from the available tools
  • Choose a prompt template and fill in the required information
  • Submit the form to add the prompt to your chat message and hit return

See the for details on using the prompt templates.

Command Line Options

  • --debug: Enable detailed logging
  • --help or -h: Show help information

Key Features

  • Programming Focus: Optimized for coding tasks and problem-solving
  • Structured Thinking: Break down complex problems into manageable steps
  • Thought Branching: Explore multiple solution paths in parallel
  • Thought Revision: Refine earlier reasoning as understanding improves
  • Safety Limits: Automatically stops after 20 thought steps to prevent loops
  • Ready-to-Use Prompts: Pre-defined templates for common development tasks

Documentation

Detailed documentation available in the docs directory:

  • : Examples of sequential thinking with the MCP server
  • : All configuration options for the MCP server
  • : Using and customizing prompts with the MCP server
  • : Testing information

Project Structure

├── index.ts                  # Entry point
├── src/                      # Implementation source files
└── test/                     # Testing framework

Prompt Evaluation

The Code Reasoning MCP Server includes a prompt evaluation system that assesses Claude's ability to follow the code reasoning prompts. This system allows:

  • Testing different prompt variations against scenario problems
  • Verifying parameter format adherence
  • Scoring solution quality

To use the prompt evaluation system, run:

npm run eval

Prompt Comparison and Development

Significant effort went into developing the optimal prompt for the Code Reasoning server. The current implementation uses the HYBRID_DESIGN prompt, which emerged as the winner from our evaluation process.

We compared four different prompt designs:

Prompt DesignDescription
SEQUENTIALThe original sequential thinking prompt design
DEFAULTThe baseline prompt previously used in the server
CODE_REASONING_0_30An experimental variant focusing on code-specific reasoning
HYBRID_DESIGNA refined design incorporating the best elements of other approaches

Our evaluation across seven diverse programming scenarios showed that HYBRID_DESIGN outperformed other prompts:

ScenarioHYBRID_DESIGNCODE_REASONING_0_30DEFAULTSEQUENTIAL
Algorithm Selection89%82%92%88%
Bug Identification92%91%88%94%
Multi-Stage Implementation87%67%82%87%
System Design Analysis87%87%83%82%
Code Debugging Task96%87%91%93%
Compiler Optimization83%78%72%78%
Cache Strategy87%88%89%87%
Average89%83%85%87%

The HYBRID_DESIGN prompt demonstrates the highest average solution quality (89%) and the most consistent performance across all scenarios, with no scores below 80%. It also produces the most thoughts. The src/server.ts file has been updated to use this optimal prompt design.

Personally, I think the biggest improvement was adding this to the end of the prompt: "✍️ End each thought by asking: "What am I missing or need to reconsider?"

See for more details on the prompt evaluation system.

License

This project is licensed under the MIT License. See the LICENSE file for details.