scrapedo-mcp-server

umutc/scrapedo-mcp-server

3.2

If you are the rightful owner of scrapedo-mcp-server and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to henry@mcphub.com.

Scrapedo MCP Server is a comprehensive Model Context Protocol server designed for advanced web scraping operations.

Tools
5
Resources
0
Prompts
0

Scrapedo MCP Server

npm version License: MIT MCP Version

Enable Claude Desktop to scrape websites using the Scrapedo API. Simple setup, powerful features.

What is this?

This is an MCP (Model Context Protocol) server that gives Claude Desktop the ability to:

  • Scrape any website (with or without JavaScript)
  • Take screenshots
  • Use proxies from different countries
  • Convert web pages to markdown

Quick Start (2 minutes)

1. Get your Scrapedo API key

Sign up at scrape.do to get your free API key.

2. Setup Claude Desktop

Run this command:

npx scrapedo-mcp-server init

That's it! The tool will automatically configure Claude Desktop for you.

3. Start using it in Claude

Just ask Claude to scrape any website:

"Can you scrape the latest news from https://example.com?"
"Take a screenshot of the Google homepage"
"Get the product prices from this e-commerce site using a US proxy"

Manual Installation Options

Install globally with npm
npm install -g scrapedo-mcp-server
scrapedo-mcp init
Install from source
git clone https://github.com/umutc/scrapedo-mcp-server.git
cd scrapedo-mcp-server
npm install
npm run build
Manual Claude Desktop configuration

Add to ~/Library/Application Support/Claude/claude_desktop_config.json:

{
  "mcpServers": {
    "scrapedo": {
      "command": "npx",
      "args": ["scrapedo-mcp-server", "start"],
      "env": {
        "SCRAPEDO_API_KEY": "your_api_key_here"
      }
    }
  }
}

What Can It Do?

Basic Web Scraping

// Claude can now do this:
scrape("https://example.com")

JavaScript-Rendered Pages

// For modern SPAs and dynamic content:
scrape_with_js("https://app.example.com", {
  waitSelector: ".content-loaded"
})

Screenshots

// Capture any webpage:
take_screenshot("https://example.com", {
  fullPage: true
})

Residential & Mobile Proxies

// Super requests with geo-targeting:
scrape("https://example.com", {
  super: true,     // Switch to residential & mobile pool
  geoCode: "us",   // Country code (defaults to "us" if omitted)
  sessionId: 42    // Optional sticky session
})

Play with Browser Actions

scrape_with_js("https://app.example.com", {
  playWithBrowser: JSON.stringify([
    { action: "wait", selector: ".cookie-banner", timeout: 2000 },
    { action: "click", selector: ".accept" },
    { action: "type", selector: "#search", value: "gaming laptop" },
    { action: "press", key: "Enter" },
    { action: "wait", selector: ".results" }
  ])
})

Screenshots & Async Callbacks

scrape_with_js("https://example.com/product", {
  screenShot: true,          // or fullScreenShot / particularScreenShot
  returnJSON: true,          // enforced automatically when screenshot flags are set
  callback: encodeURI("https://webhook.site/your-endpoint"),
  customHeaders: false,      // override Scrape.do default headers
  forwardHeaders: true
})

Convert to Markdown

// Get clean, readable content:
scrape_to_markdown("https://blog.example.com/article")

Advanced Controls

  • Toggle headers/cookies: customHeaders, extraHeaders, forwardHeaders, setCookies, pureCookies
  • Switch response modes: transparentResponse, returnJSON, showFrames, showWebsocketRequests, output: 'markdown'
  • Combine screenshots with scraping by setting screenShot, fullScreenShot, or particularScreenShot (only one at a time; automatically forces render + returnJSON)
  • Receive results later via callback webhooks (Scrape.do posts JSON payloads to the provided URL)

Project Structure

scrapedo-mcp-server/
├── src/              # TypeScript source code
│   ├── index.ts      # MCP server entry point
│   ├── cli.ts        # CLI tool for setup
│   └── tools/        # Scraping tool implementations
├── dist/             # Compiled JavaScript (generated)
├── package.json      # Project configuration
└── README.md         # You are here

API Costs

Scrapedo charges credits per request type:

FeaturesCredits Usage
Normal Request (Datacenter)1
Datacenter + Headless Browser (JS Render)5
Residential & Mobile Request (super: true)10
Residential & Mobile Request + Headless Browser25

Check your usage anytime by asking Claude: "Check my Scrapedo usage stats"

Common Use Cases

  • Price Monitoring: Track product prices across e-commerce sites
  • News Aggregation: Collect articles from multiple sources
  • Data Research: Gather public information for analysis
  • Content Migration: Export content from websites
  • SEO Analysis: Check how pages render for search engines
  • Competitive Analysis: Monitor competitor websites

FAQ

Q: Do I need to install anything?
A: Just Node.js 18+. Everything else is handled by npx.

Q: How do I update?
A: Run npx scrapedo-mcp-server@latest init to get the newest version.

Q: Is this free?
A: The MCP server is free. Scrapedo offers a free tier with limited credits.

Q: Can I use this with other MCP clients?
A: Yes! This works with any MCP-compatible client, not just Claude Desktop.

Q: How do I see debug logs?
A: Set LOG_LEVEL=DEBUG in your environment or configuration.

Q: How do I disable logging?
A: Set LOG_LEVEL=NONE to stop writing logs entirely.

Troubleshooting

IssueSolution
"Tool not found"Restart Claude Desktop
"401 Unauthorized"Check your API key is correct
"Insufficient credits"Check usage with get_usage_stats
"Empty response"The site might need JavaScript rendering

Need Help?

Contributing

We welcome contributions! See for guidelines.

License

MIT - See file


Built with ❤️ to make web scraping easy in Claude Desktop