Give Claude, Cursor, and Windsurf the ability to scrape any website, extract structured data, and take screenshots — with automatic anti-bot bypass.
Get Started Free → $1 free balance on signup — up to 5,000 scrapes.
claude mcp add alterlab -- npx -y alterlab-mcp-serverThen set your API key: export ALTERLAB_API_KEY=sk_live_... or add it to .claude.json (see full setup below).
# Add to .cursor/mcp.json — see full config belownpx -y @smithery/cli install alterlab-mcp-server --client claudeClaude's built-in WebFetch tool and open-source browser MCP servers fail on most real-world websites. They cannot bypass Cloudflare, render JavaScript SPAs, or extract structured data.
AlterLab replaces broken fetch tools with one MCP server that actually works:
| Capability | WebFetch / fetch() | Browser MCP | AlterLab MCP |
|---|---|---|---|
| Anti-bot bypass (Cloudflare, DataDome, Akamai) | No | Partial | Yes — automatic |
| JavaScript rendering (React, Angular, Vue SPAs) | No | Yes (slow) | Yes — headless Chromium |
| Structured data extraction (JSON, Schema.org) | No | No | Yes — built-in profiles |
| Smart tier escalation (cheapest method first) | N/A | N/A | Yes — saves 60-80% |
| Residential proxy rotation (195+ countries) | No | No | Yes |
| Screenshot and PDF capture | No | Screenshot only | Yes — both |
| OCR text extraction from images | No | No | Yes |
| Cost per request | Free (but fails) | Free (but slow) | From $0.0002 |
AlterLab uses a multi-tier scraping architecture. It automatically selects the cheapest method capable of fetching each URL:
- Curl ($0.0002/req) — Direct HTTP for static pages, RSS feeds, public APIs
- HTTP ($0.0003/req) — TLS fingerprint rotation for moderately protected sites
- Stealth ($0.0005/req) — Browser impersonation for Cloudflare/DataDome-protected sites
- Light JS ($0.0007/req) — Lightweight JS extraction from server-rendered HTML
- Browser ($0.001/req) — Full headless Chromium for JavaScript-heavy SPAs
Auto mode starts at Tier 1 and escalates only when blocked. Most websites resolve at Tiers 1-2, so $1 gets you 1,000 to 5,000 scrapes depending on the sites you target.
Add to your Claude config file (~/.claude.json for Claude Code, or Settings for Claude Desktop):
{
"mcpServers": {
"alterlab": {
"command": "npx",
"args": ["-y", "alterlab-mcp-server"],
"env": {
"ALTERLAB_API_KEY": "sk_live_your_key_here"
}
}
}
}Add to .cursor/mcp.json in your project root:
{
"mcpServers": {
"alterlab": {
"command": "npx",
"args": ["-y", "alterlab-mcp-server"],
"env": {
"ALTERLAB_API_KEY": "sk_live_your_key_here"
}
}
}
}Add to Windsurf MCP settings (~/.codeium/windsurf/mcp_config.json):
{
"mcpServers": {
"alterlab": {
"command": "npx",
"args": ["-y", "alterlab-mcp-server"],
"env": {
"ALTERLAB_API_KEY": "sk_live_your_key_here"
}
}
}
}npx -y @smithery/cli install alterlab-mcp-server --client claude- Sign up free — $1 free balance on signup
- Go to Dashboard → API Keys and copy your key
- Paste it into the
ALTERLAB_API_KEYfield in your MCP config
Scrape a URL and return its content as markdown, text, HTML, or JSON. Automatically handles anti-bot protection with tier escalation. Returns markdown by default — optimized for LLM context windows.
"Scrape https://www.amazon.com/dp/B0BSHF7WHW and summarize the product"
| Parameter | Type | Default | Description |
|---|---|---|---|
url |
string | required | URL to scrape |
mode |
auto | html | js | pdf | ocr |
auto |
Scraping mode |
formats |
(text|json|html|markdown)[] |
["markdown"] |
Output formats |
render_js |
boolean | false |
Use headless browser (+3 credits) |
use_proxy |
boolean | false |
Premium proxy (+1 credit) |
proxy_country |
string | — | ISO country code for geo-targeting (e.g., US, DE) |
wait_for |
string | — | CSS selector to wait for before extraction |
timeout |
number | 90 |
Timeout in seconds (1-300) |
include_raw_html |
boolean | false |
Include raw HTML alongside formatted content |
session_id |
string (UUID) | — | Stored session ID for authenticated scraping |
cookies |
Record<string, string> |
— | Inline cookies for one-off authenticated requests |
Extract structured fields from any webpage using pre-built profiles or custom JSON Schema. Returns clean JSON — ready for databases, spreadsheets, or downstream processing.
"Extract the product name, price, and rating from this Amazon page"
| Parameter | Type | Default | Description |
|---|---|---|---|
url |
string | required | URL to extract from |
extraction_profile |
enum | auto |
Profile: product, article, job_posting, faq, recipe, event |
extraction_schema |
object | — | Custom JSON Schema for structured output |
extraction_prompt |
string | — | Natural language extraction instructions |
render_js |
boolean | false |
Use headless browser |
use_proxy |
boolean | false |
Premium proxy |
Extraction profiles:
- Product — name, price, currency, rating, reviews, availability, images, description
- Article — title, author, published date, body text, featured image
- Job Posting — title, company, location, salary, description, requirements
- FAQ — question-answer pairs
- Recipe — ingredients, instructions, prep time, servings
- Event — name, date, location, description, organizer
Take a full-page screenshot of any URL. Returns a PNG image directly in the conversation — no URLs to copy, no files to download.
"Take a screenshot of our landing page at https://alterlab.io"
| Parameter | Type | Default | Description |
|---|---|---|---|
url |
string | required | URL to screenshot |
wait_for |
string | — | CSS selector to wait for before capture |
wait_until |
enum | networkidle |
networkidle, domcontentloaded, or load |
Check how much a scrape will cost before running it. Returns the predicted tier, cost per request, and confidence level.
"How much would it cost to scrape linkedin.com?"
| Parameter | Type | Default | Description |
|---|---|---|---|
url |
string | required | URL to estimate |
mode |
enum | auto |
Scraping mode |
render_js |
boolean | false |
Include JS rendering cost |
use_proxy |
boolean | false |
Include proxy cost |
Check your account balance, total deposited, and total spent. No parameters needed.
"Check my AlterLab balance"
List all stored sessions for authenticated scraping. Sessions contain cookies for specific domains, allowing you to scrape content behind login walls.
"List my stored sessions"
Create a new stored session with cookies from a logged-in browser. The session is stored securely and can be reused across multiple scrape requests.
"Create an Amazon session with these cookies: session-id=abc123, session-token=xyz789"
| Parameter | Type | Default | Description |
|---|---|---|---|
name |
string | required | Human-readable name (e.g., "My Amazon Account") |
domain |
string | required | Domain (e.g., "amazon.com") |
cookies |
Record<string, string> |
required | Cookie key-value pairs |
user_agent |
string | — | Browser User-Agent to use with this session |
Check whether a stored session is still active and its cookies are valid.
"Is my Amazon session still valid?"
| Parameter | Type | Default | Description |
|---|---|---|---|
session_id |
string (UUID) | required | Session ID to validate |
Permanently delete a stored session and its cookies.
"Delete session abc-123-def"
| Parameter | Type | Default | Description |
|---|---|---|---|
session_id |
string (UUID) | required | Session ID to delete |
AlterLab MCP supports scraping pages that require authentication. This enables AI agents to access user-specific content like order histories, account dashboards, and member-only pricing.
- Create a session with cookies from a logged-in browser using
alterlab_create_session - Scrape authenticated pages by passing the
session_idtoalterlab_scrape - Manage sessions with list, validate, and delete tools
User: "What's my Prime member price for this product?"
Claude: [calls alterlab_list_sessions → finds Amazon session]
Claude: [calls alterlab_scrape with session_id for authenticated pricing]
Claude: "The Prime member price is $24.99 (public price: $34.99)"
- Stored sessions (
session_id): Best for repeated access to the same domain. Create once, reuse across requests. - Inline cookies (
cookies): Best for one-off authenticated requests where you don't need to save the session.
Ask Claude to scrape and analyze websites in real-time:
- "Scrape the top 5 results from this Google search and summarize them"
- "Extract all product prices from this Amazon category page"
- "Compare the pricing pages of these 3 competitors"
Let Cursor or Windsurf fetch live data while building:
- "Scrape this API documentation page and generate TypeScript types from it"
- "Extract the color palette from this website and create a Tailwind config"
- "Screenshot this design and recreate it in React"
Use Claude to analyze content at scale:
- "Scrape this blog post and suggest improvements for SEO"
- "Extract all FAQ entries from this help center and create a structured dataset"
- "Compare our landing page to the competitor's and identify gaps"
Build agentic workflows that watch the web:
- "Check if this product is back in stock"
- "Scrape this page daily and alert me when the price drops below $50"
- "Monitor this job board for new senior engineering positions"
No subscriptions. No monthly minimums. Add balance and use it whenever you need it.
| Tier | Method | Cost per Request | Use Case |
|---|---|---|---|
| Curl | Direct HTTP | $0.0002 | Static pages, RSS feeds, public APIs |
| HTTP | TLS fingerprinting | $0.0003 | Sites with basic bot detection |
| Stealth | Browser impersonation | $0.0005 | Cloudflare, DataDome, PerimeterX protected sites |
| Light JS | JSON extraction | $0.0007 | Server-rendered pages needing structured data |
| Browser | Headless Chromium | $0.001 | Full JavaScript SPAs (React, Angular, Vue) |
| Add-On | Extra Cost | Description |
|---|---|---|
| JavaScript Rendering | +$0.0006 | Headless Chromium for dynamic content |
| Screenshot Capture | +$0.0002 | Full-page PNG screenshot |
| Premium Proxy | +$0.0002 | Geo-targeted residential proxy (195+ countries) |
| OCR Text Extraction | +$0.001 | Extract text from images on the page |
$1 = 5,000 light scrapes. New accounts get $1 free balance on signup.
| Variable | Required | Default | Description |
|---|---|---|---|
ALTERLAB_API_KEY |
Yes | — | Your API key (get one free) |
ALTERLAB_API_URL |
No | https://api.alterlab.io |
API base URL (for self-hosted or development) |
Install the AlterLab MCP server. For Claude Code, run claude mcp add alterlab -- npx -y alterlab-mcp-server and set your ALTERLAB_API_KEY. For Claude Desktop, Cursor, or Windsurf, add the JSON config block to your MCP settings file. Once configured, your AI assistant can scrape any URL, extract structured data, and take screenshots directly in conversation.
Yes. AlterLab automatically handles Cloudflare, DataDome, PerimeterX, Akamai, and other anti-bot systems. It uses a multi-tier approach that starts with the cheapest method and escalates only when blocked. You don't need to configure anything — anti-bot bypass is fully automatic.
MCP (Model Context Protocol) is Anthropic's open standard for connecting AI assistants to external tools and data sources. An MCP server is a small program that exposes tools — like web scraping — that Claude, Cursor, or Windsurf can call during a conversation. The AlterLab MCP server gives your AI assistant 5 tools: scrape, extract, screenshot, estimate cost, and check balance.
AlterLab starts at $0.0002 per request — 5-20x cheaper than most scraping APIs — because it only uses expensive browser rendering when a site actually requires it. Smart tier escalation means you pay for what each site needs, not the maximum. AlterLab also includes built-in structured data extraction with pre-built profiles (product, article, job posting, etc.) at no extra cost.
Yes. AlterLab handles all major e-commerce anti-bot protection. Use the alterlab_extract tool with extraction_profile: "product" to get structured JSON: product name, price, currency, rating, review count, availability, and images — ready for analysis, comparison, or data pipelines.
Yes. With AlterLab MCP installed in Cursor, you can ask it to scrape API documentation, library docs, or any reference page and generate TypeScript types, API clients, or component code from the live content. This is more reliable than relying on the LLM's training data, which may be outdated.
Yes. Use render_js: true or set mode: "js" to enable full headless Chromium rendering. AlterLab renders the complete page including all JavaScript, waits for dynamic content to load, then extracts content from the fully rendered DOM. This works for React, Angular, Vue, Next.js, and any other JavaScript framework.
Use markdown (the default). It preserves document structure — headings, tables, lists, links — while being 60-80% smaller than raw HTML. Claude, GPT-4, and other LLMs process markdown significantly better than HTML. AlterLab's markdown output is specifically optimized for LLM context windows.
Free-tier accounts have rate limits. Adding any balance removes rate limits. The MCP server includes automatic retry with exponential backoff for transient rate limit errors (429).
Yes. The MCP server processes one request at a time through the conversation interface, but you can build agentic workflows that scrape many URLs sequentially. For batch processing, use the AlterLab API directly or the n8n integration.
The MCP server returns helpful error messages with suggested next actions:
| Error | What Happens | Suggested Action |
|---|---|---|
| 401 Unauthorized | Invalid API key | Check ALTERLAB_API_KEY is set correctly |
| 402 Insufficient Credits | Balance too low | Run alterlab_check_balance, add funds |
| 403 Forbidden | Site blocked the request | Try render_js: true + use_proxy: true |
| 429 Rate Limited | Too many requests | Automatic retry with backoff |
| 504 Gateway Timeout | Scrape took too long | Increase timeout, simplify request |
git clone https://github.com/RapierCraft/alterlab-mcp-server.git
cd alterlab-mcp-server
npm install
npm run build