Use ScrapeGraphAPI through the Model Context Protocol
Play around with our MCP Server on MCP.so’s playground or on Klavis AI.
env SCRAPEGRAPH_API_KEY=your-api-key npx -y @scrapegraphai/mcp-server
If you are using Windows and are running into issues, try cmd /c "set SCRAPEGRAPH_API_KEY=your-api-key && npx -y @scrapegraphai/mcp-server"
Replace your-api-key
with your ScrapeGraphAI API key. If you don’t have one yet, you can create an account and get it from ScrapeGraphAI Dashboard
After adding, refresh the MCP server list to see the new tools. The Composer Agent will automatically use ScrapeGraphAI MCP when appropriate, but you can explicitly request it by describing your web scraping needs. Access the Composer via Command+L (Mac), select “Agent” next to the submit button, and enter your query.
./codeium/windsurf/model_config.json
:
http://localhost:3000/sse
or https://mcp.scrapegraphai.com/{SCRAPEGRAPH_API_KEY}/sse
Ctrl + Shift + P
and typing Preferences: Open User Settings (JSON)
.
.vscode/mcp.json
in your workspace. This will allow you to share the configuration with others:
SCRAPEGRAPH_API_KEY
: Your ScrapeGraphAI API key
SCRAPEGRAPH_API_URL
SCRAPEGRAPH_API_URL
(Optional): Custom API endpoint for self-hosted instances
http://localhost:8000
for local developmentSSE_LOCAL
: Set to true
to run in SSE mode locallyLOG_LEVEL
: Set logging level (debug, info, warn, error)PORT
: Custom port for SSE mode (default: 3000)scrapegraph_scrape
)scrapegraph_batch_scrape
)scrapegraph_check_batch_status
)scrapegraph_search
)scrapegraph_crawl
)scrapegraph_extract
)urls
: Array of URLs to extract information fromprompt
: Custom prompt for the LLM extractionsystemPrompt
: System prompt to guide the LLMschema
: JSON schema for structured data extractionallowExternalLinks
: Allow extraction from external linksenableWebSearch
: Enable web search for additional contextincludeSubdomains
: Include subdomains in extractionscrapegraph_deep_research
)query
(string, required): The research question or topic to explore.maxDepth
(number, optional): Maximum recursive depth for crawling/search (default: 3).timeLimit
(number, optional): Time limit in seconds for the research session (default: 120).maxUrls
(number, optional): Maximum number of URLs to analyze (default: 50).data.finalAnalysis
)scrapegraph_generate_llmstxt
)url
(string, required): The base URL of the website to analyze.maxUrls
(number, optional): Max number of URLs to include (default: 10).showFullText
(boolean, optional): Whether to include llms-full.txt contents in the response.data.llmstxt
and/or data.llmsfulltxt
)npm test