Documentation Index
Fetch the complete documentation index at: https://docs.scrapegraphai.com/llms.txt
Use this file to discover all available pages before exploring further.
Transition from v1 to v2
If youβre onn8n-nodes-scrapegraphai@0.x, this is your migration checkpoint.
Before anything else, update the community node in n8n at Settings β Community Nodes β n8n-nodes-scrapegraphai β Update to 1.0.2 (or later). Your existing SGAI-APIKEY works as-is β no re-auth needed.
Method-by-method migration
Use this table to map old resources to the new ones. Details and field changes follow below.| v1 | v2 | Notes |
|---|---|---|
Markdownify | Scrape with format Markdown | One Scrape node with a Markdown format entry replaces Markdownify. |
SmartScraper | Extract | Same job β structured extraction from a URL. |
SearchScraper | Search | Renamed; the prompt-style query field is now called Query. |
SmartCrawler (single call) | Crawl.Start, then Crawl.GetStatus, Crawl.Stop, Crawl.Resume, Crawl.Delete | Crawl is async β Start returns a job ID, poll Get Status. |
Scrape | Scrape | Same name, expanded β multi-format per call (markdown, HTML, JSON, screenshot, links, summary, branding). |
AgenticScraper | Removed | Use Extract with Fetch Config (mode JS, stealth, wait) for hard pages. |
| β | Monitor (new) | Cron-scheduled fetches with diff detection and webhooks. |
| β | History (new) | Look up past results by scrapeRefId. |
| β | Credit (new) | Check remaining credits and plan. |


Step-by-step rebuild
1. Markdownify β Scrape
Before: A dedicated Markdownify resource that always returned markdown.
After: Use the Scrape resource with one Markdown format entry. Same job, more flexible β you can mix in HTML, Links, Summary, or Branding in the same call.
2. SmartScraper β Extract
Before (v1): Website URL + User Prompt, plus optional flat fields like Render Heavy JS and Number of Scrolls.
After (v2): URL + Prompt, optional Schema (JSON) behind a Use JSON Schema toggle. All fetch knobs move into a single Fetch Config collection shared across every resource.
| v1 field | v2 field |
|---|---|
Website URL (websiteUrl) | URL |
User Prompt (userPrompt) | Prompt |
Output Schema (outputSchema) | Schema (JSON) (behind Use JSON Schema toggle) |
Render Heavy JS (renderHeavyJs) | Fetch Config β Mode set to JS |
Number of Scrolls (numberOfScrolls) | Fetch Config β Scrolls |
Stealth, Wait (Ms), Timeout (Ms), Country, Headers (JSON), Cookies (JSON).

3. SearchScraper β Search
Before: User Prompt + a few flat options.
After: Query (the search string) plus optional Rollup Prompt for AI extraction across all fetched results, optional Schema (JSON) behind a toggle, and new fields like Time Range and Location (Country Code).
| v1 field | v2 field |
|---|---|
User Prompt (userPrompt) | Query |
Output Schema (outputSchema) | Schema (JSON) (behind Use JSON Schema toggle) |
| β | Rollup Prompt (new β AI extraction across results) |
| β | Time Range / Location (Country Code) (new) |
4. SmartCrawler β Crawl jobs
Before: A single synchronous SmartCrawler operation.
After: Crawl is explicitly async. Start the job, then poll. Five operations are exposed: Start, Get Status, Stop, Resume, Delete.
A typical chain in n8n:
- Crawl β Start β returns a
cronId - Wait node (~60s)
- Crawl β Get Status β returns the
pages[]array - (Optional) Split Out + History β Get β fetch full content per crawled page
5. Output shape
Downstream nodes (Set, IF, HTTP Request) that reference v1 paths like$json.result.markdown will break β v2 returns a different shape.
The new node ships an Output parameter on every content-producing operation (Scrape, Extract, Search) with three modes: Simplified, Raw, Selected Fields. Pick Simplified when migrating β itβs the closest match to v1.

What else changed in v2
- New Fetch Config knobs that didnβt exist in v1:
Stealth,Wait (Ms),Timeout (Ms),Country,Headers (JSON),Cookies (JSON) - New resources:
Monitor(cron + diff + webhook),History(look up past results byscrapeRefId),Credit(check usage) - Async crawl model with five lifecycle ops instead of one synchronous call
- AI-Agent friendly β every content-producing op exposes
Simplified/Raw/Selected Fieldsoutput modes - Cleaner credentials test β n8n hits
GET /api/creditsto verify keys
Recommended path
- Update the community node: Settings β Community Nodes β
n8n-nodes-scrapegraphaiβ Update to1.0.2(or later) - Open each affected workflow β v1 ScrapeGraphAI nodes will surface as deprecated or fail to execute
- Drop in fresh ScrapeGraphAI nodes and pick the matching v2 resource from the migration table
- Re-map fields per the step-by-step rebuild above
- Set Output to
Simplified(closest to v1) - Test the node, fix downstream expressions, delete the v1 node
FAQ
- Will my workflows keep running until I touch them? Yes β until the next execution opens the v1 node, which then fails against the deprecated v1 API.
- Can I run v1 and v2 side-by-side? No β same package, version-pinned.
- Self-hosted vs n8n Cloud? Self-hosted: bump the version in Community Nodes. n8n Cloud doesnβt yet allow community nodes.
- I used Agentic Scraper β what now? Use Extract with Fetch Config β Mode = JS plus Stealth and Wait (Ms).
Related guides
n8n integration guide
Full reference for the v2 node β every resource, operation, and field
API transition guide
Python / JavaScript / REST migration β the underlying API changes
GitHub repo
Source code, issue tracker, release notes
Legacy v1 npm
The last 0.x release (deprecated)
