Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.scrapegraphai.com/llms.txt

Use this file to discover all available pages before exploring further.

Overview

The ScrapeGraphAI app for Make.com lets you connect any automation scenario to ScrapeGraph’s v2 API β€” no code required. Fetch pages, extract structured data with an AI prompt, run web searches, kick off multi-page crawls, and schedule monitors, all as native Make modules.

ScrapeGraphAI on Make

Install the app from Make’s marketplace

ScrapeGraphAI Dashboard

Get your API key

Installation

  1. Open your Make.com workspace and go to Connections.
  2. Search for ScrapeGraphAI and click Install.
  3. When prompted, enter your SGAI-APIKEY from the dashboard.
  4. Click Save β€” the connection is shared across all modules in your scenario.
ScrapeGraphAI connection dialog in Make

Example: Extract product data into Google Sheets

This scenario runs daily, extracts all products from an Amazon search page, and saves each one as a row in Google Sheets β€” no code required. Full scenario flow:
Full scenario: Schedule β†’ Extract β†’ Iterator β†’ Google Sheets
Step 1 β€” Schedule trigger: Set the scenario to run daily (or any interval). Step 2 β€” Extract module: Configure with your target URL, an extraction prompt, and an output schema.
Extract module configuration
  • URL: The product listing page to extract from
  • Extraction Prompt: Extract all products on the page with their name, price, rating, and number of reviews
  • Output Schema (JSON):
{
  "type": "object",
  "properties": {
    "products": {
      "type": "array",
      "items": {
        "type": "object",
        "properties": {
          "name": {"type": "string"},
          "price": {"type": "string"},
          "rating": {"type": "number"},
          "reviews": {"type": "number"}
        }
      }
    }
  }
}
Step 3 β€” Iterator: Add a Flow Control β†’ Iterator module and set the Array field to {{2.json.products}}. This loops through each product and passes it to the next module one at a time.
Iterator module configuration
Step 4 β€” Google Sheets: Add a Row: Map each field from the Iterator output:
  • Name β†’ {{value.name}}
  • Price β†’ {{value.price}}
  • Rating β†’ {{value.rating}}
  • Reviews β†’ {{value.reviews}}
Google Sheets module configuration
Result: Every product on the page is saved as a separate row.
Google Sheets result with extracted product data

Modules

Scrape

Fetch a URL and return its content in one or more formats: Markdown, HTML, links, images, a plain-text summary, or branding elements.
Scrape module configuration
FieldDescription
URLThe page to fetch
FormatOutput format β€” Markdown, HTML, Links, Images, Summary, Branding
HTML ModeRendering mode β€” Normal, Reader, Prune

Extract

Send a URL to ScrapeGraph and get back structured JSON β€” driven by a natural-language prompt and an optional JSON schema.
Extract module configuration
FieldDescription
Website URLPage to extract from
Extraction PromptNatural-language instruction, e.g. Extract product name and price
Output Schema (JSON)Optional JSON schema to enforce output shape
HTML Processing ModeNormal, Reader, or Prune

Run a web search and get page content returned inline β€” optionally with AI extraction applied to each result.
Search module configuration
FieldDescription
QuerySearch query string
Number of Results1–20, default 3
FormatContent format for each result
Extraction PromptOptional AI extraction applied to each page
Output Schema (JSON)Optional schema β€” requires Extraction Prompt
Country Code2-letter country code for localised results

Start Crawl

Start an async multi-page crawl from an entry URL. Returns a Crawl Job ID to pass into Get Crawl Status.
Start Crawl module configuration
FieldDescription
URLEntry point for the crawl
FormatOutput format per page
Max PagesCap on total pages crawled (1–1000)
Max DepthHow many link levels deep to traverse
Max Links Per PageMaximum links to follow per page
Include / Exclude PatternsURL glob patterns, e.g. /blog/*

Get Crawl Status

Poll the status and results of a crawl job started by Start Crawl.
Get Crawl Status module
FieldDescription
Crawl Job IDThe id output from Start Crawl β€” map with {{1.id}}
Returns status (running / completed / failed) and a pages array when completed.
Add a Tools β†’ Sleep module (60 seconds) between Start Crawl and Get Crawl Status to give the crawl time to finish before polling. For large crawls, use two separate scenarios with a Make Data Store to persist the job ID.

Create Monitor

Schedule ScrapeGraph to fetch a URL on a recurring cron schedule and detect changes between runs.
Create Monitor module configuration
FieldDescription
URLPage to watch
Monitor NameOptional display name
Interval (cron)Cron expression β€” see table below
FormatContent format to capture
Webhook URLOptional URL to POST results to on each tick
Common cron expressions
ScheduleCron
Every hour0 * * * *
Every 6 hours0 */6 * * *
Daily at 09:00 UTC0 9 * * *
Weekly on Monday0 9 * * 1
Run Create Monitor once manually to set up the monitor, then use Get Monitor Activity in a separate scheduled scenario to fetch what changed.

Get Monitor Activity

Fetch the latest activity ticks from an existing monitor.
Get Monitor Activity module
FieldDescription
Monitor IDThe id returned by Create Monitor
LimitNumber of ticks to return (1–100, default 20)
Returns a ticks array where each entry has changed (boolean), diffs, status, and createdAt.

Deprecated modules

The following modules from the v1 integration are still visible but no longer functional. Use the v2 modules above instead.
DeprecatedReplacement
[Deprecated] SmartScrapeScrape
[Deprecated] MarkdownifyScrape (Markdown format)
[Deprecated] Generate JSON SchemaExtract