smart-scraper
Copy
# Extract product listings
just-scrape smart-scraper https://store.example.com/shoes \
-p "Extract all product names, prices, and ratings"
# Enforce output schema + scroll to load more content
just-scrape smart-scraper https://news.example.com \
-p "Get all article headlines and dates" \
--schema '{"type":"object","properties":{"articles":{"type":"array","items":{"type":"object","properties":{"title":{"type":"string"},"date":{"type":"string"}}}}}}' \
--scrolls 5
# Anti-bot bypass for JS-heavy SPAs
just-scrape smart-scraper https://app.example.com/dashboard \
-p "Extract user stats" \
--stealth
search-scraper
Copy
# Research across multiple sources
just-scrape search-scraper "What are the best Python web frameworks in 2025?" \
--num-results 10
# Get raw markdown only (cheaper — 2 credits vs 10)
just-scrape search-scraper "React vs Vue comparison" \
--no-extraction --num-results 5
# Structured output with schema
just-scrape search-scraper "Top 5 cloud providers pricing" \
--schema '{"type":"object","properties":{"providers":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string"},"free_tier":{"type":"string"}}}}}}'
markdownify
Copy
# Convert a blog post to markdown
just-scrape markdownify https://blog.example.com/my-article
# Save to a file
just-scrape markdownify https://docs.example.com/api \
--json | jq -r '.result' > api-docs.md
# Bypass Cloudflare
just-scrape markdownify https://protected.example.com --stealth
crawl
Copy
# Crawl a docs site and collect code examples
just-scrape crawl https://docs.example.com \
-p "Extract all code snippets with their language" \
--max-pages 20 --depth 3
# Crawl only blog pages, skip everything else
just-scrape crawl https://example.com \
-p "Extract article titles and summaries" \
--rules '{"include_paths":["/blog/*"],"same_domain":true}' \
--max-pages 50
# Raw markdown from all pages (no AI extraction, cheaper)
just-scrape crawl https://example.com \
--no-extraction --max-pages 10
scrape
Copy
# Get raw HTML
just-scrape scrape https://example.com
# Geo-targeted + anti-bot bypass
just-scrape scrape https://store.example.com \
--stealth --country-code DE
# Extract branding info (logos, colors, fonts)
just-scrape scrape https://example.com --branding
sitemap
Copy
# List all pages on a site
just-scrape sitemap https://example.com
# Pipe URLs to another tool
just-scrape sitemap https://example.com --json | jq -r '.urls[]'
agentic-scraper
Copy
# Log in and extract dashboard data
just-scrape agentic-scraper https://app.example.com/login \
-s "Fill email with user@test.com,Fill password with secret,Click Sign In" \
--ai-extraction -p "Extract all dashboard metrics"
# Navigate a multi-step form
just-scrape agentic-scraper https://example.com/wizard \
-s "Click Next,Select Premium plan,Fill name with John,Click Submit"
# Persistent browser session across multiple runs
just-scrape agentic-scraper https://app.example.com \
-s "Click Settings" --use-session
generate-schema
Copy
# Generate a schema from a description
just-scrape generate-schema "E-commerce product with name, price, ratings, and reviews array"
# Refine an existing schema
just-scrape generate-schema "Add an availability field" \
--existing-schema '{"type":"object","properties":{"name":{"type":"string"},"price":{"type":"number"}}}'
history
Copy
# Interactive history browser
just-scrape history smartscraper
# Fetch a specific request by ID
just-scrape history smartscraper abc123-def456-7890
# Export last 100 crawl jobs as JSON
just-scrape history crawl --json --page-size 100 \
| jq '.requests[] | {id: .request_id, status}'