
Overview
SearchScraper is our advanced LLM-powered search service that intelligently searches and aggregates information from multiple web sources. Using state-of-the-art language models, it understands your queries and extracts relevant information across the web, providing comprehensive answers with full source attribution.Try SearchScraper instantly in our interactive playground - no coding required!
Getting Started
Quick Start
Parameters
Parameter | Type | Required | Description |
---|---|---|---|
apiKey | string | Yes | The ScrapeGraph API Key. |
prompt | string | Yes | A textual description of what you want to achieve. |
numResults | number | No | Number of websites to search (3-20). Default: 3. Higher = deeper research. |
schema | object | No | The Pydantic or Zod object that describes the structure and format of the response |
mock | boolean | No | Optional parameter to enable mock mode. When set to true, the request will return mock data instead of performing an actual search. Useful for testing and development. Default: false |
NEW: You can now control the number of websites to search (3-20) for deeper research. More sites = more credits. See advanced usage below!
Advanced: Website Limits & Credit Costs
Advanced: Website Limits & Credit Costs
You can now configure how many websites SearchScraper will search (from 3 up to 20). This allows you to balance research depth and credit usage.
- Default: 3 websites (30 credits)
- Enhanced: 5 websites (50 credits)
- Maximum: 20 websites (200 credits)
- 3 websites: 30 credits
- 5 websites: 50 credits
- 10 websites: 100 credits
- 20 websites: 200 credits
num_results
(Python) or numResults
(JS) in your request.Example Response
Example Response
request_id
: Unique identifier for tracking your requeststatus
: Current status of the search (“completed”, “running”, “failed”)result
: The extracted data in structured JSON formatreference_urls
: Source URLs for verificationerror
: Error message (if any occurred during search)
Key Features
Multi-Source Search
Intelligent search across multiple reliable web sources
AI Understanding
Advanced LLM models for accurate information extraction
Structured Output
Clean, structured data in your preferred format
Source Attribution
Full transparency with reference URLs
Use Cases
Research & Analysis
- Academic research and fact-finding
- Market research and competitive analysis
- Technology trend analysis
- Industry insights gathering
Data Aggregation
- Product research and comparison
- Company information compilation
- Price monitoring across sources
- Technology stack analysis
Content Creation
- Fact verification and citation
- Content research and inspiration
- Data-driven article writing
- Knowledge base building
Want to learn more about our AI-powered search technology? Visit our main website to discover how we’re revolutionizing web research.
Other Functionality
Retrieve a previous request
If you know the response id of a previous request you made, you can retrieve all the information.Parameters
Parameter | Type | Required | Description |
---|---|---|---|
apiKey | string | Yes | The ScrapeGraph API Key. |
requestId | string | Yes | The request ID associated with the output of a previous searchScraper request. |
Custom Schema Example
Define exactly what data you want to extract using Pydantic or Zod:Advanced Schema Usage
The schema system in SearchScraper is a powerful way to ensure you get exactly the data structure you need. Here are some advanced techniques for using schemas effectively:Nested Schemas
You can create complex nested structures to capture hierarchical data:Schema Validation Rules
Enhance data quality by adding validation rules to your schema:Quality Improvement Tips
To get the highest quality results from SearchScraper, follow these best practices:1. Detailed Field Descriptions
Always provide clear, detailed descriptions for each field in your schema:2. Structured Prompts
Combine schemas with well-structured prompts for better results:3. Data Validation
Implement comprehensive validation to ensure data quality:4. Error Handling
Implement robust error handling for schema validation:Async Support
Example of using the async searchscraper functionality to search for information concurrently:Integration Options
Official SDKs
- Python SDK - Perfect for data science and backend applications
- JavaScript SDK - Ideal for web applications and Node.js
AI Framework Integrations
- LangChain Integration - Use SearchScraper in your LLM workflows
- LlamaIndex Integration - Build powerful search and QA systems
- CrewAI Integration - Create AI agents with search capabilities
Best Practices
Query Optimization
- Be specific in your prompts
- Use descriptive queries
- Include relevant context
- Specify time-sensitive requirements
Schema Design
- Start with essential fields
- Use appropriate data types
- Add field descriptions
- Make optional fields nullable
- Group related information
Rate Limiting
- Implement reasonable delays between requests
- Use async clients for better performance
- Monitor your API usage
Example Projects
Check out our cookbook for real-world examples:API Reference
For detailed API documentation, see:Support & Resources
Documentation
Comprehensive guides and tutorials
API Reference
Detailed API documentation
Community
Join our Discord community
GitHub
Check out our open-source projects
Ready to Start?
Sign up now and get your API key to begin searching and extracting data with SearchScraper!