๐ฆ Batch Scrape
Scrape multiple webpages in a single batch request. This endpoint is ideal for bulk extraction jobs where you need to process multiple URLs with shared options.
๐งฐ Using with SDKs
Prefer code over curl? Crawlio offers official SDKs for seamless integration with your stack:
- Node.js SDK (npm) โ Perfect for backend automation, agents, and JS projects.
- Python SDK (PyPI) โ Ideal for data science, AI/ML workflows, and scripting.
๐ View full usage docs: ๐ Node.js SDK Docs ๐ Python SDK Docs
We are working on an extensive documentation on our SDKs. Thanks for your cooperation!
Cost
Name | Cost | Type |
---|---|---|
Scrape | Number of scraped pages | Scrape |
๐ฆ POST /scrape/batch
Scrape multiple webpages in a single batch request. This endpoint is ideal for bulk extraction jobs where you need to process multiple URLs with shared options.
๐ฅ Request
Endpoint:
Headers:
Request Body Parameters:
Field | Type | Required | Description |
---|---|---|---|
url | array of strings | โ Yes | A list of URLs to scrape in one batch. |
options | object | โ No | Additional config such as headers, timeouts, etc. (future use) |
exclude | array of strings | โ No | CSS selectors to remove elements from all pages in the batch. |
markdown | boolean | โ No | Return scraped content in Markdown format for each page. |
๐งพ Example Request
๐ค Response
On success, Crawlio will respond with a unique batch identifier which you can use to retrieve results when the job is complete.
Field | Type | Description |
---|---|---|
batchId | string | Unique ID for the batch scraping job. |
๐ฆ Example Response
You can use the batch ID with a job result or status endpoint (if available) to monitor progress and retrieve full output once processing completes.
See Postman Collection for more detailed information on getting the batch data and job status :
What and Why?
The Batch Scrape feature is designed to let you extract content from multiple URLs in one go, making it ideal for bulk scraping tasks where efficiency and consistency are key.
Instead of sending individual requests for each page, you can group your targets and handle them together โ saving time, reducing overhead, and simplifying result tracking.
Use Cases:
- ๐ Scraping product detail pages from a category listing
- ๐ Collecting multiple blog posts or news articles
- ๐ Monitoring a set of competitor pages
- ๐ Feeding a batch of URLs from a search or database into a scraper
Key Capabilities:
- โ Submit multiple URLs at once for scraping
- ๐งน Optionally exclude elements like footers, ads, or navigation bars
- ๐ Return content as Markdown for easier storage or post-processing
- ๐ช Perfect for pairing with
/search
or internal lists of known URLs
The /scrape/batch
endpoint is a fast, scalable solution when you need to extract structured content from many pages with shared configuration.
๐ท๏ธ Scrape Endpoint
Initiate a one-time scrape of a specific webpage. This endpoint allows fine-grained control over what content is retrieved and how it's formatted.
๐ Crawl Endpoint
Initiate a full website crawl starting from a given URL. Crawlio will recursively follow links and extract content from each page, subject to the options you provide.