crawlio

๐Ÿ“ฆ Batch Scrape

Scrape multiple webpages in a single batch request. This endpoint is ideal for bulk extraction jobs where you need to process multiple URLs with shared options.

๐Ÿงฐ Using with SDKs

Prefer code over curl? Crawlio offers official SDKs for seamless integration with your stack:

๐Ÿ“– View full usage docs: ๐Ÿ‘‰ Node.js SDK Docs ๐Ÿ‘‰ Python SDK Docs

We are working on an extensive documentation on our SDKs. Thanks for your cooperation!

Cost

NameCostType
ScrapeNumber of scraped pagesScrape

๐Ÿ“ฆ POST /scrape/batch

Scrape multiple webpages in a single batch request. This endpoint is ideal for bulk extraction jobs where you need to process multiple URLs with shared options.

๐Ÿ“ฅ Request

Endpoint:

POST https://crawlio.xyz/api/scrape/batch

Headers:

Authorization: Bearer YOUR_API_KEY  
Content-Type: application/json

Request Body Parameters:

FieldTypeRequiredDescription
urlarray of stringsโœ… YesA list of URLs to scrape in one batch.
optionsobjectโŒ NoAdditional config such as headers, timeouts, etc. (future use)
excludearray of stringsโŒ NoCSS selectors to remove elements from all pages in the batch.
markdownbooleanโŒ NoReturn scraped content in Markdown format for each page.

๐Ÿงพ Example Request

POST /scrape/batch
Content-Type: application/json
Authorization: Bearer YOUR_API_KEY

{
  "url": [
    "https://example.com/page1",
    "https://example.com/page2"
  ],
  "exclude": [".ads", "footer"],
  "markdown": true
}

๐Ÿ“ค Response

On success, Crawlio will respond with a unique batch identifier which you can use to retrieve results when the job is complete.

FieldTypeDescription
batchIdstringUnique ID for the batch scraping job.

๐Ÿ“ฆ Example Response

{
  "batchId": "batch_789ghi"
}

You can use the batch ID with a job result or status endpoint (if available) to monitor progress and retrieve full output once processing completes.

See Postman Collection for more detailed information on getting the batch data and job status :


What and Why?

The Batch Scrape feature is designed to let you extract content from multiple URLs in one go, making it ideal for bulk scraping tasks where efficiency and consistency are key.

Instead of sending individual requests for each page, you can group your targets and handle them together โ€” saving time, reducing overhead, and simplifying result tracking.

Use Cases:

  • ๐Ÿ›’ Scraping product detail pages from a category listing
  • ๐Ÿ“ Collecting multiple blog posts or news articles
  • ๐Ÿ“ˆ Monitoring a set of competitor pages
  • ๐Ÿ”„ Feeding a batch of URLs from a search or database into a scraper

Key Capabilities:

  • โœ… Submit multiple URLs at once for scraping
  • ๐Ÿงน Optionally exclude elements like footers, ads, or navigation bars
  • ๐Ÿ“ Return content as Markdown for easier storage or post-processing
  • ๐Ÿช Perfect for pairing with /search or internal lists of known URLs

The /scrape/batch endpoint is a fast, scalable solution when you need to extract structured content from many pages with shared configuration.

On this page