Scrapeless API
Dashboard
Dashboard
  1. Crawl
  • User
    • Get User Info
      GET
  • Scraping Browser
    • CDP API
    • Connect
      WSS
    • Running sessions
      GET
    • Live URL
      GET
  • Scraping API
    • shopee
      • Actor List
      • Shopee Product
      • Shopee Search
      • Shopee Rcmd
    • br sites
      • Solucoes cnpjreva
      • Solucoes certidaointernet
      • Servicos receita
      • Consopt
    • amazon
      • API Parameters
      • product
      • seller
      • keywords
    • google search
      • API Parameters
      • Google Search
      • Google Images
      • Google Local
    • google trends
      • API Parameters
      • AutoComplete
      • Interest Over Time
      • Compared Breakdown By Region
      • Interest By Subregion
      • Related Queries
      • Related Topics
      • Trending Now
    • google flights
      • API Parameters
      • Round trip
      • One way
      • Multi-city
    • google flights chart
      • API Parameters
      • chart
    • google maps
      • API Parameters(Google Maps)
      • API Parameters(Google Maps Autocomplete)
      • API Parameters(Google Maps Contributor Reviews)
      • API Parameters(Google Maps Directions)
      • API Parameters(Google Maps Reviews)
      • Google Maps
      • Google Maps Autocomplete
      • Google Maps Contributor Reviews
      • Google Maps Directions
      • Google Maps Reviews
    • google scholar
      • API Parameters(Google Scholar)
      • API Parameters(Google Scholar Author)
      • API Parameters(Google Scholar Cite)
      • API Parameters(Google Scholar Profiles)
      • Google Scholar
      • Google Scholar Author
      • Google Scholar Cite
      • Google Scholar Profiles
    • google jobs
      • API Parameters
      • Google Jobs
    • google shopping
      • API Parameters
      • Google Shopping
    • google hotels
      • API Parameters
      • Supported Google Vacation Rentals Property Types
      • Supported Google Hotels Property Types
      • Supported Google Vacation Rentals Amenities
      • Supported Google Hotels Amenities
      • Google Hotels
    • google news
      • API Parameters
      • Google News
    • google lens
      • API Parameters
      • Google Lens
    • google finance
      • API Parameters
      • Google Finance
      • Google Finance Markets
    • google product
      • API Parameters
      • Google Product
    • google play store
      • API Parameters
      • Supported Google Play Categories
      • Google Play Games
      • Google Play Books
      • Google Play Movies
      • Google Play Product
      • Google Play Apps
    • google ads
      • API Parameters
      • Google Ads
    • Scraper Request
      POST
    • Scraper GetResult
      GET
  • Universal Scraping API
    • JS Render Docs
    • JS Render
    • Web Unlocker
    • Akamaiweb Cookie
    • Akamaiweb Sensor
  • Crawler
    • Scrape
      • Scrape a single URL
      • Scrape multiple URLs
      • Cancel a batch scrape job
      • Get the status of a scrape
      • Get the status of a batch scrape job
      • Get the errors of a batch scrape job
    • Crawl
      • Crawl multiple URLs based on options
        POST
      • Cancel a crawl job
        DELETE
      • Get the status of a crawl job
        GET
      • Get the errors of a crawl job
        GET
  • Public
    • actor status
    • actor status
  1. Crawl

Crawl multiple URLs based on options

POST
/api/v1/crawler/crawl
Crawling
Last modified:2025-06-04 10:04:20

Request

Authorization
Add parameter in header
x-api-token
Example:
x-api-token: ********************
Body Params application/json
url
string <uri>
required
The base URL to start crawling from
limit
integer 
optional
Maximum number of pages to crawl. Default limit is 10000.
Default:
10000
excludePaths
array[string]
optional
URL pathname regex patterns that exclude matching URLs from the crawl. For example, if you set "excludePaths": ["blog/.*"] for the base URL firecrawl.dev, any results matching that pattern will be excluded, such as https://www.scrapeless.com/blog/firecrawl-launch-week-1-recap.
includePaths
array[string]
optional
URL pathname regex patterns that include matching URLs in the crawl. Only the paths that match the specified patterns will be included in the response. For example, if you set "includePaths": ["blog/.*"] for the base URL firecrawl.dev, only results matching that pattern will be included, such as https://www.scrapeless.com/blog/firecrawl-launch-week-1-recap.
maxDepth
integer 
optional
Maximum depth to crawl relative to the base URL. Basically, the max number of slashes the pathname of a scraped URL may contain.
Default:
10
maxDiscoveryDepth
integer 
optional
Maximum depth to crawl based on discovery order. The root site and sitemapped pages has a discovery depth of 0. For example, if you set it to 1, and you set ignoreSitemap, you will only crawl the entered URL and all URLs that are linked on that page.
ignoreSitemap
boolean 
optional
Ignore the website sitemap when crawling
Default:
false
ignoreQueryParameters
boolean 
optional
Do not re-scrape the same path with different (or none) query parameters
Default:
false
deduplicateSimilarURLs
boolean 
optional
Controls whether similar URLs should be deduplicated.
regexOnFullURL
boolean 
optional
Controls whether the regular expression should be applied to the full URL.
allowBackwardLinks
boolean 
optional
Enables the crawler to navigate from a specific URL to previously linked pages.
Default:
false
allowExternalLinks
boolean 
optional
Allows the crawler to follow links to external websites.
Default:
false
delay
number 
optional
Delay in seconds between scrapes. This helps respect website rate limits.
scrapeOptions
object (ScrapeOptions) 
optional
formats
array[string]
optional
Formats to include in the output.
Allowed values:
markdownhtmlrawHtmllinksscreenshotscreenshot@fullPagejson
Default:
markdown
onlyMainContent
boolean 
optional
Only return the main content of the page excluding headers, navs, footers, etc.
Default:
true
includeTags
array[string]
optional
Tags to include in the output.
excludeTags
array[string]
optional
Tags to exclude from the output.
headers
object 
optional
Headers to send with the request. Can be used to send cookies, user-agent, etc.
waitFor
integer 
optional
Specify a delay in milliseconds before fetching the content, allowing the page sufficient time to load.
Default:
0
timeout
integer 
optional
Timeout in milliseconds for the request
Default:
30000
browserOptions
object (BrowserOptions) 
optional
session_name
string 
optional
Set a name for your session to facilitate searching and viewing in the historical session list.
session_ttl
string 
optional
Controls the session duration and automatically closes the browser instance after timeout. Measured in seconds (s), defaults to 180 seconds (3 minutes), customizable between 60 seconds (1 minute) and 900 seconds (recommended maximum 15 minutes, but longer times can be set). Once the specified TTL is reached, the session will expire and Scraping Browser will close the browser instance to free resources.
session_recording
string 
optional
Whether to enable session recording. When enabled, the entire browser session execution process will be automatically recorded, and after the session is completed, it can be replayed and viewed in the historical session list details. Defaults to false.
proxy_country
string 
optional
Sets the target country/region for the proxy, sending requests via an IP address from that region. You can specify a country code (e.g., US for the United States, GB for the United Kingdom, ANY for any country). See country codes for all supported options.
proxy_url
string 
optional
Used to set the browser’s proxy URL, for example: http://user:pass@ip:port. If this parameter is set, all other proxy_* parameters will be ignored.
💡Custom proxy functionality is currently only available to Enterprise and Enterprise Enhanced subscription users Upgrade Now
💡Enterprise-level custom users can contact us to use custom proxies.
fingerprint
string 
optional
A browser fingerprint is a nearly unique “digital fingerprint” created using your browser and device configuration information, which can be used to track your online activity even without cookies. Fortunately, configuring fingerprints in Scraping Browser is optional. We offer deep customization of browser fingerprints, such as core parameters like browser user agent, time zone, language, and screen resolution, and support extending functionality through custom launch parameters. Suitable for multi-account management, data collection, and privacy protection scenarios, using scrapeless’s own Chromium browser completely avoids detection. By default, our Scraping Browser service generates a random fingerprint for each session. Reference
Example
{
    "url": "http://example.com",
    "limit": 10000,
    "excludePaths": [
        "string"
    ],
    "includePaths": [
        "string"
    ],
    "maxDepth": 10,
    "maxDiscoveryDepth": 0,
    "ignoreSitemap": false,
    "ignoreQueryParameters": false,
    "deduplicateSimilarURLs": true,
    "regexOnFullURL": true,
    "allowBackwardLinks": false,
    "allowExternalLinks": false,
    "delay": 0,
    "scrapeOptions": {
        "formats": [
            "markdown"
        ],
        "onlyMainContent": true,
        "includeTags": [
            "string"
        ],
        "excludeTags": [
            "string"
        ],
        "headers": {},
        "waitFor": 0,
        "timeout": 30000
    },
    "browserOptions": {
        "session_name": "string",
        "session_ttl": "string",
        "session_recording": "string",
        "proxy_country": "string",
        "proxy_url": "string",
        "fingerprint": "string"
    }
}

Request samples

Shell
JavaScript
Java
Swift
Go
PHP
Python
HTTP
C
C#
Objective-C
Ruby
OCaml
Dart
R
Request Request Example
Shell
JavaScript
Java
Swift
curl --location --request POST 'https://api.scrapeless.com/api/v1/crawler/crawl' \
--header 'Content-Type: application/json' \
--header 'x-api-token;' \
--data-raw '{
    "url": "http://example.com",
    "limit": 10000,
    "excludePaths": [
        "string"
    ],
    "includePaths": [
        "string"
    ],
    "maxDepth": 10,
    "maxDiscoveryDepth": 0,
    "ignoreSitemap": false,
    "ignoreQueryParameters": false,
    "deduplicateSimilarURLs": true,
    "regexOnFullURL": true,
    "allowBackwardLinks": false,
    "allowExternalLinks": false,
    "delay": 0,
    "scrapeOptions": {
        "formats": [
            "markdown"
        ],
        "onlyMainContent": true,
        "includeTags": [
            "string"
        ],
        "excludeTags": [
            "string"
        ],
        "headers": {},
        "waitFor": 0,
        "timeout": 30000
    },
    "browserOptions": {
        "session_name": "string",
        "session_ttl": "string",
        "session_recording": "string",
        "proxy_country": "string",
        "proxy_url": "string",
        "fingerprint": "string"
    }
}'

Responses

🟢200OK
application/json
Successful response
Body
success
boolean 
optional
id
string 
optional
Example
{
    "success": true,
    "id": "string"
}
🟠402402
🟠429429
🔴500Server Error
Previous
Get the errors of a batch scrape job
Next
Cancel a crawl job
Built with