Prerequisites
- A scraper saved to production in the Bright Data Scraper Studio IDE
- An API key for the API-trigger and API-delivery paths (create one)
How do I save a scraper to production?
While you edit code in the Bright Data Scraper Studio IDE, the system auto-saves your work as a development draft. To make the scraper runnable outside the IDE, click Save to Production in the top-right corner of the IDE. All production scrapers appear under My Scrapers in the control panel. Inactive scrapers are shown faded.
How do I trigger a scraper run?
Bright Data Scraper Studio supports three ways to initiate a collection.- Initiate by API
- Initiate manually
- Schedule a scraper
Start a collection through the REST API without opening the control panel. See Getting started with the API for authentication, request format, and response schema.Before you send a request, create an API key. Go to Dashboard > Account settings > API key.
The API-trigger panel gives you three things:

- Inputs: provide input values manually or through the API request body
- Trigger behavior: queue multiple requests to run in parallel or sequentially; queued jobs run in the order they are submitted
- Preview of the API request: Bright Data shows you a ready-to-run
curlcommand. Select the Linux Bash viewer forcurl. The response includes ajob_idyou use to fetch the data later.
When delivery is set to API download, you must call the “Receive data” API endpoint to retrieve results. Webhook and cloud-storage destinations push data automatically.
What are the rate limits and concurrency limits?
Bright Data Scraper Studio enforces concurrency limits per scraper, based on whether the request is batch or real-time.| Collection type | Concurrency limit |
|---|---|
| Batch | Up to 1,000 concurrent requests per scraper |
| Real-time | No limit |
Maximum limit of 1000 jobs per scraper has been exceeded. Please reduce the number of parallel jobs.
Batch vs real-time collection
Bright Data Scraper Studio offers two collection methods, each optimized for a different use case.| Batch collection | Real-time collection | |
|---|---|---|
| Input size | Many inputs per job (list of URLs or keywords) | One input per request |
| Response timing | Results returned after the full job completes | Response returned in real time |
| Retention | 16 days | 7 days |
| Concurrency limit | 1,000 concurrent jobs | None |
| Use when | You are building a dataset and can wait | You need an answer inside a live request |
How do I configure delivery?
Open My Scrapers, click a scraper row, and choose Delivery preferences to set where and how Bright Data Scraper Studio delivers results.When should I receive the data?
When should I receive the data?
- Batch: get results once the whole job finishes; efficient for large datasets
- Split batch: deliver partial results in smaller chunks as they become ready
- Real-time: get a fast response to a single request
- Skip retries: do not retry on error (speeds up collection at the cost of completeness)
Which file formats are supported?
Which file formats are supported?
- JSON
- NDJSON
- CSV
- XLSX
- Parquet
Which delivery destinations are supported?
Which delivery destinations are supported?
- API download (pull via REST API)
- Webhook (push via HTTPS POST)
- Cloud storage: Amazon S3, Google Cloud Storage, Azure Blob Storage, Alibaba Cloud OSS
- SFTP / FTP
Media files cannot be delivered via Email or API download. Use cloud storage, SFTP, or webhook when collecting images, videos, or other binary content.
How can I control what goes in the batch output?
How can I control what goes in the batch output?
- Results and errors in separate files
- Results and errors in one combined file
- Only successful results
- Only errors
Which notifications can I enable?
Which notifications can I enable?
- Notify when a collection completes
- Notify on success-rate thresholds
- Notify when an error occurs
How do I configure the output schema?
The output schema defines the structure of your collected data: field names, data types, default values, and any additional metadata you want Bright Data Scraper Studio to attach (timestamps, screenshots, WARC snapshots).
| Control | Description |
|---|---|
| Input / Output schema | Tab switch for the two schema views |
| Custom validation | Define validation rules that run on every collected record |
| Parsed data | The raw fields the scraper’s parser code emits |
| Add new field | Add a new field by name and type |
| Additional data | Optional metadata: timestamp, screenshot, WARC snapshot, and more |
Related
Scraper Studio specifications
Infrastructure limits, billing, and data retention
WARC snapshots
Archive raw HTTP responses alongside collected data


