/trigger endpoint. Use this when you have more than 20 URLs, need discovery by keyword or subreddit, or want delivery to a webhook or S3.
Prerequisites
- A Bright Data account with an active API token
- Familiarity with the synchronous request flow
Step 1: Trigger the collection
Send aPOST request to the /trigger endpoint with your input URLs. This example collects five Reddit posts in a single batch:
200 response with a snapshot_id:
Discovery with async
The async endpoint is the best fit for discovery jobs, because Reddit discovery can return many results. Trigger a subreddit or keyword discovery by adding the relevant query parameters: Discover by subreddit URL:Step 2: Monitor progress
Poll the snapshot status until it showsready. This takes 30 seconds to several minutes depending on the number of URLs and whether discovery is involved.
| Status | Meaning |
|---|---|
collecting | Scraping is in progress |
digesting | Data is being processed |
ready | Results are available for download |
failed | The collection encountered an error |
Step 3: Download results
Once the status isready, download the scraped data:
Skip polling with webhooks
If you don’t want to poll for status, add awebhook parameter to receive results automatically:
Limits and constraints
| Constraint | Value |
|---|---|
| Max input file size | 1 GB |
| Max concurrent batch requests | 100 |
| Max concurrent single-input requests | 1,500 |
| Webhook delivery size | Up to 1 GB |
| API download size | Up to 5 GB |
Troubleshooting
Getting a 429 Too Many Requests error?
Getting a 429 Too Many Requests error?
You’ve exceeded the concurrent request limit. Reduce the number of parallel requests or combine inputs into fewer, larger batches. Each batch can include up to 1 GB of input data.
Snapshot status shows 'failed'?
Snapshot status shows 'failed'?
Check that all input URLs are valid, publicly accessible Reddit URLs. Review the error details in the snapshot response or in the Logs tab of your Bright Data dashboard.
Results are incomplete or missing some URLs?
Results are incomplete or missing some URLs?
Some URLs may fail individually while the overall job succeeds. Private subreddits, deleted posts and removed comments cannot be scraped. Check the snapshot response for any
errors field and retry failed URLs in a separate request.Discovery returned zero results?
Discovery returned zero results?
For keyword discovery, make sure the
date value matches one of Reddit’s accepted ranges (e.g. Past hour, Past day, Past week, Past month, Past year, All time). For subreddit discovery, confirm the subreddit URL is valid and not private.Next steps
Delivery options
Webhooks, S3, Snowflake, Azure and GCS delivery.
API reference
Full endpoint specs, parameters and response schemas.