Skip to main content
Send a Reddit URL or a keyword, get structured JSON back. The Bright Data Reddit Scraper API handles proxies, CAPTCHAs and parsing so you can focus on your data pipeline.
New to Bright Data? Create a free account and get $2 credit to start scraping.

How it works

You send one or more Reddit URLs or keywords to the Bright Data Reddit Scraper API. Bright Data handles the scraping infrastructure and returns clean, structured JSON.
Your app  -->  Bright Data API  -->  Structured JSON
           POST /datasets/v3/scrape
           Authorization: Bearer YOUR_API_TOKEN
All requests use a dataset_id to specify the data type (posts or comments) and return results in JSON, NDJSON or CSV.

Supported data types

Posts

Titles, descriptions, upvotes, community stats and attached media. Collect by post URL or discover by keyword or subreddit URL.

Comments

Comment text, replies, upvotes and post context. Collect by post or comment URL, with optional day-based filtering.

Endpoints

The Reddit Scraper API exposes two dataset families. Each family supports one or more input shapes.
DatasetInputWhat it returns
Posts — Collect by URL (gd_lvz8ah06191smkebj4)url (post URL)A single post with metadata, community info and media
Posts — Discover by keyword (gd_lvz8ah06191smkebj4)keyword, date, num_of_postsPosts matching a search term, filtered by date
Posts — Discover by subreddit URL (gd_lvz8ah06191smkebj4)url (subreddit URL), sort_byPosts from one subreddit, sorted by new, top or hot
Comments — Collect by URL (gd_lvzdpsdlw09j6t702)url (post or comment URL), days_backComments from a post with full thread context

Request methods

The Bright Data Reddit Scraper API supports two request methods. Choose based on your volume and latency needs.
MethodEndpointBest for
Synchronous/scrapeReal-time lookups, up to 20 URLs
Asynchronous/triggerBatch jobs, 20+ URLs, discovery, production pipelines
Learn more in Understanding sync vs. async requests.

Capabilities and limits

CapabilityDetail
Output formatsJSON, NDJSON, CSV
Max URLs per sync request20
Max URLs per async request5,000
Data freshnessReal-time (scraped on demand)
Delivery optionsAPI download, webhook, Amazon S3, Snowflake, Azure, GCS (all options)
PricingPay per successful record (see pricing)

Common questions

Yes. Each request triggers a live scrape against Reddit. There is no cached or stale data. Response times vary by endpoint: Collect-by-URL requests typically return in 10 to 30 seconds, while discovery requests may take longer depending on result volume.
Collect by URL scrapes a specific Reddit page you provide — a single post, a comment thread or a subreddit. Discover finds Reddit posts matching search criteria such as a keyword or a subreddit listing, then scrapes the results. Discovery is most useful via async requests when you don’t know the exact URLs up front.
Yes. The Comments API accepts an optional days_back parameter. Pass an integer to limit results to comments posted within that many days.
When scraping using proxies or Web Unlocker, you still need to write and maintain your own parsing logic and update it whenever Reddit changes its page structure. The Reddit Scraper API handles the entire stack: proxy rotation, anti-bot bypassing and parsing. You simply send a Reddit URL or keyword and get clean, structured JSON back with no scraping infrastructure or parser maintenance required on your end.

Next steps

Quickstart

Scrape your first Reddit post in 5 minutes.

Send your first request

Full code examples for every endpoint in cURL, Python and Node.js.

API reference

Endpoint specs, parameters and response schemas.