Download the Make.com Reddit Scrapeer
Turn Reddit into your personal idea factory. This plug-and-play Make.com scenario pulls the best posts from any subreddit, cleans the mess, lets AI pre-qualify what’s worth writing about, and hands you tidy data for Sheets, Notion, WordPress, Discord – wherever you publish. It’s fast, fun, and totally Reddit API-key-free.
Why scrape reddit?
If you have ever wished you could turn Reddit into a steady stream of blog ideas, newsletter links, or WordPress drafts, this guide is for you. Below I walk through a simple end-to-end workflow that shows exactly how to Scrape Reddit Make.com using nothing more than Reddit’s public RSS, a bit of text cleaning, and one smart AI check. You can import the ready-made scenario in minutes and start shipping content the same day.
Why this Make.com Reddit scraper works so well
This setup deliberately avoids the Reddit API, so there are no keys to apply for and no quotas to worry about. Everything runs from the public RSS feed, which means fewer moving parts and far less maintenance. Inside Make.com, the scenario exposes four simple controls at the very start: the subreddit you want to watch, the sort you prefer (hot, new, or top), the time window you care about (day, week, month, year, or all), and the limit on how many items to fetch. Those inputs shape a clean RSS request, which is then parsed and stripped of HTML so you are always working with readable text.
The clever bit is the built-in AI step. Rather than dumping every post into your system, the scenario asks a model to decide whether an item is actually worth your time. It replies with a strict yes or no and, when it says yes, it also proposes a clearer, search-friendly blog title. The final output is a tidy array that you can push into Sheets, Notion, WordPress, Discord, or anywhere else you run your publishing workflow.
What’s inside the blueprint
When you open the scenario you will see a short, readable chain of modules. It starts with a Set Variables module that acts as your control panel for subreddit, sort, time window, and limit. The RSS module builds the feed URL from those variables and fetches the latest items. A small text processing step removes the HTML clutter from each description so the AI is judging the content, not the formatting. An Aggregator then batches the items so the AI can make a single, consistent decision pass. The OpenAI module reviews the batch and returns a compact JSON object telling us which posts are worth writing about and what their improved titles should be. A simple filter keeps only the winners, and a final mapping step preserves the original title, author, URL, and cleaned description so your downstream tool receives exactly what it needs.
Quick start to Scrape Reddit in Make.com
Import the blueprint into your Make.com account and open the first Set Variables module. Change the subreddit, the sort, the time window, and the limit to match your niche. If you want the AI triage and title suggestion, connect your OpenAI account to the relevant module. Run the scenario once to confirm everything looks right. From there you can plug in your favourite output: append rows in Google Sheets, create Notion pages for your content calendar, draft WordPress posts, trigger a Webhook for custom apps, or post into Discord and Telegram channels.
The exact RSS pattern this scenario uses
https://www.reddit.com/r/{{subreddit}}/{{sort}}/.rss?t={{time_window}} Try it with investing, technology, tennis, ukinvesting, personalfinance, entrepreneur, startups, or artificial to get a feel for different communities. Because you control sort and time window, you can surface either the latest chatter or the all-time classics with a single tweak.
The AI prompt you can paste into the OpenAI module
Use a GPT-4-class model with a low temperature so it always returns valid JSON. The scenario expects this shape:
{ "instruction": "You will receive an array of Reddit posts. For each, decide if it's worth writing a UK-focused blog post about (topical, useful, or evergreen). Return STRICT JSON with a single object containing a result array. Each element must be: { \"title\": <original_title>, \"write\": \"yes\"|\"no\", \"suggested_blog_post_title\": <if write=yes, a clearer, SEO-friendly UK English title; else empty string> }. No extra keys, no prose.", "posts": {{array_of_posts}} } From scraper to publishing pipeline
Because the whole flow is mapped to a clean array, you can grow from “Scrape Reddit Make.com” to a full publishing pipeline without rewriting anything. Many people start with a Google Sheets idea bank, appending each approved post with the date, subreddit, URL, suggested title, and a short summary. Others build a Notion content calendar and create pages with publish dates and statuses so research, drafting, editing, and publishing all live in one place. If you blog on WordPress, add a module that creates draft posts using the AI-suggested title. Place the Reddit URL and a cleaned summary in the post content and tag the draft with the subreddit name and your niche so it is easy to find later.
It is also simple to send yourself a daily email of the top five picks, pipe approved links into a private Discord or Telegram group, or spin up an outline generator that creates H2 and H3 headings for every yes result. If you want more control, extend the AI step to score each idea from 1 to 10 for relevance or commercial intent and then pass only the sevens and above. You can keep your topics tight with whitelists and blacklists, or add a de-duplication check by hashing each URL and storing it in a Make Data Store or a dedicated sheet so the scenario never serves the same link twice.
Scheduling that actually fits your week
For discovery mode, schedule the scenario every hour from 08:00 to 18:00 on weekdays so you are never short of prompts. On Saturdays, switch the time window to week and run it once to catch the big hits you may have missed. For evergreen hunting, set it to month, year, or all with sort set to top. That single change turns the same scenario into a long-term research tool.
WordPress mapping that feels native
When you connect WordPress, set the post status to draft and map the AI-suggested title into the post title field. Drop the Reddit URL and the cleaned description into the post content so you always keep attribution and context. Use tags for the subreddit and your topic area to keep your draft queue organised. This is the easiest way to go from Scrape Reddit Make.com to publish-ready drafts without touching your CMS by hand.
Troubleshooting without the guesswork
If nothing passes your filter, check that the AI is returning the exact yes or no string you expect and keep the temperature low to avoid JSON drift. If you see odd characters in descriptions, confirm the HTML stripping step happens before aggregation. If duplicates slip through, add a simple URL hash or store each GUID in a Data Store and skip items already seen. If your queue is too noisy, lower the fetch limit or tighten your filters. You can always scale back up once you are happy with the signal.
Good internet manners
Reddit is a brilliant source of signals, but it deserves respect. Always credit the original post, avoid copying long bodies verbatim, and focus on adding commentary, context, and value. If you do not want to pass SEO value, use nofollow on outbound links. This keeps communities healthy and your content useful.