The Reddit Myth: Why Aggregate AI Citation Studies Are Misleading Your Strategy
Studies showing Reddit and Wikipedia dominate AI citations are technically accurate and strategically useless. Here is why the aggregate number actively harms how brands allocate effort.
- 1Aggregate AI citation studies pool responses across millions of query types. The “average” source profile is meaningless for any specific industry or category.
- 2Wikipedia dominates AI citations for factual and definitional queries. If your buyers are not asking factual trivia, Wikipedia dominance in aggregate data is irrelevant to you.
- 3Reddit citation rates are heavily driven by consumer product and lifestyle queries. Many B2B categories have near-zero Reddit citation signal regardless of community activity.
- 4Citation sources differ dramatically by industry vertical, query stage, AI model, and geographic market. A single aggregate number tells you none of this.
- 5Brands misallocate months of effort chasing aggregate-driven channels that carry no weight in their actual category. The cost is real and rarely diagnosed correctly.
- 6The right question is not “what does AI cite most overall?” but “what does AI cite when my specific buyers ask their specific questions?”
The slide appears in every GEO presentation now. A bar chart. Reddit near the top. Wikipedia usually first. YouTube somewhere in the middle. The headline reads something like: "These are the sources AI cites most. Your strategy should reflect this."
Marketers take notes. Teams realign. Someone gets tasked with a Reddit strategy. Someone else starts drafting Wikipedia edits. A content manager asks whether the company should be doing more video because the chart said YouTube.
None of this is the researchers' fault. The studies are, in their own terms, accurate. The problem is the step that happens next: a finding about AI citation behaviour across all of the internet gets applied to a single company operating in a single category, serving a specific kind of buyer. That step is where the logic breaks, quietly and expensively.
Updated March 2026
What These Studies Actually Measured
Aggregate citation studies measure AI behaviour across all query types simultaneously — cooking, history, product recommendations, science, and everything else — then rank sources by total frequency. The result is an accurate portrait of the whole internet's citation landscape and a misleading guide to any single category within it.
These studies typically work by running thousands of queries across ChatGPT, Perplexity, Claude, or some combination, recording which URLs or domains appear in responses, and ranking sources by total citation frequency. The query sets are necessarily broad. To get a statistically meaningful picture of AI citation behaviour, researchers need volume. So they pull from everything: cooking questions, history queries, celebrity lookups, product recommendations, medical questions, geography trivia, sports statistics, software comparisons, science explainers. The resulting dataset reflects AI behaviour across the full surface area of human inquiry.
Of course Wikipedia comes first. When someone asks an AI what year the Berlin Wall fell, or what photosynthesis is, or who wrote Hamlet, Wikipedia is the single most reliable structured reference in existence. For factual, definitional, and encyclopaedic queries, Wikipedia is an extraordinarily strong source. And there are enormous numbers of those queries.
Of course Reddit appears prominently. When someone asks an AI which running shoes are most comfortable for wide feet, or what the best budget mirrorless camera is for a beginner, or how to deal with a difficult landlord in Ontario, Reddit threads are the richest repository of peer experience on the open web. Consumer lifestyle queries generate massive Reddit citation rates. And consumer lifestyle queries are a very large slice of all AI queries.
“All queries ever run on AI” is not the population you belong to.
The studies are not wrong. They are describing a real population. The problem is what happens when you apply it to yours.
The Base Rate Illusion in Action
Applying an all-query aggregate to a specific B2B category is a textbook base rate fallacy. The overall population (everyone asking AI everything) behaves very differently from the sub-population that matters to you (professional buyers researching specialist software). One number cannot describe both, and the aggregate is almost never the one that applies.
Consider a B2B HR software company. Their buyers are HR directors and People Operations leaders asking things like: "What HRIS platforms integrate natively with Workday?" or "Best performance management tools for a hybrid workforce of 200 people." They are not asking Wikipedia-style factual questions. They are not asking Reddit-style consumer lifestyle questions. They are asking expert vendor evaluation questions in a specialised professional domain.
Worked Example: HR Software
For HR director queries, AI models draw from a completely different source hierarchy:
Reddit has a small HR community. But in citations that appear when AI answers questions about enterprise HR software specifically, its presence is a rounding error. The aggregate study says Reddit is major. For this company's buyers, Reddit is almost irrelevant.
The company reads the study, builds a Reddit strategy, and wonders why nothing moves.
How Citation Profiles Differ by Vertical
Citation source hierarchies are not just somewhat different across industries — they are structurally different. The sources AI cites for a healthcare query share almost nothing with those it cites for a fintech query, which share almost nothing with those for a consumer electronics query. Each vertical has its own source ecosystem, and aggregate data flattens all of them into a single misleading bar chart.
| Industry Vertical | Primary AI Citation Sources | Do Aggregate Stars Apply? |
|---|---|---|
| B2B SaaS | G2, Capterra, SourceForge, industry comparison articles, vendor documentation | Low. Reddit has pockets; Wikipedia irrelevant for niche tools |
| Healthcare / Clinical | PubMed, NHS, CDC, Mayo Clinic, peer-reviewed journals, professional associations | Very low. AI deprioritises Reddit for clinical queries |
| B2B Fintech | Regulatory bodies, trade press (Finextra, Payments Journal), Stack Overflow, LinkedIn | Low to moderate. Developer-adjacent fintech has Stack Overflow signal |
| Consumer Electronics | Reddit (r/photography, r/buildapc), YouTube reviews, RTINGS, Wirecutter, Amazon | High. This is where aggregate findings actually apply |
| Professional Services | Clutch, GoodFirms, industry association publications, LinkedIn profiles, case studies | Low. Professional buyers do not research agencies on Reddit |
| Legal / Compliance SaaS | Bar association publications, Law Technology Today, Capterra, regulatory guidance | Very low. Legal research has its own source hierarchy entirely |
When Aggregates DO Apply
The consumer electronics row is instructive. For that vertical, aggregate study findings broadly apply. Reddit, YouTube, and review aggregators are legitimately the primary citation sources when AI recommends a laptop or a camera. A consumer electronics brand reading the aggregate study and investing in Reddit and YouTube is making a reasonable inference.
But a legal compliance SaaS platform making the same inference from the same study is operating on completely wrong assumptions. Its buyers are not on Reddit. Its citation sources are regulatory bodies, bar association guidance, and trade press. The aggregate study is describing a world that does not overlap with the world its buyers inhabit.
Query Stage Changes Everything, Too
Even within a single category, the dominant citation sources shift dramatically by buyer journey stage. The sources that drive top-of-funnel problem recognition responses are often completely different from the sources that drive vendor evaluation responses. Aggregate data collapses this distinction entirely, leaving brands optimising for the wrong stage as often as the wrong channel.
Take B2B project management software. A buyer at the start of their journey might ask: "What problems do remote engineering teams run into with project coordination?" That is a problem recognition query. AI responses to this kind of question draw heavily from general productivity and workplace content, including some Reddit threads from engineering communities like r/programming or r/devops.
"What problems do remote teams have with project coordination?"
Sources: Productivity blogs, r/devops threads, general workplace content
Reddit signal: Present
"Best project management tools for engineering teams"
Sources: G2, Capterra, listicles, comparison articles
Reddit signal: Declining
"Jira vs Linear for a 50-person engineering team"
Sources: G2, SourceForge, Capterra, structured comparison articles
Reddit signal: Near zero
A strategy built on the aggregate finding that "Reddit is significant" might produce decent top-of-funnel coverage and almost no impact on vendor evaluation queries, which is where purchasing decisions actually form. The channels that matter at each stage are not the same channels.
Add Model Variation and Region, and the Aggregates Collapse Further
Aggregate studies typically blend results across AI models with fundamentally different architectures and across geographic markets with different source authority patterns. Averaging Perplexity (real-time web search) with ChatGPT (training-data heavy) with Gemini (blended) produces a citation profile that accurately describes none of them — and may not describe the model your buyers actually use.
How Different AI Models Cite Sources
Real-time search engine + LLM
Retrieves live web results and cites them directly. Source profile looks closer to Bing than a training model.
For your strategy: Freshness and current indexing dominate
Training-data driven (with optional browsing)
Without browsing, draws primarily on training data with a different vintage and source composition.
For your strategy: Training corpus recency and coverage matters most
Blended training + search
Mixes training knowledge with real-time search in a third distinct pattern. Neither fully live nor fully static.
For your strategy: Both training signals and live indexing contribute
A study that averages across all three is averaging across fundamentally different information retrieval systems. Geography adds another layer. The same query asked in the Australian market versus the US market can produce different citation profiles, because regional sources have different authority in different markets. An Australian HR software company optimising based on a US-centric aggregate study is building a strategy calibrated to a different buyer population entirely.
What the Misallocation Actually Costs
Aggregate-driven strategy does not produce zero results — it produces results in the wrong places while the right places go unworked. For teams with limited bandwidth, this is the difference between gaining AI visibility traction in 60 days and spending 18 months wondering why nothing has moved.
A B2B SaaS company that builds a Reddit engagement strategy because the aggregate study said so will find some relevant communities. They will post useful answers. Some of those posts will be cited by AI occasionally. The effort is not wasted to zero. But the same hours spent completing their G2 profile, getting 40 more detailed reviews, and building out their SourceForge listing would have driven three to five times the AI citation improvement. They chose the flashier channel because a study made it look important. The study was measuring the wrong population.
The Bandwidth Problem
Most B2B marketing teams working on AI visibility have one, maybe two people on it alongside other responsibilities. Sending that bandwidth toward channels with low signal in their specific category is not a minor inefficiency.
60 days
to traction with category-specific research
18 months
wondering why nothing moved on aggregate advice
The Three Questions Aggregate Studies Cannot Answer
Which sources does AI cite for my specific category?
Aggregate data tells you what is cited across all of the internet. Your category has its own source hierarchy that may look nothing like the aggregate.
Which sources does AI cite at the buyer journey stage that matters most to me?
Problem recognition, solution research, and vendor evaluation queries each have different source profiles, even within the same category.
Which specific gaps in my current source presence are costing me citations?
The most actionable finding is always specific: a missing G2 listing, a thin Capterra profile, an absent SourceForge page. Aggregate data never gets you there.
What to Do Instead
Replace the aggregate study with a category-specific query simulation. Run 30 to 50 queries that mirror your actual buyers' research process, trace the citations in each response, and build a source map of what is driving recommendations in your specific competitive landscape. That map — not a global bar chart — is what your strategy should be built on.
What you will typically find, after doing this properly, is that your category has three to six source types that drive the majority of AI citations. Some will be obvious once you see them. Others will surprise you. The surprises are the valuable findings, because they are the places your competitors have already figured out what you have not yet noticed.
A few principles that hold across most B2B categories, regardless of vertical:
Review platforms first
G2, Capterra, SourceForge, and TrustRadius provide exactly the kind of structured, aggregated, third-party signal AI models are designed to trust. If your profile on any of these is thin, fixing it is almost always the highest-leverage action regardless of what else your category research shows.
Your own content at vendor evaluation stage
AI models do draw from branded content, especially for feature-specific and comparison queries. If your documentation, comparison pages, and FAQ content directly answers the questions buyers ask at the bottom of the funnel, you have a legitimate citation advantage over competitors who have left that space empty.
Structured data is underutilised in almost every B2B category
Schema markup (JSON-LD) gives AI a machine-readable description of what your product is, what category it belongs to, and what problems it solves. Most B2B SaaS companies either have none or have minimal generic schema. A well-implemented SoftwareApplication or Product schema with rich attributes consistently outperforms competitors with no schema for precise vendor-specific queries.
Find the community where your buyers actually are
For developer tools: Stack Overflow and GitHub discussions. For marketing technology: LinkedIn thought leadership and specialised communities. For HR software: SHRM forums and People Operations Slack groups. For none of these categories is it "Reddit broadly." Find the community where your buyers actually talk, rather than the community that shows up in aggregate citation data.
Recognising the Pattern Before It Costs You
Any recommendation about AI visibility that cites a cross-category aggregate should trigger one immediate question: does the underlying query set resemble my buyers' queries? If the study included millions of general knowledge and consumer lifestyle queries, the answer for most B2B companies is almost certainly no — and the finding does not apply to you.
The reason aggregate citation studies spread so successfully is that they are genuinely interesting and easy to share. A chart showing which sources AI cites most is legible to anyone. It requires no domain knowledge to interpret. It feels actionable immediately. These are qualities that make content shareable, which is not the same as making it useful for decision-making.
The study that correctly says “Reddit is significant in AI citations” is technically right and practically wrong for most B2B companies reading it.
Good AI visibility strategy starts with the humility to recognise that your category is not the internet. Run your buyers' queries. Trace the citations. Build your strategy on that data instead of on a bar chart that was never measuring your world in the first place. The companies winning in AI visibility right now are not the ones who read the most studies. They are the ones who looked most carefully at what AI actually does when their specific buyers ask their specific questions, and then fixed the gaps that mattered.
For a grounding in why AI citation signals differ structurally from SEO signals, see SEO vs GEO: Why Strong Search Rankings Don't Guarantee AI Visibility. And if your starting question is why a specific competitor is appearing in AI responses where you are not, the diagnostic framework in Why Are My Competitors Showing Up in ChatGPT and I'm Not? is more relevant than any aggregate study.
Frequently Asked Questions
Are aggregate AI citation studies useless?
No. Aggregate studies are genuinely useful for understanding how AI systems work at a structural level: what types of sources AI tends to trust, how different model architectures affect citation behaviour, and how citation patterns have changed over time. That is valuable context for anyone doing serious work in this space.
The problem is not the studies. It is applying aggregate findings directly to category-level strategy without verifying whether those findings hold in your specific vertical, for your specific buyer queries, on the specific AI platforms your buyers use. That verification step is where most brands skip to their detriment.
Does Reddit ever matter for B2B categories?
Yes, but narrowly. Developer-adjacent B2B tools have meaningful Reddit signal because r/devops, r/programming, r/MachineLearning, and similar communities are actively used by the same buyers. Security software, developer tools, and infrastructure products all have real Reddit citation presence in AI responses to the right queries.
For most B2B software categories above the developer layer, including HR tech, legal tech, finance software, marketing platforms, and enterprise workflows, Reddit signal in AI citations is weak. The buyers are not there, so the discussions are not there, so the citations are not there. Aggregate data that pools developer tool queries with enterprise HR software queries will show Reddit as significant, obscuring the fact that it is significant for one and negligible for the other.
How do I find my actual category citation sources?
The core method is systematic query simulation. Write down 30 to 50 queries that represent how your buyers research your category, covering problem recognition, solution research, and vendor evaluation. Run each query in ChatGPT, Claude, Gemini, and Perplexity. Record not just whether your brand appears, but which sources each model cites in responses where your brand does and does not appear.
After 30 to 50 queries across four models, you will have 120 to 200 data points. Aggregate the sources cited and a clear pattern will emerge: three to six source types will account for the majority of citations in your category. That is your actual citation landscape. That is what your strategy should be built around.
What if my category is genuinely well-represented in aggregate studies?
Consumer-facing products, especially in electronics, lifestyle, and entertainment, are often well-represented in aggregate studies because consumer queries make up a large share of the total. If you are a consumer brand in one of these categories, aggregate findings may actually be a reasonable starting point.
Even then, verify at the category level before committing. A consumer brand selling professional photography equipment is in a consumer category but serving a specialised buyer who has different source preferences than the general consumer. The aggregate study even when the aggregate findings seem to point in roughly the right direction.
If you want to see where your brand actually stands in your specific category, run a free AI visibility report. It simulates 25 buying scenarios through ChatGPT and surfaces your actual citation gaps in about 10 minutes, grounded in your specific category rather than an aggregate average.