The AI Citation Arms Race: Why Your Optimization Tactics Have an Expiration Date
There's a coding bootcamp somewhere that lost 80% of its revenue because someone got mad at them on Reddit. Not mad enough to file a lawsuit or organize a boycott or even write a particularly viral Medium post. Just mad enough to coordinate a small group of people to say bad things about them in the right corners of the internet at the right frequency.
The bootcamp didn't know this was happening. They kept teaching the same curriculum, kept getting the same student reviews, kept showing up in the same Google results. But ChatGPT and Perplexity started citing them differently. Started hedging. Started including phrases like "some users report" and "mixed feedback suggests." The calls stopped coming. The Slack channel where potential students asked questions went quiet.
A 2025 study on AI search manipulation found what you'd expect if you thought about it for thirty seconds: large language models are easily gamed by tactics that would have gotten you algorithmically slaughtered by Google in 2010. When these systems pull information from the web to answer questions, repeated mentions can outweigh source quality. Publication dates can be manipulated. Coordinated campaigns get treated as genuine consensus because the models haven't developed the filtering mechanisms to distinguish authentic patterns from manufactured ones.
Which means we're in a very specific moment. A loophole. The kind where early movers make absurd amounts of money doing things that will seem obviously stupid in retrospect.
At first the models only mirrored the web. Catalogued what existed, ranked what was linked. Now they're starting to eavesdrop on it.
The thing about domain authority is that it doesn't really exist anymore
Or it exists in the way that phrenology exists. As a framework that made sense given what people knew at the time, but that we've now quietly replaced with something else without really announcing it.
The data here is weirdly precise. Someone at an agency tried to score it. How much does each platform care about your domain's history versus whether your content actually explains things well? ChatGPT barely cares about domain authority anymore. Google's AI Overviews still do, which tells you something about how slowly institutional systems adapt.
This is the kind of shift that sounds incremental until you realize what it means: a three-person company with a really good explanation of how SSL certificates work can beat Cloudflare's entire marketing team if Cloudflare's explanation is middling. The byline doesn't matter. The backlinks don't matter. What matters is whether you're the best teacher.
Except that's not quite true either. Someone tracked citations in AI Overviews for personal care products and found that Amazon and Walmart alone account for a staggering portion of everything getting referenced. They're not the best teachers. They're just the biggest names.
So you have this strange dual reality where authority is both completely democratized and more concentrated than ever. Small sites are getting cited alongside The New York Times, which suggests a kind of meritocracy. But major brands are also getting cited reflexively, which suggests the opposite.
The mechanism here is worth understanding. Large language models don't develop opinions about authority. They learn patterns from their training data. If a source gets cited frequently in that training set, the model picks up on it. It's not deciding that The New York Times is authoritative; it's observing that other sources cite The New York Times all the time, so when similar contexts appear, the same names come up again. This creates a feedback loop: the already-cited get cited more because citation itself becomes the signal.
But here's where it gets interesting: that same mechanism also creates the long tail. If you're the only clear signal on a narrow enough topic, you become the match for that context. The democratization isn't pure meritocracy. It's a side effect of how these systems compress attention. The models aren't being fair; they're being precise. And precision, it turns out, is something you can actually compete on.
Maybe this isn't new. Maybe we're just watching reputation reinvent itself with better math. The mechanisms change. Links to citations to token frequencies. But the underlying question stays the same: who gets believed, and why? Every few years the internet finds a new way to ask it.
Everyone is running the same playbook and it still works
The playbook is: make your content as easy as possible for a machine to parse. Use listicles. Use schema markup. Use FAQ sections. Use clear hierarchies. Basically, write like you're talking to someone who's reading very quickly and won't remember what you said three paragraphs ago.
The data backs this up. Listicles are genuinely easier for AI systems to extract and cite. Structured data makes your content machine-readable in a way that prose doesn't. FAQ schemas increase the odds that an AI will quote you when someone asks that question.
The problem is that everyone knows this now. Which means everyone's cranking the same knobs, chasing volume over tone. The bar keeps rising not because content is getting better but because it's getting louder. And at some point (probably soon) the sheer noise of over-optimized content is going to force these systems to develop better filters.

Google already started. In August, they rolled out an update targeting sites that use structured data to lie about things like authorship or review ratings. If you're using schema markup to tell Google that your blog post is actually a peer-reviewed study, or that your three-star product secretly has five stars, the algorithm now knows to ignore you.
This is how it always goes. A tactic works until enough people do it that the platform has to filter it out, and then the tactic becomes a liability. Exact-match domains used to be worth something in SEO. Then everyone bought them. Then Google downgraded them. Then having one became a signal that you were probably spammy.
We're watching this happen in real-time with AI citations, except faster, because the people building these systems saw what happened with Google and they're not interested in repeating that decade-long arms race.
The models are learning to check your references
There's an academic paper from April that analyzed references generated by large language models and found something called the Matthew Effect: the models systematically favor papers that are already highly cited, which means they're amplifying existing citation inequalities rather than discovering new sources.
This sounds bad if you're a researcher hoping your obscure-but-brilliant paper gets noticed. It sounds worse if you're trying to get your blog cited and you're competing against Forbes.
But here's the interesting part: a different study looked at ChatGPT citations and found that while the top domains accounted for nearly half, the rest was split among tens of thousands of smaller sites. Which means the long tail is real. Which means you can win on expertise even if you don't have brand recognition.
The catch is that you have to actually be the best source. You have to be the Wikipedia of your niche. You have to have the most complete, most clearly explained, most thoroughly researched answer to whatever question someone is asking. And you have to do this while also making your content machine-readable, because the models aren't going to do you any favors.
What you can't do, what increasingly won't work, is fake it. You can't just format mediocre content as a listicle and expect that to be enough. The models are getting better at distinguishing between "this is structured because structure aids comprehension" and "this is structured because someone told me AI likes structure."
Authority is becoming something you earn in public
But something subtler is happening beneath all these tactics. Something that matters more than whether you use listicles or schema markup or any of the surface-level adjustments everyone's focused on.
This is maybe the strangest shift. For years, the way you built authority was by getting other websites to link to you. Backlinks were the currency of the internet. You'd do guest posts and press releases and all these baroque link-building strategies because that's how you signaled to Google that you were legitimate.
That's not gone, but it's being supplemented by something messier and harder to game: what people say about you in places where real humans congregate. Reddit threads. TikTok comments. LinkedIn discussions. The places where someone might say "oh yeah, I used that company's tool and it actually worked" or "I read their article and it was helpful."
These mentions matter now. They matter because the models are learning to treat social proof as a trust signal. When someone asks ChatGPT about the best project management software and seventeen Reddit threads mention the same tool, that's data. When a creator with 50,000 followers makes a video explaining a concept and credits your article, that's a signal that your article is probably good.
You can't manufacture this at scale, which is exactly why it's valuable. You can buy backlinks. You can't really buy authentic community respect. I mean, you can try. That coding bootcamp's competitors apparently did. But it's expensive and risky and obvious when it goes wrong.
The strategic implication is that content strategy now extends beyond your own domain. You need to be good at your thing, obviously. You need to structure your content well. But you also need to exist in public in a way that makes people want to talk about you. You need to collaborate with other creators. You need to show up in communities where your expertise is relevant. You need to do the thing that's always worked but that we briefly convinced ourselves we could optimize our way out of: you need to be genuinely helpful to actual humans.
The window is closing but it hasn't closed
Right now, today, formatting everything as a listicle still works. Schema markup still works. Being strategic about Reddit mentions still works. These tactics have maybe eighteen months left before they stop working or start backfiring.
The question is whether to build a strategy that depends on them. The models are getting smarter. The platforms are developing better spam filters. The tactics that work now are becoming signals rather than strategies precisely because everyone is using them.
The alternative is slower but more durable: become genuinely authoritative in a small enough space that you can dominate it. Publish original research that other people cite. Explain things more clearly than anyone else. Show up in communities and be helpful without asking for anything in return. Build relationships with other creators whose audiences overlap with yours.
This is the same advice that worked in 2007 when blogrolls mattered, or in the Usenet days when reputation was built through consistent, helpful participation in newsgroups. Every platform shift produces a temporary window where format beats substance, and every time, the platform eventually closes that window by learning to recognize the difference. We're watching it happen again, just faster, because the people building these systems already know how this story ends.
The coding bootcamp didn't lose 80% of its revenue because its content was bad. It lost revenue because the public conversation about it turned negative and the AI believed the crowd. In 2025, your authority is only as strong as your reputation in the spaces where people talk about things like you.
The fundamental shift here is from indexing to listening. For decades, search engines catalogued what existed. They ranked pages based on links and keywords and metadata. The model was archival: here's everything, sorted by signals we can measure. But these AI systems are learning to do something different. They're asking what communities trust, what creators vouch for, what gets mentioned when real humans talk to each other. They're not just ranking sources anymore; they're learning to distinguish between published authority and earned credibility.
They're learning to gossip.
Member discussion