Why SEO Just Became More Important Than Ever

October 15, 2025

AI was supposed to kill SEO. Instead, it made search optimization the most critical business function of 2025.

For the past two years, the marketing world has been bracing for SEO’s extinction. ChatGPT would replace Google. AI chatbots would make search engines obsolete. Organic traffic would vanish as users asked questions directly to language models instead of clicking through search results.

That’s not what happened.

Instead, something unexpected emerged: SEO has become more valuable, not less. The companies seeing this shift early are adjusting their content strategies accordingly. The ones ignoring it are watching their digital presence slowly evaporate from both traditional search and AI-powered discovery systems.

The reason comes down to economics and physics. AI models can’t magic information out of thin air. They need sources. And obtaining those sources just got exponentially more expensive and technically complex.

The billion-dollar retraining problem

Training a frontier AI model has become obscenely expensive. Google reportedly spent $192 million training Gemini 1.0 Ultra. OpenAI’s GPT-4 cost an estimated $79 million. Industry analysts expect the largest models to exceed a billion dollars in training costs by 2027.

Those aren’t one-time expenses. Models need updating. New information emerges daily. Without fresh data, AI systems become outdated reference libraries spouting information from their last training cutoff.

But retraining isn’t like updating software. A single retraining run can cost millions of dollars, consume weeks of compute time, and emit hundreds of tons of CO2. For context, the cost of training frontier models has grown 2.4 times annually since 2016.

No company can afford to retrain massive models every time new information appears. OpenAI famously chose not to fix a known mistake in GPT-3 because retraining would have been too expensive. Google’s DeepMind avoided certain architectural experiments for its StarCraft AI because the training costs were prohibitive.

So what do AI companies do instead? They scrape the web. Constantly.

Google just declared war on AI scrapers

In September 2025, Google quietly removed a feature that had existed for years: the ability to view 100 search results on a single page. The change seemed minor. It wasn’t.

The removal targeted a specific URL parameter that SEO tools, researchers, and AI companies had used to efficiently scrape large batches of search results. Instead of making one request for 100 results, scrapers now need to make ten separate requests.

The cost just increased tenfold.

Google’s public statement was carefully neutral: “The use of this URL parameter is not something that we formally support.” But the timing tells a different story. AI platforms like ChatGPT, Perplexity, and others had been aggressively scraping Google’s results to train models and provide real-time answers.

Graph showing impact of Google's num=100 parameter removal
After Google disabled the num=100 parameter in September 2025, search impression data dropped 80-90% for many sites as bot traffic vanished from analytics.

The change had immediate ripple effects. Rank-tracking tools broke. Search Console impression data plummeted as bot traffic disappeared from reporting. SEO researchers estimate the change effectively hides 80-90% of indexed pages from bulk data collection.

More importantly, it signals that Google views AI scrapers as a competitive threat worth fighting. The move forces AI companies to work harder and pay more to access the same information.

AI models still need the open web

Here’s the paradox: AI was supposed to replace search engines, but AI models depend entirely on content that’s optimized for search engines to find.

Language models don’t generate knowledge. They synthesize information from sources. When ChatGPT answers a question about recent events, it’s either searching the web in real-time or pulling from content it previously indexed. When Perplexity provides citations, those citations come from web pages that were discoverable, crawlable, and well-structured.

AI-powered web scraping has become a massive industry. The global web scraping market is projected to grow from current levels to over $1 billion by 2030, with AI integration driving much of that expansion. Modern AI scrapers use machine learning to adapt to website changes, bypass anti-scraping measures, and extract data from JavaScript-heavy sites.

But they’re still fundamentally doing web scraping. They still need to find your content, access it, parse it, and understand it. The same factors that make content discoverable to Google make it discoverable to AI systems.

What AI systems look for

AI models and their scraping systems prefer certain content characteristics:

Structured data. Clean HTML, semantic markup, proper heading hierarchies. Schema.org markup that explicitly defines what content represents. AI parsers work better when content follows predictable patterns.

Authoritative sources. Original research, expert analysis, proper citations. AI systems need to assess reliability. Content from established domains with strong backlink profiles and consistent publishing histories ranks higher in both traditional search and AI training pipelines.

Fresh information. Models can’t rely solely on stale training data. Real-time scraping focuses on recently published or updated content. Sites that publish regularly and update existing content signal ongoing value.

Accessible content. Paywalls, aggressive bot protection, and complex JavaScript can make content invisible to scrapers. Ironically, the same technical factors that hurt traditional SEO also limit AI discoverability.

You’re now optimizing for multiple discovery channels

The competitive landscape has shifted. Your content used to compete primarily in Google search results. Now it competes across multiple discovery channels simultaneously:

Traditional search engines still drive 90%+ of web traffic for most businesses. Google processes over 8 billion searches daily. Bing, DuckDuckGo, and other engines collectively handle billions more. This hasn’t changed.

AI-powered search is growing rapidly. Google’s Gemini AI chatbot received over 1 billion visits in September 2025, up 46% from the previous month. Perplexity, ChatGPT’s search feature, and other AI search tools are seeing similar growth.

Direct AI citations represent a new traffic source. When AI systems cite sources in their responses, they’re creating new referral traffic. Some marketers report that citations in AI-generated answers now drive measurable traffic, particularly for technical, educational, and authoritative content.

Training data pipelines determine long-term visibility. Content that makes it into model training datasets gains persistent visibility. Every time someone asks a related question, your expertise influences the response even without explicit citation.

The businesses winning in this environment aren’t choosing between traditional SEO and AI optimization. They’re building content strategies that work across all discovery channels simultaneously.

The new metrics that actually matter

Traditional SEO metrics still apply, but they’re no longer sufficient. Forward-thinking marketing teams are tracking additional signals:

AI Overview appearances. How often does your content appear in Google’s AI-generated summaries? These featured positions drive significant visibility even when users don’t click through.

Citation frequency. Are AI systems citing your content when answering questions in your domain? Some teams use custom scripts to query ChatGPT, Perplexity, and other tools with relevant questions, then log which sources get cited.

Structured data coverage. What percentage of your content includes proper schema markup? AI parsers rely heavily on structured data to understand context and relationships.

Content freshness signals. How frequently are you publishing and updating content? Recency matters more in an environment where AI systems need current information but can’t afford constant retraining.

Source authority metrics. Traditional measures like domain authority, backlink quality, and expert authorship have taken on new importance. AI systems use these same signals to assess source reliability.

The visibility gap just got wider

Google’s scraping restrictions have created an unexpected consequence: top-ranking content matters more than ever.

When AI systems and SEO tools could easily access 100 search results at once, lower-ranked content still had visibility. Position 45 was discoverable. Position 78 showed up in comprehensive data pulls.

Now that data collection requires ten times as many requests, systems focus on top results. The first page of search results gets scraped frequently. Page two occasionally. Pages three through ten rarely.

The practical effect: content that doesn’t rank on page one has become functionally invisible not just to human users but to AI systems building knowledge bases.

This creates a reinforcement loop. Top-ranking content gets indexed by AI systems. AI systems then cite and amplify that content. Citations and traffic improve search rankings. Better rankings lead to more AI citations.

Meanwhile, lower-ranked content becomes increasingly marginalized in both traditional search and AI discovery channels.

Quality finally became the differentiator

For years, SEO had a reputation problem. Too many businesses treated it as a technical game of manipulating algorithms rather than a discipline of creating genuinely valuable content.

AI has changed that calculation. Language models are remarkably good at assessing content quality, originality, and expertise. They can detect thin content, keyword stuffing, and manipulative link schemes. They prioritize sources that demonstrate real knowledge and authority.

The businesses benefiting most from the AI-powered discovery landscape share common characteristics:

They publish original research and unique insights rather than rehashing common knowledge. They employ genuine experts who contribute specialized knowledge. They invest in comprehensive, well-researched content that thoroughly addresses topics. They update existing content regularly to maintain accuracy and relevance. They structure information clearly with proper formatting, citations, and references.

In other words, they do SEO the way it was always supposed to be done: by creating genuinely valuable content that serves user needs.

The strategic imperative

Understanding the economics changes the strategic calculation. AI companies will continue scraping the web because retraining remains prohibitively expensive. Search engines will continue serving results because that’s their business model. Content creators who understand this dynamic have an opportunity.

The companies thriving in this environment treat SEO not as a marketing tactic but as foundational infrastructure for digital discoverability. Their content strategies explicitly account for both human readers and AI systems.

They’re asking different questions: Does our content structure help AI parsers understand our expertise? Are we building the kind of authoritative presence that AI systems consider reliable? When AI tools answer questions in our domain, are we getting cited?

These aren’t separate from traditional SEO. They’re extensions of the same principles: create valuable content, structure it clearly, build authority, make it discoverable.

The difference is scale and consequence. Traditional SEO determined whether humans could find you. AI-era SEO determines whether both humans and AI systems can find you, understand you, cite you, and amplify you.

What this means for businesses

The practical implications vary by industry and business model, but several patterns are emerging across successful organizations:

Content investment is increasing, not decreasing. Companies that cut content budgets expecting AI to fill the gap are finding the opposite. Quality content requires more investment in an AI-powered world, not less.

Technical SEO fundamentals matter more. Clean code, fast loading times, mobile optimization, structured data implementation. These technical factors affect both traditional search visibility and AI scraping efficiency.

Authority building has become critical. Backlinks, expert authorship, consistent publishing, industry recognition. AI systems use these same signals to assess source reliability.

Content freshness drives ongoing value. Publishing new content and updating existing content signals ongoing relevance to both search engines and AI systems.

Cross-channel optimization is necessary. Successful strategies work for traditional search, AI search tools, training data pipelines, and direct traffic simultaneously.

The competitive advantage

Companies with strong SEO foundations are discovering an unexpected advantage. The same content strategies that drove Google rankings now drive AI citations. The same technical infrastructure that helped search engines crawl sites helps AI scrapers access content. The same authoritative positioning that built search visibility builds AI credibility.

Meanwhile, competitors who dismissed SEO as obsolete are finding themselves invisible in both traditional and AI-powered discovery.

The gap will widen. AI systems amplify existing authority. Top-ranking content gets cited more, which improves rankings, which drives more citations. Lower-visibility content becomes increasingly marginalized.

This creates a window of opportunity. Organizations that recognize the shift and invest now in comprehensive, authoritative, well-optimized content are building compounding advantages. They’re positioning themselves as the sources AI systems reference, the authorities human users trust, and the destinations both types of searchers ultimately reach.

The bottom line

SEO didn’t die when AI emerged. It evolved into something more fundamental: the infrastructure layer of digital discoverability in a world where both humans and machines search for information.

The economics are clear. AI companies can’t afford constant retraining. They need to scrape the web for fresh information. That means content creators who understand how to be discoverable, authoritative, and useful maintain control over their digital destiny.

The question isn’t whether to invest in SEO. It’s whether you’re investing enough, in the right ways, to remain visible as discovery channels multiply and competition intensifies.

The companies getting this right aren’t treating SEO as a marketing channel. They’re treating it as core infrastructure for how their business gets found, understood, and trusted in an AI-powered world.

That’s not a nice-to-have capability. That’s existential.

By The Numbers

  • $192M: Estimated cost to train Google’s Gemini 1.0 Ultra
  • 2.4x: Annual growth rate of AI model training costs since 2016
  • $1B+: Expected cost of largest AI models by 2027
  • 10x: Cost increase for scraping Google after num=100 removal
  • 80-90%: Percentage of indexed pages effectively hidden from bulk scraping
  • 1.1B: Monthly visits to Google’s Gemini AI chatbot (October 2025)
  • 46%: Month-over-month growth in Gemini usage

What do you think?

More notes