Google’s UCP Checkout Brings New Tradeoffs For Retailers via @sejournal, @MattGSouthern

When Google announced that shoppers could complete purchases directly in AI Mode, the focus was on convenience and technical capability. A retailer who emailed Search Engine Journal raised different questions about what gets lost when the transaction moves to Google’s surfaces.

The retailer cited concerns that customers never visit the store, see accessory recommendations from other sellers, and lose brand connection when making purchases on Google.

The concern shows a tradeoff in Google’s Universal Commerce Protocol. Retailers gain potential access to customers at the moment of purchase intent. However, they may lose some of the brand environment, discovery patterns, and relationship-building that occur when shoppers visit owned sites.

What Changes When Checkout Leaves Your Site

The change affects several parts of how retailers interact with customers.

Cross-selling

Cross-selling may change shape. A customer buying a camera on your site might see lens recommendations, memory cards, or cases based on your merchandising strategy.

Google says it plans to add capabilities like discovering related products, applying loyalty rewards, and powering custom shopping experiences on Google, but it hasn’t detailed reporting, fees, or data-sharing for AI Mode checkout.

If loyalty rewards, saved preferences, and checkout work more smoothly on Google surfaces, some shoppers may prefer that experience even if retailers have less control over it. Whether that tradeoff benefits retailers depends on details Google hasn’t disclosed yet.

Brand Connection

Brand storytelling can get compressed into whatever product data feeds into Google’s systems. Retailers invest in site design, content, and navigation to communicate what makes them different. That investment may not fully transfer when the interaction happens in AI Mode’s standardized interface.

The customer relationship dynamics change. Retailers traditionally owned the full transaction flow: discovery, consideration, purchase, and post-purchase communication. For orders completed inside AI Mode, Google would host more of the discovery and checkout experience on its own surfaces, while retailers remain the seller of record.

The degree to which retailers can access customer journey data that normally informs merchandising and marketing is unknown.

The Amazon Parallel

The situation resembles dynamics that already exist with Amazon marketplace sellers. Third-party sellers on Amazon get access to massive customer traffic. Marketplace sellers often accept less control over the customer experience and limited access to relationship signals compared with selling on their own sites.

Google’s protocol creates similar dynamics but extends them across the open web rather than within a single marketplace. Google positions UCP as an open standard, in contrast to Amazon’s closed marketplace model. The key difference: Amazon requires sellers to list products on its platform. UCP lets Google insert checkout capabilities into AI Mode while products technically remain on participating retailers’ inventory systems.

Whether that distinction leads to more data for retailers or a different platform dependency depends on reporting and data-sharing details Google hasn’t specified.

When It Makes Sense, When It Doesn’t

Some retail business models rely heavily on price, convenience, and fulfillment speed. For these retailers, losing the site visit may matter less if UCP delivers customers when they’re ready to buy.

Other retailers compete on curation, brand experience, and discovery. A customer visiting a specialty outdoor gear retailer expects to explore complementary products, read buying guides, and engage with brand content. Moving more of the purchase flow onto Google surfaces could reduce how much of that value proposition happens on a retailer’s site.

The calculation also depends on customer acquisition costs. For example, if you’re paying $30 to acquire a customer through Google Ads and they buy a $50 product on your site, the unit economics work when you can cross-sell or build long-term relationship value. If checkout happens on Google’s surface and you can’t cross-sell or retarget, the same acquisition cost may not be worth it.

What’s Known Versus What’s Speculation

Google said eligible U.S. retailers will be able to participate in UCP checkout through AI Mode in Search and the Gemini app. Google says retailers remain the seller of record and can customize the integration.

A separate Google Developers blog post explains that merchants remain the Merchant of Record and highlights an embedded option for a customized checkout experience. But the announcement didn’t detail the data-sharing arrangement, fee structure, or the funnel-level reporting retailers will receive for AI Mode checkout events.

The protocol is described as “open,” but adoption requirements, integration complexity, and whether non-Google AI systems can use it are unclear.

Google’s Business Agent feature demonstrates one use of the new protocol: branded AI chat appears in Search results for participating retailers, but the interaction occurs on Google’s platform.

Some analysts frame the change as existential, using terms like “extinction event” for certain retail models. That’s based on assumptions about adoption rates, customer behavior, and competitive dynamics that haven’t played out yet.

The more measured question retailers are asking: Does this create fragmentation where they need to optimize for multiple checkout flows, or consolidation where Google becomes the dominant transaction layer for product searches?

Questions Without Clear Answers

Three implementation details will likely determine how disruptive AI Mode checkout becomes for retailers:

  1. Merchant Center control: whether participation is explicitly opt-in and retailers can limit checkout to specific products or categories.
  2. Measurement: what reporting retailers get for actions on Google surfaces and whether AI Mode orders can be distinguished from standard site conversions.
  3. Customer and journey data: what signals, if any, come back to retailers to support lifecycle marketing and merchandising decisions.

Google has outlined the direction for UCP but hasn’t detailed these operational components.

Looking Ahead

Google said UCP checkout will roll out to eligible U.S. retailers soon, but hasn’t provided specific timing. Business Agent, which puts branded AI chat on Search results, went live Jan. 12.

Retailers questioning the tradeoffs between visibility and control face a pattern that’s played out before with Amazon, Google Shopping, and social commerce. Early participants gain access to new traffic sources but accept platform rules they don’t control. Late adopters may find themselves at a disadvantage.

The core question several retailers have raised is: Can they maintain the brand differentiation and relationship-building that justified creating owned channels when the transaction occurs on someone else’s platform?

The protocol is too new to know yet.


Featured Image: michnik101/Shutterstock

WordPress X Account’s ‘Childish’ Trolling Causes Backlash via @sejournal, @martinibuster

An official WordPress.org social media account was used to troll the open source movement to decentralize the WordPress plugins and themes repository, creating what some feel was an undignified, even “childish”, representation of the WordPress community.

What Is The FAIR Project?

The Federated And Independent Repositories project is an open-source initiative that was launched in 2025 in response to actions by Matt Mullenweg and Automattic that exposed a weakness in how plugins and themes are distributed to WordPress sites. The project was initiated after Mullenweg cut off WP Engine from updating their plugins, disrupting the proper functioning of thousands of websites.

The FAIR goal of the FAIR project is to decentralize the distribution of WordPress plugins and themes to protect against one person from disrupting the free distribution of software.

FAIR is backed by open source giant Linux, announced in June 2025. The official announcement explained that the purpose of FAIR is to create a “vendor-neutral” method for distributing WordPress software within a trusted environment, writing:

“Vendor-neutral package management for content management systems like WordPress provides critical universal infrastructure that addresses the new realities of content, e-commerce and AI.

The FAIR Package Manager project helps make plugins and tools more discoverable and lets developers choose where to source those plugins depending on the needs of their supply chain. By giving commercial plugin developers, hosts, and application developers more options to control the tools they rely on, the FAIR Package Manager project promotes innovation and protects business continuity.”

What Caused An Issue With FAIR?

A WordPress user recently experienced a temporary problem updating their website using the FAIR repository, forcing them to manually SFTP the software updates to their server.

They posted on X:

“Here I am updating one of my sites for the new year, and it looks like FAIR broke my plugin and theme updates.”

After updating their site they returned to X with more thoughts about their experience with FAIR:

“Glad this was just a “for fun” site and not something critical. I like experimenting with stuff in the WordPress ecosystem, but this is a bit too experimental for my taste. Going back to stock updates, at least until 2.0.

…This is making me rethink how I organize my domains and sites. Should probably just set up a sandbox for things like this, but then again… the squeaky wheel gets the grease. If it’s all locked away in a sandbox, I’ll forget to ever touch it.”

There was an issue with an update to FAIR version 1.2.2. According to the release notes:

“FAIR Connect 1.2.2 Release Announcement

Version 1.2.2 of FAIR Connect is a fast follow up to our version 1.2.1 release. This release fixes a fatal error introduced in 1.2.1 that impacts the updating process.

If you previously updated to 1.2.1, you will need to perform this update manually.”

So apparently there’s an issue with updating the FAIR Connect plugin which requires manually deactivating the FAIR Connect plugin, downloading the updated version of the plugin from the FAIR repository, then manually uploading the plugin from the WordPress admin plugin dashboard (unless the site is unavailable, which necessitates SFTP’ing the updated plugin).

WordPress Trolls The FAIR Project

The official WordPress.org X account posted the following comment about the FAIR project:

“Looks like the Federated and Independent Repository project is going great. This is clearly going to rock the WordPress world. We don’t know how we’ll continue without these contributors. Maybe they need some REST.”

The post was highly unusual for the WordPress X account because it’s normally a feel-good destination of announcements and inspiration related to WordPress. The unprofessional tone of the post caught many in the WordPress community by surprise.

One person shared their disappointment:

“Hi Matt! These comments aren’t clearly going to rock the atmosphere in our community too. So, http://WP.org never had issues?”

RapidLightnings responded:

“These people working at or for WordPress are so childish and unprofessional. Professional people wouldn’t care or would not post stuff like that on official accounts.”

Responses Hidden By WordPress

There were additional responses that were hidden by WordPress:

Like this by o_be_one:

“For an OpenSource project, your take is toxic af.”

Rohan K called the post by the official WordPress account immature:

“Growing pains. Why are you gleefully gloating about this, when your immature and short-sighted actions led the creation of it? It makes you look bad.

Grow up.”

Aron Prins posted a one-word response:

“Ewww”

Thisbit commented on how it reflects poorly on the WordPress leadership:

“Shameful leadership.”

Jono Alderson reflected on the childishness of the tweet:

“Oh hush. Your misuse of this account for sniping is childish and tedious. Be better.”

Other posts were directed at Matt Mullenweg, with this one prematurely dancing on WordPress’s grave:

“SO HAPPY that AI is ending WordPress for good.
Ciao CattyMatty”

And this one:

“I’d say get a clue, but you’d probably steal it from another developer.”

Jono Alderson’s Response

Alderson started a new discussion to express his opinion about the WordPress troll-post:

“I love WordPress-the-software, but this kind of childish nonsense makes me ashamed and embarrassed to be associated with WordPress-the-brand. What childish, petty, unprofessional, shameful, amateur nonsense. All of these people need firing and replacing with capable grown-ups.”

The responses to Jono’s post generally expressed disappointment that the official WordPress account was used for trolling, with one person responding that it seemed crazy.

Featured Image by Shutterstock/AYO Production

Google Announces AI Mode Checkout Protocol, Business Agent via @sejournal, @MattGSouthern

Google announced tools that let shoppers complete purchases directly within AI Mode and chat with branded AI agents in Search results.

Users can purchase from eligible product listings on Google. Retailers are still the seller of record, while the checkout happens on Google surfaces instead of the retailer’s website.

Universal Commerce Protocol Powers AI Mode Checkout

Google launched the Universal Commerce Protocol, an open standard for what it calls “agentic commerce.” The protocol will power checkout on eligible Google product listings in AI Mode in Search and the Gemini app.

Google developed UCP with Shopify, Etsy, Wayfair, Target, and Walmart. More than 20 additional companies endorsed it, including Adyen, American Express, Best Buy, Mastercard, Stripe, The Home Depot, and Visa.

Shoppers will use Google Pay with payment methods and shipping info from Google Wallet. PayPal support is coming. UCP checkout starts with eligible U.S. retailers, with global expansion planned.

Business Agent Brings Branded Chat To Search

Business Agent lets shoppers chat with brands in Search results. Google describes it as a “virtual sales associate” that can answer product questions in the brand’s voice.

The feature goes live January 12 with Lowe’s, Michael’s, Poshmark, Reebok, and others. Eligible U.S. retailers can activate and customize the agent through Merchant Center.

Google plans to add capabilities for training agents on retailer data, providing product offers, and enabling purchases within the chat experience.

Direct Offers Pilot Tests Ads In AI Mode

Google also announced Direct Offers, a new ad pilot in AI Mode. It allows advertisers to offer exclusive discounts to people searching for products.

Google gave an example of a rug search where relevant retailers could feature a special 20% discount. Retailers set up offers in campaign settings, and Google determines when to display them.

Early partners include Petco, e.l.f. Cosmetics, Samsonite, Rugs USA, and Shopify merchants.

Why This Matters

Checkout in AI Mode means a user searching for a product can research, compare, and buy without ever reaching the retailer’s site.

For ecommerce sites, this changes the traffic equation. The sale still happens, but the site visit may not. Retailers participating in UCP gain access to high-intent buyers at the moment of decision. Those who don’t participate may find their products harder to surface when users expect to complete transactions without leaving Google.

Looking Ahead

Checkout in AI Mode rolls out to eligible U.S. retailers soon. Business Agent launches January 12. Direct Offers is in pilot with select advertisers.

Google said it plans to add new Merchant Center data attributes designed for discovery in AI Mode, Gemini, and Business Agent. The company will roll out the new attributes with a small group of retailers soon before expanding more broadly.


Featured Image: hafakot/Shutterstock

Google: AI Overviews Show Less When Users Don’t Engage via @sejournal, @MattGSouthern

AI Overviews don’t show up consistently across Google Search because the system learns where they’re useful and pulls them back when people don’t engage.

Robby Stein, Vice President of Product at Google Search, described in a CNN interview how Google tests the summaries, measures interaction, and reduces their appearance for certain kinds of searches where they don’t help.

How Google Decides When To Show AI Overviews

Stein explained that AI Overviews appear based on learned usefulness rather than showing up by default.

“The system actually learns where they’re helpful and will only show them if users have engaged with that and find them useful,” Stein said. “For many questions, people just ask like a short question or they’re looking for very specific website, they won’t show up because they’re not actually helpful in many many cases.”

He gave a concrete example. When someone searches for an athlete’s name, they typically want photos, biographical details, and social media links. The system learned people didn’t engage with an AI Overview for those queries.

“The system will learn that if it tried to do an AI overview, no one really clicked on it or engaged with it or valued it,” Stein said. “We have lots of metrics we look at that and then it won’t show up.”

What “Under The Hood” Queries Mean For Visibility

Stein described the system as sometimes expanding a search beyond what you type. Google “in many cases actually issues additional Google queries under the hood to expand your search and then brings you the most relevant information for a given question,” he said.

That may help explain why pages sometimes show up in AI Overview citations even when they don’t match your exact query wording. The system pulls in content answering related sub-questions or providing context.

For image-focused queries, AI Overviews integrate with image results. For shopping queries, they connect to product information. The system adapts based on what serves the question.

Where AI Mode Fits In

Stein described AI Mode as the next step for complicated questions that need follow-up conversation. The design assumes you start in traditional Search, get an Overview if it helps, then go deeper into AI Mode when you need more.

“We really designed AI Mode to really help you go deeper with a pretty complicated question,” Stein said, citing examples like comparing cars or researching backup power options.

During AI Mode testing, Google saw “like a two to three … full increase in the query length” compared to typical Search queries. Users also started asking follow-up questions in a conversational pattern.

The longer AI Mode queries included more specificity. Stein’s example: instead of “things to do in Nashville,” users asked “restaurants to go to in Nashville if one friend has an allergy and we have dogs and we want to sit outside.”

Personalization Exists But Is Limited

Some personalization in AI Mode already exists. Users who regularly click video results might see videos ranked higher, for example.

“We are personalizing some of these experiences,” Stein said. “But right now that’s a smaller adjustment probably to the experience because we want to keep it as consistent as possible overall.”

Google’s focus is on maintaining consistency across users while allowing for individual preferences where it makes sense.

Why This Matters

In July 2024, research showed Google had dialed back AIO presence by 52%, from widespread appearance to showing for just 8% of queries. Stein’s description offers one possible explanation for that pattern.

If you’re tracking AIO presence week to week, the fluctuations may reflect user behavior patterns for different question types rather than algorithm changes.

The “under the hood” query expansion means content can appear in citations even without matching your exact phrasing. That matters when you’re explaining CTR drops internally or planning content for complex queries where Overviews are more likely to surface.

Looking Ahead

Google’s AI Overviews earn placement based on usefulness rather than appearing by default.

Personalization is limited today, but the direction is moving toward more tailored experiences that maintain overall consistency.

See the full interview with Stein below:


Featured Image: nwz/Shutterstock

Google Gemini Gains Share As ChatGPT Declines In Similarweb Data via @sejournal, @MattGSouthern

ChatGPT accounted for 64% of worldwide traffic share among gen AI chatbot websites as of January, while Google’s Gemini reached 21%, according to Similarweb’s Global AI Tracker.

Similarweb’s tracker (PDF link) measures total visits at the domain level, so it reflects people who go to these tools directly on the web. It doesn’t capture API usage, embedded assistants, or other integrations where much of the AI usage occurs now.

ChatGPT Down, Gemini Up In A Year Of Share Gains

The share movement is easiest to see year-over-year.

A year ago, Similarweb estimated ChatGPT accounted for 86% of worldwide traffic among tracked chatbot sites. Now, that figure is 64%. Over the same period, Gemini rose from 5% to 21%.

Other tools are much smaller by this measure. DeepSeek was at 3.7%, Grok at 3.4%, and Perplexity and Claude both at 2.0%.

Google has been promoting Gemini through products like Android and Workspace, which may help explain why it’s gaining share among users who access these tools directly.

Winter Break Pulled Down Total Visits

Similarweb pointed to seasonality during the holiday period:

Similarweb wrote on X:

“Driven by the winter break, the daily average visits to all tools dropped to August-September levels.”

That context matters because it helps distinguish overall category softness from shifts in market share.

Writing Tool Domain Traffic Declines

Writing and content generation sites were down 10% over the most recent 12-week window in Similarweb’s category view.

At the individual tool level, Similarweb’s table shows steep drops for several writing platforms. Growthbarseo was down 100%, while Jasper fell 16%, Writesonic dropped 17%, and Rytr declined 9%. Originality was up 17%.

These are still domain-level visit counts, so the clearest takeaway is that fewer people are going directly to specialized writing sites online. That can happen for several reasons, including users relying more on general assistants, switching to apps, or using these models through integrations.

Code Completion Shows Mixed Results

The developer tools category looked more mixed than the writing tools.

Similarweb’s code completion table shows Bolt down 39% over 12 weeks, while Cursor (up 8%), Replit (up 2%), and Base44 (up 49%) moved in different directions.

Traditional Search Looks Close To Flat

In Similarweb’s “disrupted sectors” view, traditional search traffic is down roughly 1% to 3% year-over-year across recent periods, which doesn’t indicate a sharp drop in overall search usage in this dataset.

The same table shows Reddit up 12% year-over-year and Quora down 53%, consistent with the idea that some Q&A behavior is being redistributed even as overall search remains relatively steady.

Why This Matters

When making sense of how AI is changing discovery and demand, these numbers can help you understand where direct, web-based attention is concentrating. That can influence which assistants you monitor for brand mentions, citations, and referral behavior.

Though you should treat this a snapshot, not the full picture. If your audience is interacting with AI through browsers, apps, or embedded assistants, your own analytics will be a better barometer than any domain-level tracker.

Looking Ahead

The next report should clarify whether category traffic rebounds after the holiday period and whether Gemini continues to gain share at the same pace. It will also be a useful read on whether writing tools stabilize or whether more of that usage continues to consolidate into general assistants and bundled experiences.


Featured Image: vfhnb12/Shutterstock

Most Major News Publishers Block AI Training & Retrieval Bots via @sejournal, @MattGSouthern

Most top news publishers block AI training bots via robots.txt, but they’re also blocking the retrieval bots that determine whether sites appear in AI-generated answers.

BuzzStream analyzed the robots.txt files of 100 top news sites across the US and UK and found 79% block at least one training bot. More notably, 71% also block at least one retrieval or live search bot.

Training bots gather content to build AI models, while retrieval bots fetch content in real time when users ask questions. Sites blocking retrieval bots may not appear when AI tools try to cite sources, even if the underlying model was trained on their content.

What The Data Shows

BuzzStream examined the top 50 news sites in each market based on SimilarWeb traffic share, then deduplicated the list. The study grouped bots into three categories: training, retrieval/live search, and indexing.

Training Bot Blocks

Among training bots, Common Crawl’s CCBot was the most frequently blocked at 75%, followed by Anthropic-ai at 72%, ClaudeBot at 69%, and GPTBot at 62%.

Google-Extended, which trains Gemini, was the least blocked training bot at 46% overall. US publishers blocked it at 58%, nearly double the 29% rate among UK publishers.

Harry Clarkson-Bennett, SEO Director at The Telegraph, told BuzzStream:

“Publishers are blocking AI bots using the robots.txt because there’s almost no value exchange. LLMs are not designed to send referral traffic and publishers (still!) need traffic to survive.”

Retrieval Bot Blocks

The study found 71% of sites block at least one retrieval or live search bot.

Claude-Web was blocked by 66% of sites, while OpenAI’s OAI-SearchBot, which powers ChatGPT’s live search, was blocked by 49%. ChatGPT-User was blocked by 40%.

Perplexity-User, which handles user-initiated retrieval requests, was the least blocked at 17%.

Indexing Blocks

PerplexityBot, which Perplexity uses to index pages for its search corpus, was blocked by 67% of sites.

Only 14% of sites blocked all AI bots tracked in the study, while 18% blocked none.

The Enforcement Gap

The study acknowledges that robots.txt is a directive, not a barrier, and bots can ignore it.

We covered this enforcement gap when Google’s Gary Illyes confirmed robots.txt can’t prevent unauthorized access. It functions more like a “please keep out” sign than a locked door.

Clarkson-Bennett raised the same point in BuzzStream’s report:

“The robots.txt file is a directive. It’s like a sign that says please keep out, but doesn’t stop a disobedient or maliciously wired robot. Lots of them flagrantly ignore these directives.”

Cloudflare documented that Perplexity used stealth crawling behavior to bypass robots.txt restrictions. The company rotated IP addresses, changed ASNs, and spoofed its user agent to appear as a browser.

Cloudflare delisted Perplexity as a verified bot and now actively blocks it. Perplexity disputed Cloudflare’s claims and published a response.

For publishers serious about blocking AI crawlers, CDN-level blocking or bot fingerprinting may be necessary beyond robots.txt directives.

Why This Matters

The retrieval-blocking numbers warrant attention here. In addition to opting out of AI training, many publishers are opting out of the citation and discovery layer that AI search tools use to surface sources.

OpenAI separates its crawlers by function: GPTBot gathers training data, while OAI-SearchBot powers live search in ChatGPT. Blocking one doesn’t block the other. Perplexity makes a similar distinction between PerplexityBot for indexing and Perplexity-User for retrieval.

These blocking choices affect where AI tools can pull citations from. If a site blocks retrieval bots, it may not appear when users ask AI assistants for sourced answers, even if the model already contains that site’s content from training.

The Google-Extended pattern is worth watching. US publishers block it at nearly twice the UK rate, though whether that reflects different risk calculations around Gemini’s growth or different business relationships with Google isn’t clear from the data.

Looking Ahead

The robots.txt method has limits, and sites that want to block AI crawlers may find CDN-level restrictions more effective than robots.txt alone.

Cloudflare’s Year in Review found GPTBot, ClaudeBot, and CCBot had the highest number of full disallow directives across top domains. The report also noted that most publishers use partial blocks for Googlebot and Bingbot rather than full blocks, reflecting the dual role Google’s crawler plays in search indexing and AI training.

For those tracking AI visibility, the retrieval bot category is what to watch. Training blocks affect future models, while retrieval blocks affect whether your content shows up in AI answers right now.


Featured Image: Kitinut Jinapuck/Shutterstock

Google’s Mueller Weighs In On SEO vs GEO Debate via @sejournal, @MattGSouthern

Google Search Advocate John Mueller says businesses that rely on referral traffic should think about how AI tools fit into the picture.

Mueller responded to a Reddit thread asking whether SEO is still enough or whether practitioners need to start considering GEO, a term some in the industry use for optimizing visibility in AI-powered answer engines like ChatGPT, Gemini, and Perplexity.

“If you have an online business that makes money from referred traffic, it’s definitely a good idea to consider the full picture, and prioritize accordingly,” Mueller wrote.

What Mueller Said

Mueller didn’t endorse or reject the GEO terminology. He framed the question in terms of practical business decisions rather than new optimization techniques.

“What you call it doesn’t matter, but ‘AI’ is not going away, but thinking about how your site’s value works in a world where ‘AI’ is available is worth the time,” he wrote.

He also pushed back on treating AI visibility as a universal priority. Mueller suggested practitioners look at their own data first.

Mueller added:

“Also, be realistic and look at actual usage metrics and understand your audience (what % is using ‘AI’? what % is using Facebook? what does it mean for where you spend your time?).”

Why This Matters

I’ve been tracking Mueller’s public statements for years, and this one lands differently than the usual “it depends” responses he’s known for. He’s reframing the GEO question as a resource allocation problem rather than a terminology debate.

The GEO conversation has picked up steam over the past year as AI answer engines started sending measurable referral traffic. I’ve covered the citation studies, the traffic analyses, and the research comparing Google rankings to LLM citations. What’s been missing is a clear signal from Google: is this a distinct discipline, or just rebranded SEO?

Mueller’s answer is consistent with what Google said at Search Central Live, when Gary Illyes emphasized that AI features share infrastructure with traditional Search. The message from both is that you probably don’t need a separate framework, but you do need to understand how discovery is changing.

What I find more useful is his emphasis on checking your own numbers. Current data shows ChatGPT referrals at roughly 0.19% of traffic for the average site. AI assistants combined still drive less than 1% for most publishers. That’s growing, but it’s not yet a reason to reorganize your entire strategy.

The industry has a habit of chasing trends that apply to some sites but not others. Mueller’s pushing back on that pattern. Look at what percentage of your audience actually uses AI tools before reallocating resources toward them.

Looking Ahead

The GEO terminology will likely stick, regardless of Google’s stance. Mueller’s framing puts the decision back on individual businesses to measure their own audience behavior.

For practitioners, this means the homework is in your analytics. If AI referrals are showing up in your traffic sources, they’re worth understanding. If they’re not, you have other priorities.


Featured Image: Roman Samborskyi/Shutterstock

Google’s Mueller Explains ‘Page Indexed Without Content’ Error via @sejournal, @MattGSouthern

Google Search Advocate John Mueller responded to a question about the “Page Indexed without content” error in Search Console, explaining the issue typically stems from server or CDN blocking rather than JavaScript.

The exchange took place on Reddit after a user reported their homepage dropped from position 1 to position 15 following the error’s appearance.

What’s Happening?

Mueller clarified a common misconception about the cause of “Page Indexed without content” in Search Console.

Mueller wrote:

“Usually this means your server / CDN is blocking Google from receiving any content. This isn’t related to anything JavaScript. It’s usually a fairly low level block, sometimes based on Googlebot’s IP address, so it’ll probably be impossible to test from outside of the Search Console testing tools.”

The Reddit user had already attempted several diagnostic steps. They ran curl commands to fetch the page as Googlebot, checked for JavaScript blocking, and tested with Google’s Rich Results Test. Desktop inspection tools returned “Something went wrong” errors while mobile tools worked normally.

Mueller noted that standard external testing methods won’t catch these blocks.

He added:

“Also, this would mean that pages from your site will start dropping out of the index (soon, or already), so it’s a good idea to treat this as something urgent.”

The affected site uses Webflow as its CMS and Cloudflare as its CDN. The user reported the homepage had been indexing normally with no recent changes to the site.

Why This Matters

I’ve covered this type of problem repeatedly over the years. CDN and server configurations can inadvertently block Googlebot without affecting regular users or standard testing tools. The blocks often target specific IP ranges, which means curl tests and third-party crawlers won’t reproduce the problem.

I covered when Google first added “indexed without content” to the Index Coverage report. Google’s help documentation at the time noted the status means “for some reason Google could not read the content” and specified “this is not a case of robots.txt blocking.” The underlying cause is almost always something lower in the stack.

The Cloudflare detail caught my attention. I reported on a similar pattern when Mueller advised a site owner whose crawling stopped across multiple domains simultaneously. All affected sites used Cloudflare, and Mueller pointed to “shared infrastructure” as the likely culprit. The pattern here looks familiar.

More recently, I covered a Cloudflare outage in November that triggered 5xx spikes affecting crawling. That was a widespread incident. This case appears to be something more targeted, likely a bot protection rule or firewall setting that treats Googlebot’s IP addresses differently from other traffic.

Search Console’s URL Inspection tool and Live URL test remain the primary ways to identify these blocks. When those tools return errors while external tests pass, server-level blocking becomes the likely cause. Mueller made a similar point in August when advising on crawl rate drops, suggesting site owners “double-check what actually happened” and verify “if it was a CDN that actually blocked Googlebot.”

Looking Ahead

If you’re seeing the “Page Indexed without content” error, check the CDN and server configurations for rules that affect Googlebot’s IP ranges. Google publishes its crawler IP addresses, which can help identify whether security rules are targeting them.

The Search Console URL Inspection tool is the most reliable way to see what Google receives when crawling a page. External testing tools won’t catch IP-based blocks that only affect Google’s infrastructure.

For Cloudflare users specifically, check bot management settings, firewall rules, and any IP-based access controls. The configuration may have changed through automatic updates or new default settings rather than manual changes.

Google Ads Using New AI Model To Catch Fraudulent Advertisers via @sejournal, @martinibuster

Google published a research paper about a new AI model for detecting fraud in the Google Ads system that’s a strong improvement over what they were previously using. What’s interesting is that the research paper, dated December 31, 2025,  says that the new AI is deployed, resulting in an improvement in the detection rate of over 40 percentage points and achieving 99.8% precision on specific policies.

ALF: Advertiser Large Foundation Model

The new AI is called ALF (Advertiser Large Foundation Model), the details of which were published on December 31, 2025. ALF is a multimodal large foundation model that analyzes text, images, and video, together with factors like account age, billing details, and historical performance metrics.

The researchers explain that many of these factors in isolation won’t flag an account as potentially problematic, but that comparing all of these factors together provides a better understanding of advertiser behavior and intent.

They write:

“A core challenge in this ecosystem is to accurately and efficiently understand advertiser intent and behavior. This understanding is critical for several key applications, including matching users with ads and identifying fraud and policy violations.

Addressing this challenge requires a holistic approach, processing diverse data types including structured account information (e.g., account age, billing details), multi-modal ad creative assets (text, images, videos), and landing page content.

For example, an advertiser might have a recently created account, have text and image ads for a well known large brand, and have had a credit card payment declined once. Although each element could exist innocently in isolation, the combination strongly suggests a fraudulent operation.”

The researchers address three challenges that previous systems were unable to overcome:

1. Heterogeneous and High-Dimensional Data
Heterogeneous data refers to the fact that advertiser data comes in multiple formats, not just one type. This includes structured data like account age and billing type and unstructured data like creative assets such as images, text, and video. High-dimensional data refers to the hundreds or thousands of data points associated with each advertiser, causing the mathematical representation of each one to become high-dimensional, which presents challenges for conventional models.

2. Unbounded Sets of Creative Assets
Advertisers could have thousands of creative assets, such as images, and hide one or two malicious ones among thousands of innocent assets. This scenario overwhelmed the previous system.

3. Real-World Reliability and Trustworthiness
The system needs to be able to generate trustworthy confidence scores that a business has malicious intent because a false positive would otherwise affect an innocent advertiser. The system must be expected to work without having to constantly retune it to catch mistakes.

Privacy and Safety

Although ALF analyzes sensitive signals like billing history and account details, the researchers emphasize that the system is designed with strict privacy safeguards. Before the AI processes any data, all personally identifiable information (PII) is stripped away. This ensures that the model identifies risk based on behavioral patterns rather than sensitive personal data.

The Secret Sauce: How It Spots Outliers

The model also uses a technique called “Inter-Sample Attention” to improve its detection skills. Instead of analyzing a single advertiser in a vacuum, ALF looks at “large advertiser batches” to compare their interactions against one another. This allows the AI to learn what normal activity looks like across the entire ecosystem and make it more accurate in spotting suspicious outliers that don’t fit into normal behavior.

Alf Outperforms Production Benchmarks

The researchers explain that their tests show that ALF outperforms a heavily tuned production baseline:

“Our experiments show ALF significantly outperforms a heavily tuned production baseline while also performing strongly on public benchmarks. In production, ALF delivers substantial and simultaneous gains in precision and recall, boosting recall by over 40 percentage points on one critical policy while increasing precision to 99.8% on another.”

This result demonstrates that ALF can deliver measurable gains across multiple evaluation criteria under actual real-world production conditions, rather than just in offline or benchmarked environments.

Elsewhere they mention tradeoffs in speed:

“The effectiveness of this approach was validated against an exceptionally strong production baseline, itself the result of an extensive search across various architectures and hyperparameters, including DNNs, ensembles, GBDTs, and logistic regression with feature cross exploration.

While ALF’s latency is higher due to its larger model size, it remains well within the acceptable range for our production environment and can be further optimized using hardware accelerators. Experiments show ALF significantly outperforms the baseline on key risk detection tasks, a performance lift driven by its unique ability to holistically model content embeddings, which simpler architectures struggled to leverage. This trade-off is justified by its successful deployment, where ALF serves millions of requests daily.”

Latency refers to the amount of time the system takes to produce a response after receiving a request, and the researcher data shows that although ALF increases this response time relative to the baseline, the latency remains acceptable for production use and is already operating at scale while delivering substantially better fraud detection performance.

Improved Fraud Detection

The researchers say that ALF is now deployed to the Google Ads Safety system for identifying advertisers that are violating Google Ads policies. There is no indication that the system is being used elsewhere such as in Search or Google Business Profiles. But they did say that future work could focus on time-based factors (“temporal dynamics”) for catching evolving patterns. They also indicated that it could be useful for audience modeling and creative optimization.

Read the original PDF version of the research paper:

ALF: Advertiser Large Foundation Model for Multi-Modal Advertiser Understanding

Featured Image by Shutterstock/Login

The Guardian: Google AI Overviews Gave Misleading Health Advice via @sejournal, @MattGSouthern

The Guardian published an investigation claiming health experts found inaccurate or misleading guidance in some AI Overview responses for medical queries. Google disputes the reporting and says many examples were based on incomplete screenshots.

The Guardian said it tested health-related searches and shared AI Overview responses with charities, medical experts, and patient information groups. Google told The Guardian the “vast majority” of AI Overviews are factual and helpful.

What The Guardian Reported Finding

The Guardian said it tested a range of health queries and asked health organizations to review the AI-generated summaries. Several reviewers said the summaries included misleading or incorrect guidance.

One example involved pancreatic cancer. Anna Jewell, director of support, research and influencing at Pancreatic Cancer UK, said advising patients to avoid high-fat foods was “completely incorrect.” She added that following that guidance “could be really dangerous and jeopardise a person’s chances of being well enough to have treatment.”

The reporting also highlighted mental health queries. Stephen Buckley, head of information at Mind, said some AI summaries for conditions such as psychosis and eating disorders offered “very dangerous advice” and were “incorrect, harmful or could lead people to avoid seeking help.”

The Guardian cited a cancer screening example too. Athena Lamnisos, chief executive of the Eve Appeal cancer charity, said a pap test being listed as a test for vaginal cancer was “completely wrong information.”

Sophie Randall, director of the Patient Information Forum, said the examples showed “Google’s AI Overviews can put inaccurate health information at the top of online searches, presenting a risk to people’s health.”

The Guardian also reported that repeating the same search could produce different AI summaries at different times, pulling from different sources.

Google’s Response

Google disputed both the examples and the conclusions.

A spokesperson told The Guardian that many of the health examples shared were “incomplete screenshots,” but from what the company could assess they linked “to well-known, reputable sources and recommend seeking out expert advice.”

Google told The Guardian the “vast majority” of AI Overviews are “factual and helpful,” and that it “continuously” makes quality improvements. The company also argued that AI Overviews’ accuracy is “on a par” with other Search features, including featured snippets.

Google added that when AI Overviews misinterpret web content or miss context, it will take action under its policies.

The Broader Accuracy Context

This investigation lands in the middle of a debate that’s been running since AI Overviews expanded in 2024.

During the initial rollout, AI Overviews drew attention for bizarre results, including suggestions involving glue on pizza and eating rocks. Google later said it would reduce the scope of queries that trigger AI-written summaries and refine how the feature works.

I covered that launch, and the early accuracy problems quickly became part of the public narrative around AI summaries. The question then was whether the issues were edge cases or something more structural.

More recently, data from Ahrefs suggests medical YMYL queries are more likely than average to trigger AI Overviews. In its analysis of 146 million SERPs, Ahrefs reported that 44.1% of medical YMYL queries triggered an AI Overview. That’s more than double the overall baseline rate in the dataset.

Separate research on medical Q&A in LLMs has pointed to citation-support gaps in AI-generated answers. One evaluation framework, SourceCheckup, found that many responses were not fully supported by the sources they cited, even when systems provided links.

Why This Matters

AI Overviews appear above ranked results. When the topic is health, errors carry more weight.

Publishers have spent years investing in documented medical expertise to meet. This investigation puts the same spotlight on Google’s own summaries when they appear at the top of results.

The Guardian’s reporting also highlights a practical problem. The same query can produce different summaries at different times, making it harder to verify what you saw by running the search again.

Looking Ahead

Google has previously adjusted AI Overviews after viral criticism. Its response to The Guardian indicates it expects AI Overviews to be judged like other Search features, not held to a separate standard.