Why Product Feeds Shouldn’t Be The Most Ignored SEO System In Ecommerce

Most ecommerce brands obsess over category pages and backlinks or product optimizations, while their product feeds remain auto-generated and underoptimized. Product feeds act as the backbone of ecommerce site catalogs and have long been the sole remit of PPC teams, but in the new era of AI Search, this is changing.

Back in 2023, Search Console added enhancements to the Shopping tab Listings report to help brands to get a better understanding of how their products were being seen in the Merchant Center.

We’ve also seen the emergence of OpenAI’s Product Feed specification as a specific requirement to allow ChatGPT to accurately index and display products. Although more recently, we’ve seen announcements that OpenAI has ended Instant Checkout and considering new directions.

These changes are pulling product feed visibility directly into the SEO performance ecosystem and aligning it as general “search infrastructure,” not just “ads infrastructure.”

In this article, we’ll be talking you through the value that product feeds can bring to businesses and how SEO aligns with this.

SEO’s Role In Product Feeds

In ecommerce, product feeds are often seen as “set it and forget it” assets, but treating these feeds as simply raw data is an immediate missed opportunity to boost visibility across organic search, shopping, and agentic commerce in the future.

While a standard product feed provides basic data to search bots, an optimized feed enhances attribute accuracy to ensure your products appear for high-intent search queries. By refining your product data, you bridge the gap between technical specs and consumer needs, increasing both visibility and click-through rates.

SEO can help to optimize feeds across four main pillars:

1. Semantic Query Mapping

SEOs don’t just use basic product names. They use consumer language built out of query mapping and intent-matching.

By front-loading titles with high-intent keywords and “long-tail” descriptions that include attributes like color, material, or use-case, products are more likely to appear where the user’s intent is highest.

Example:

Instead of “Men’s Waterproof Jacket Black”

SEO Driven Product Feed: “Brand X Men’s Waterproof Running Jacket – Black Lightweight Performance Shell”

2. Taxonomy Logic

Taxonomy is important to stop your products from being lost in the void. A misplaced product can quickly become a lost sale.

By refining categorization and product grouping, general terms like “tactical hiking boots” won’t get buried under generalized categories like “general footwear.”

Building a logical hierarchy allows algorithms to crawl and understand the catalog with higher confidence of exactly who the product is targeting. All products within your feed will be automatically assigned a product category.

Ensuring your taxonomy, as well as the titles, descriptions, and GTIN information, will help to ensure that products are correctly categorized according to [google_product_category] attribute.

3. Structured Data

In Google Shopping, structured data acts as the anchor of “truth” that connects your website to your Merchant Center feed.

Structured data allows Google and other bots to directly pull product data from your HTML, creating a form of automated data validation. If, for example, your feed says a product is $50, but your schema says $60, Google will likely disapprove the listing.

In many cases, high-performing feeds rely on structured data to update price and availability in real-time. If you run a flash sale, Google’s crawler can detect the change via schema and updates your Shopping Ads, preventing “out of stock” clicks.

When it comes to agentic commerce, agents will query schema properties to see if your product fits the user’s specific constraints.

Structured data provides hard facts and allows agents to see if a product is “agent-ready” for checkout.

4. Analytical Review

Having a highly analytical mind that is always looking for opportunity, SEOs can help to identify any “ghost products” and diagnose whether the issues are down to attributes, images, or descriptions, providing ongoing optimization recommendations.

As we move into an era of AI-driven discovery, the quality of a brand’s feed data can quickly become a reflection of a brand’s reputation.

By providing more context within the feed, you are more likely to see your brand get recommended in conversational search and show up in organic shopping.

What Ecommerce Brands Get Wrong With Product Feed Optimization

The majority of issues that we see in product feeds come from inconsistencies and a lack of depth within the feed.

From conversations with brand managers, this seems to stem from a lack of ownership within a channel and a lack of understanding of the impact of what these inconsistencies can have.

In some cases, feeds can be disapproved due to having inaccurate price status due to inconsistency between the feed and a landing page.

Other common issues include:

  • Auto-generated Shopify titles.
  • No keyword layering.
  • Inconsistent variants.
  • Missing GTIN/MPN.
  • Thin descriptions.
  • Feed data not aligned with on-page SEO.

This is where having the eyes of an SEO who is used to ongoing technical auditing and hygiene maintenance, and understands the value of structured data and content for context, can be vital in product feed performance.

How Product Feeds Directly Impact Organic & AI Visibility

Quite simply, the more context you can provide in your product feed, the more chances you have of being shown or cited in traditional search and in AI engines.

If a product feed is missing critical attributes like size, color, material, compatibility, or use case, the product won’t just rank lower; it will become ineligible for more specific, high-intent queries.

As search queries grow longer and intent becomes more nuanced, i.e., searchers looking for “men’s waterproof trail running jacket black medium” rather than just “men’s trail running jacket,” feeds need to evolve past being simple descriptors.

They need to properly layer structured attributes that mirror how real customers search and filter online. The more complete the product feed, the more opportunities there will be for your products to appear online across Shopping to AI-generated citations.

What Product Feed Optimization Actually Looks Like

There are a few stages of product feed optimization that SEOs need to be both aware of and able to deliver.

Keyword & Intent Architecture

SEOs should approach product feeds the same way they approach category and content strategy.

Keyword research should be conducted at a product level, identifying high-intent modifiers such as size, material, compatibility, and demographic, and layer those attributes both into product titles and feed data.

Rather than relying on generic exports from Shopify or another ecommerce platform, product titles should reflect real organic search behavior around how customers actually query products.

Structured Data Alignment

SEOs should also make sure that feed attributes match on-page schema.

Keeping a close eye on Merchant Center for any potential issues, such as missing GTINs or prices not matching, and making any necessary adjustments to schema/structured data, will help to ensure that the feed is consistent and context is fully delivered to bots.

Variant Consolidation Strategy

This leans heavily into faceted navigation – which ecommerce SEOs have been battling for years.

By determining when product variations should be grouped under a single parent entity versus a standalone URL, SEOs can have more control over any unnecessary duplication and cannibalization.

This can also help to protect crawl efficiencies across large product catalogs and declutter product feeds.

Feed Health Monitoring

Similar to how SEOs regularly run technical crawls of websites to maintain hygiene and pick up any issues, SEOs should also treat feed governance as part of their regular checks.

This includes actively monitoring feed errors and addressing any Merchant Center issues that might limit visibility.

Prioritizing AI Search Readiness

A large opportunity for the future of search comes with agentic commerce, and product feeds are going to align directly with this.

By ensuring feeds are clearly structured and contain complete and accurate attributes, SEOs can reinforce strong product entity signals and provide clarity, which AI systems rely on to determine what to display in comparisons and recommendations.

Final Thoughts

Product feeds are no longer just paid media assets; they are core search infrastructure that directly impacts organic shopping visibility and AI-driven discovery.

Even the strongest category pages can’t compensate for inconsistent or poorly structured data at scale.

As search becomes more conversational and comparative, structured product clarity is going to be the difference between brands that are cited and brands that are not.

More Resources:


Featured Image: Roman Samborskyi/Shutterstock

Google Says It Can Handle Multiple URLs To The Same Content via @sejournal, @martinibuster

Google’s John Mueller answered a question about duplicate URLs appearing after a site structure change. His response offers clarity about how Google handles duplicate content and what actually influences indexing and ranking decisions.

Concern About Duplicate URLs And Ranking Impact

A site owner had changed the URL structure of their web pages then later discovered that older versions of those URLs were still accessible and appearing in Google Search Console.

The person asking the question on Reddit was concerned that requesting recrawls of the older URLs might confuse Google or lead to ranking issues.

They asked:

“I switched over themes a while back and did some redesign and at some point …I changed all my recipes urls by taking the /recipe/ part out of site.com/recipe/actualrecipe so it’s now just site.com/actualrecipe but there are urls that still work when you put the /recipe/ back in the url.

I went to GSC and panicked that a bunch of my recipes weren’t indexed due to a 5xx error (I think it was when my site was down for a few days).

Now I’ve requested a bunch of them already to be recrawled, but realizing maybe google was ignoring them for a reason, like it didn’t want the duplicates.

Are my recrawl requests for /recipe/ urls going to confuse google who might penalize my ranking for the duplicates?”

The question reflects a reasonable concern that duplicate URLs and content might negatively affect rankings, especially when the error is surfaced through the search console indexing reports.

Google Is Able To Handle Duplicate URLs

Google’s John Mueller answered the question by explaining that multiple URLs pointing to the same content do not trigger a penalty or loss of search visibility. He also noted that this kind of duplication is common across the web, implying that Google’s systems are experienced with handling this kind of problem.

He explained:

“It’s fine, but you’re making it harder on yourself (Google will pick one to keep, but you might have preferences).

There’s no penalty or ranking demotion if you have multiple URLs going to the same content, almost all sites have it in variations. A lot of technical SEO is basically search-engine whispering, being consistent with hints, and monitoring to see that they get picked up.”

What Mueller is referring to is Google’s ability to canonicalize a single URL as the one that’s representative of the various similar URLs. As Mueller said, multiple URLs for essentially the same content is a frequent issue on the web.

Google’s documentation lists five reasons duplicate content happens:

  1. “Region variants: for example, a piece of content for the USA and the UK, accessible from different URLs, but essentially the same content in the same language
  2. Device variants: for example, a page with both a mobile and a desktop version
  3. Protocol variants: for example, the HTTP and HTTPS versions of a site
  4. Site functions: for example, the results of sorting and filtering functions of a category page
  5. Accidental variants: for example, the demo version of the site is accidentally left accessible to crawlers”

The point is that duplicate content is something that happens often on the the web and is something that Google is able to handles.

Technical SEO Signals

Mueller said Google will pick one URL to keep, but added that the site owner might have preferences. That means Google will canonicalize the duplicates on its own, but the site owner or SEO can still signal which URL is the best choice (the canonical one) for ranking in the search results.

That is where technical SEO comes in. Internal linking, redirects, the proper use of rel=”canonical”, sitemap consistency, and consistency in 301 redirects all work as hints that help Google identify on the version you actually want indexed.

The Real Problem Is Mixed Signals

Mueller’s remark about making it harder on yourself was about the site owner/SEO spending time requesting URLs to be recrawled and noting that Google will figure it out on its own. But then he also referenced preferences, which alluded to all the signals I previously mentioned, in particular the rel=”canonical”.

Technical SEO Is Often About Reinforcing Preferences

Mueller’s description of technical SEO as “search-engine whispering” is useful because it captures how much of SEO involves reinforcing your preferences for what URLs are crawled, which content is chosen to rank, and indicating which pages of a website are the most important. Google may still choose a canonical on its own, but consistent signals increase the chance that it chooses the version the site owner wants.

That makes this a good example of what SEO is all about: Making it easy for Google to crawl, index, and understand the content. That’s really the essence of SEO. It is about being clear and consistent in the content, URLs, internal linking, overall site navigation, and even in showing the cleanest HTML, including semantic HTML (which makes it easier for Google to annotate a web page).

Semantic HTML can be used to clearly identify the main content of a web page. It can directly help Google zero in on what’s called the Centerpiece content, which is likely used for Google’s Centerpiece Annotation. The centerpiece annotation is a summary of the main topic of the web page.

Google’s canonicalization documentation explains:

“When Google indexes a page, it determines the primary content (or centerpiece) of each page. If Google finds multiple pages that seem to be the same or the primary content very similar, it chooses the page that, based on the factors (or signals) the indexing process collected, is objectively the most complete and useful for search users, and marks it as canonical. The canonical page will be crawled most regularly; duplicates are crawled less frequently in order to reduce the crawling load on sites.”

Technical SEO And Being Consistent

Stepping back to take a forest level view, duplicate URLs are really about a website not being consistent. Being consistent is not often seen as having to do with SEO but it actually is, on a general level. Every time I have created a new website I always had a plan for how to make it consistent, from the URLs to the topics, and also how to be able to expand that in a consistent manner as the website grows to cover more topics, to build that in.

Takeaways

  • Multiple URLs to the same content do not cause a penalty or ranking demotion
  • Google will usually pick one version to keep
  • Site owners can influence that choice through consistent technical signals
  • The real issue is mixed signals, not duplicate content itself
  • Technical SEO often comes down to reinforcing clear preferences and monitoring whether Google picks them up
  • The forest-level view of SEO can be seen as being consistent

Featured Image by Shutterstock/Andrey_Kuzmin

How Consumers Navigate High-Stakes Purchases In AI Mode via @sejournal, @Kevin_Indig

Boost your skills with Growth Memo’s weekly expert insights. Subscribe for free!

AI Mode is compressing the stage where buyers compare, reject, and discover brands on their own. Our new usability study of 185 documented purchase tasks shows that 74% of AI Mode final shortlists came directly from the AI’s output – no external check, no triangulation, no second opinion.

This analysis will cover:

  • How the comparison search phase has collapsed.
  • What this means for brands competing in categories with high competitor AI Mode saturation.
  • The three levers that determine whether your brand shows up.

Why We Conducted The Study

AI transforms Search from a list of results to a list of recommendations (shortlist). Until now, we have no idea how users treat AI shortlists. Do they take it at face value or thoroughly validate it?

That’s why I partnered with Citation Labs and Clickstream Solutions to record real users and their interactions when facing high-stakes purchases. This usability study of 48 participants completing 185 major-purchase tasks reveals that AI Mode operates as a recommendation environment, not a comparison one.

In traditional search, people click through results, comparing across sources to assemble a candidate set. In AI Mode, they accept the AI’s candidates and move on. 74% of AI Mode shortlists came directly from the AI’s output with no external check. In traditional search, more than half of users built their own shortlist from scratch.

The study covers four categories (televisions, laptops, washer/dryer sets, and car insurance). Participants completed tasks using both AI Mode and traditional search in a within-subjects A/B design, producing 149 AI Mode task observations and 36 search observations. The behavioral patterns are consistent enough across categories and participants to carry weight. (Full study design is at the end.)

From Garret French, founder of Citation Labs:

“In AI Mode, buyers often use a shortlist synthesis to shortcut the cognitive effort of Standard Searching and comparing. This raises the value of onsite decision assets and third-party sources that provide AI with clear trade-offs, specific evidence, and sufficient contextual structure to describe a brand’s offering with confidence.”

From Eric Van Buskirk:

The absence of narrowness frustration is the most intellectually significant finding. 15% in AI Mode vs 11% in Search, with no meaningful statistical difference. That’s the finding that rules out the obvious alternative explanation: that users accepted the AI’s shortlist because they felt trapped. They didn’t push back. They weren’t frustrated. They were satisfied. That makes the acceptance harder to dismiss.

Here’s what happened.

1. 88% Of Users Took The AI’s Shortlist Outright

Across the laptop and insurance tasks, where participants used both search surfaces (classic search and AI Mode), the gap in constructing a product shortlist was stark.

Image Credit: Kevin Indig

Definitions:

  • AI Adopted: The participant took the AI’s recommended candidates as their shortlist with no changes or external verification.
  • User Built: The participant ignored the AI’s (or Search’s) suggestions and assembled their own candidate list from independent sources.
  • AI Verified: The participant started with the AI’s candidates but checked them against an outside source (a retailer site, a review, a manufacturer page) before finalizing.
  • Hybrid: The participant combined AI-suggested candidates with at least one candidate they found independently.

In classic search, 56% of participants built their own shortlist from multiple sources. In AI Mode, only 8 out of 147 codeable tasks produced a genuinely self-built shortlist. The user’s comparison process didn’t just shrink when using AI Mode. For most participants, it didn’t happen at all.

64% of AI Mode participants clicked nothing at all during their task. They read the AI’s text, sometimes scrolled through inline product snippets, and declared their finalists. The no-click rate varied by category:

Image Credit: Kevin Indig

Insurance participants delegated most heavily. Washer/dryer participants clicked the most, likely because appliance decisions involve specific physical constraints (capacity, stacking compatibility, dimensions) that the AI summary didn’t always resolve.

The 36% who did interact with individual results within AI Mode broke into 2 groups:

  • About 15% of the AI Adopted group (17 of 117 participants) verified inside AI Mode: They opened inline product cards or merchant pop-ups to check a price or spec, then returned to the AI’s list.
  • Others used follow-up prompts as verification tools, asking the AI for prices or narrowing by constraints.

A separate 23% of all AI Mode tasks involved at least one visit to an external website, mostly retailers (Best Buy appeared in 10 of 34 tasks with external visits) and manufacturer sites. The destination pattern matters: Users left AI Mode to confirm a candidate they’d already accepted from the AI’s list, not to find new ones.

Of the 117 participants who adopted the AI’s shortlist directly, roughly 85% showed no internal verification behavior at all. Participants who built their own lists took an average of 89 seconds longer and consulted more than twice as many sources.

  • “Given that the first paragraph says Lenovo or Apple… going with that,” said one user about laptops when searching via AI Mode. Position one in the AI response was the entire decision.
  • Another AI Mode user remarked: “I liked it more than anything else I’ve ever used for product searching. It made it a lot quicker to find the options.” They experienced speed as a valuable feature, not a shortcut.

In classic search, the pattern reversed. Nearly 89% of participants clicked on something.

  • One insurance participant clicked out to Progressive and GEICO independently, read both landing pages, consulted an Experian article, and then arrived at a shortlist.
  • A laptop participant applied hardware filters and flagged a review score discrepancy: “It shows 4.6 out of 5 stars for the reviews, but when you actually click the link: not reviewed yet.” Active skepticism of aggregated data was a behavior absent from AI Mode transcripts.

2. The AI’s Top Pick Becomes The User’s Top Pick 74% Of The Time

Just like in classic search, the top answer carries outsized weight. 74% of participants chose the item ranked first in the AI’s response as their top pick. The mean rank of the final choice was 1.35. Only 10% chose something ranked third or lower.

Image Credit: Kevin Indig

Position one in the AI’s output carries an outsized advantage because of where it sits: inside a curated section that typically contains two to five items, after the AI has already done the filtering. The first item is the AI’s top pick. When people engage with AI mode, we know they read almost all of the output: The first AI Mode study found users spend 50 to 80 seconds reading AI Mode output, more than double the dwell time on AI Overviews. Users are reading carefully. They just read within a set the AI already narrowed.

However, 26% of participants in this study overrode rank order. The driver: brand recognition. They spotted a brand lower on the list and preferred it regardless of where the AI placed it. TV and laptop categories saw this most, where participants arrived with existing preferences for Samsung, LG, Apple, or Lenovo. But overriding rank did not mean rejecting the AI’s output: 81% of rank-override participants still chose from the AI’s candidate set.

3. The AI’s Words Become The Trust Signal

“Travelers and USAA actually tell me how much, whereas State Farm and GEICO give percentages. Just knowing the exact amount makes me want to pick Travelers or USAA right off the bat.”

That quote captures a core pattern in AI Mode trust. The AI’s formatting shaped the decision: Dollar amounts versus percentage discounts determined which brands made the shortlist.

AI framing (37%), meaning how AI talks about the product, and brand recognition (34%) were the top 2 trust drivers in AI Mode. They run nearly even:

  • Brand recognition led when participants arrived with brand preferences.
  • AI’s wording filled the gaps where participants didn’t already have preferences.
Image Credit: Kevin Indig

In classic search, the dominant trust mechanism was multi-source convergence: Participants built confidence by checking whether multiple independent sources agreed about a product.

Essentially, users triangulated. One checked Progressive, then GEICO, then an Experian article. Another compared aggregated star ratings against reviews on the actual site. They were building a case from separate inputs.

That behavior was almost absent in AI Mode (5%). Instead, AI framing (how the AI worded its description of a product) and brand recognition were the top 2 trust drivers.

The split between these two signals tracked closely with product category:

Image Credit: Kevin Indig

For televisions and laptops, where most participants arrived with existing brand preferences, brand recognition dominated. For insurance and washer/dryer, where participants had less prior knowledge, AI framing dominated.

When you lack a prior view, the AI’s description becomes the trust signal. In AI Mode, the synthesis is the corroboration. Participants treated the AI’s summary as if the cross-checking had already been done for them.

The first study showed a related pattern from the supply side: AI Mode matches site type to intent, surfacing brands for transactional queries and review sites for comparisons. This study shows the demand side of the same behavior: When the AI surfaces a brand the user already knows, brand recognition drives the decision; when it doesn’t, the AI’s own framing fills that role. The site-type matching and the trust mechanism reinforce each other.

4. If You’re Not In The List, You Don’t Exist

Purchase outcomes in AI Mode concentrated heavily. For laptops, three brands captured 93% of all AI Mode final choices. In classic search, the distribution was broader: HP EliteBook variants appeared three times, ASUS once, and other brands got consideration they never received in AI Mode.

Image Credit: Kevin Indig

Two distinct problems emerged:

  1. Brands that never appeared in the AI’s output were never considered. Participants didn’t see them, so they couldn’t evaluate them. The AI decided who made the list, not the buyer.
  2. Brands that did appear but lacked recognition faced a different problem: They weren’t seriously considered. Erie Insurance showed up in AI Mode results, but multiple participants eliminated it on name recognition alone. The brand was present but hadn’t built enough awareness to survive the moment of selection. One participant dropped a brand because it lacked a hyperlink in the AI output, reading that formatting gap as a credibility signal: “There’s not even a link there.”

Another participant said when using AI Mode: “I’m already eager to believe these are good recommendations because it mentions LG and Samsung, two brands I consider very reliable.” The AI didn’t say those brands were better. The participant inferred it from familiarity.

Participants didn’t feel constrained by the narrower set. Narrowness frustration appeared in 15% of AI Mode tasks and 11% of classic search tasks, statistically indistinguishable. The option set shrank, but the feeling of having enough options didn’t change. The most skeptical AI Mode participant in the comparison set, who complained the AI kept pointing to “teen drivers, teen drivers, teen drivers,” still chose GEICO and Travelers: the consensus AI result.

5. Users Leave To Buy, Not To Research

23% of AI Mode tasks involved an external site visit, but keep in mind these prompts reflect high-stakes situations. In standard search, that figure was 67%.

Image Credit: Kevin Indig

The volume difference matters less than the intent difference:

  • AI Mode participants who left went to retailer sites and manufacturer pages to verify a price or spec for a candidate they’d already selected.
  • Standard Search participants left to discover candidates: Reddit for peer opinions, editorial review sites for expert takes, insurance aggregators for comparison.

In the first AI Overviews study, we found that high risk leads users to verify AI claims more and reference against answers from other users on UGC platforms (like Reddit).

In this study, Reddit appeared in 19% of standard search tasks and only twice across all 149 AI Mode sessions. The peer-opinion layer that shapes a large share of traditional Search barely exists in AI Mode behavior.

There’s irony in that pattern. Google leans heavily on Reddit content to train its models. However, the source that users rely on most in standard search is the one they almost never visit when the AI synthesizes those same sources for them.

The first study found the same pattern at a different scale. Across 250 sessions, clicks were “reserved for transactions:” Shopping prompts drove the highest exit share, while comparison prompts drove the lowest. The exit destinations were retailers and brand sites, not editorial or peer-opinion sources. Six months and a different task set later, the pattern holds: When users leave AI Mode, they leave to buy.

6. 3 Levers: Visibility, Framing, And Pricing Data

Three things that excite me most about the study:

First, we can apply the mental model of rankings (higher = better) to AI Mode as well. Most users choose the first product. Now, we can apply this to prompt tracking by focusing more on prompts that lead to shortlists and use our position as a goalpost.

Second, trust trumps rank. We know this since the first user behavior studies I published, but this study reinforces the importance of building trust with users before they search. It’s the ultimate cheat code.

Third, we now know buyers trust AI’s recommendations. Obviously, there’s a high risk here if the AI is wrong, but seeing how quickly buyers take the AI’s recommendation also shows us how fast consumers adopt AI. It truly is the future of Search.

Keep in mind:

1. Visibility at the model layer is the new threshold. If AI Mode doesn’t surface your brand, you have a visibility problem at the model layer. Query your own category the way a buyer would (i.e., “best car insurance for a family with a teen driver,” “best washer dryer set under $2,000”) and document which brands appear, in what order, and with what framing. Do this across multiple prompt variations. Do it regularly, because AI responses shift over time.

2. How the AI describes you matters as much as whether it appears. Brands cited with concrete attributes (specific model, specific price, named use case) held stronger positions than brands described generically. The content on your site that the AI draws from not only affects whether you show up, but also how confidently and specifically you show up. A brand with structured pricing data, clear product specs, and explicit use cases gives the AI better material to work with.

3. For categories with context-dependent pricing, AI Mode creates a false-confidence problem. 63% of insurance participants were rated overconfident about pricing. They accepted AI-quoted rate estimates without checking whether the figures applied to their actual state, driving record, or current insurer. They made elimination decisions based on numbers that may not have applied to them. Where shopping panels showed explicit retailer-confirmed prices (washer/dryer), 85% of participants understood pricing clearly. Where they didn’t (insurance, laptops), confusion and overconfidence filled the gap. Structured pricing data through Merchant Center feeds and schema markup is the most direct lever for brands selling physical products. For services, the lever is editorial: Make sure your landing pages and FAQ content frame pricing as conditional (“your rate depends on X, Y, Z”) so the AI has that framing to draw from.

Study Design

Citation Labs and Clickstream Solutions ran this as a remote, unmoderated usability study with 48 U.S.-based participants recruited through Prolific. Each participant completed up to four major-purchase shortlisting tasks across televisions, laptops, washer/dryer sets, and car insurance.

The comparison between AI Mode and traditional standard search used a within-subjects A/B design: Participants used both surfaces, not one or the other. Significance calculations were normalized for the exact number of participants in each group (149 AI Mode task observations, 36 standard search task observations). This matters because the groups are unequal in size, and raw percentage comparisons between them would overstate confidence without that correction.

Sessions were screen-recorded with think-aloud audio. Trained analysts annotated each recording for behavioral markers (click-through, shortlist origin, trust signals, external site visits) and qualitative markers (stated reasoning, brand mentions, frustration signals). The 185 task-level observations provide a larger analytical base than the 48-participant headcount suggests, but confidence intervals remain wider than a large-scale survey. Findings are directional, not population-level estimates.

Notes on terminology used throughout this report:

  • Shortlist: The final set of brands a user would consider buying from.
  • AI Adopted: The participant took the AI’s recommended candidates as their shortlist with no changes or external verification.
  • User Built: The participant ignored the AI’s (or Search’s) suggestions and assembled their own candidate list from independent sources. In Search, when there was no AIO present, they had no option for relying on AI suggestions.
  • AI Verified: The participant started with the AI’s candidates but checked them against an outside source (a retailer site, a review, a manufacturer page, further prompting, or interaction with a panel outside the main AI text block ) before finalizing.
  • Hybrid: The participant combined AI-suggested candidates with at least one candidate they found independently.
  • AI framing: The specific words and structure the AI used to describe a product, such as labels like “best for affordability” or explicit price comparisons.
  • Brand recognition: The user chose or eliminated a brand based on prior familiarity, not the AI’s description or any external research.
  • AI trust (general): The user accepted the AI’s output as credible without citing a specific reason, such as a particular label or description.
  • Source trust: The user trusted a recommendation because of where it came from, such as a retailer, manufacturer, or named publication surfaced in results.
  • Multi-source convergence: The user built confidence by checking whether multiple independent sources agreed on the same recommendation.
  • Rank override rate: The share of users who chose a brand other than the AI’s top-ranked option, regardless of whether they stayed within the AI’s candidate list.

Featured Image: Tapati Rinchumrus/Shutterstock; Paulo Bobita/Search Engine Journal

Google Explains Why It Doesn’t Matter That Websites Are Getting Larger via @sejournal, @martinibuster

A recent podcast by Google called attention to the fact that websites are getting larger than ever before. Google’s Gary Illyes and Martin Splitt explained that the idea that websites are getting “larger” is a bad thing is not necessarily true. The takeaway for publishers and SEOs is that Page Weight is not a trustworthy metric because the cause of the “excess” weight might very well be something useful.

Page Size Depends On What ‘s Being Measured

Google’s Martin Splitt explained that what many people think of as page size depends on what is being measured.

  • Is it measured by just the HTML?
  • Or are you talking about total page size, including images, CSS, and JavaScript?

It’s an important distinction. For example, many SEOs were freaked out when they heard that Googlebot was limiting their page crawl to just 2 megabytes of HTML per page. To put that into perspective, two megabytes of HTML equals about two million characters (letters, numbers, and symbols). That’s the equivalent of one HTML page with the same number of letters as two Harry Potter books.

But when you include CSS, images, and JavaScript along with the HTML, now we’re having a different conversation that’s related to page speed for users, not for the Googlebot crawler.

Martin discussed an article on HTTPArchive’s Web Almanac, which is a roundup of website trends. The article appeared to be mixing up different kinds of page weight, and that makes it confusing because there are at least two versions of page weight.

He noted:

“See that’s where I’m not so clear about their definition of page weight.

…they have a paragraph where they are trying to like explain what they mean by page weight. …I don’t understand the differences in what these things are. So they say page weight (also called page size) is the total volume of data measured in kilobytes or megabytes that a user must download to view a specific page. In my book that includes images and whatnot because I have to download that to see.

And that’s why I was surprised to hear that in 2015 that was 845 kilobytes. That to me was surprising. …Because I would have assumed that with images it would be more than 800 kilobytes.

… In July 2025, the same median page is now 2.3 megabytes.”

Data Gets Compressed

But that is only one way to understand page size. Another way to consider page size is by focusing on what is transferred over the network, which can be smaller due to compression. Compression is an algorithm on the server side that minimizes the size of the file that is sent from the server and downloaded by the browser. Most servers use a compression algorithm called Brotli.

Martin Splitt explains:

“I ask this question publicly that different people had very different notions of how they understood page size. Depending on the layer you are looking at, it gets confusing as well
because there’s also compression.

…So some people are like, ah, but this website downloads 10 megabytes onto my disk.

And I’m like, yes. …but maybe if you look at what actually goes over the wire, you might find that this is five or six megabytes, not the whole 10 megabytes. Because you can compress things on the network level and then you decompress them on the client side level…”

Technically, the page size in Martin’s example is actually five or six megabytes because of compression, and it’s able to download faster. But on the user’s side, that five or six megabytes gets decompressed, and it turns back into ten megabytes, which occupies that much space on a user’s phone, desktop, or wherever.

And that introduces an ambiguity. Is your web page ten megabytes or five megabytes?

That illustrates a wider problem: different people are talking about different things when they talk about page size.

Even widely used definitions don’t fully resolve the ambiguity. Page weight is described as “the total volume of data measured in kilobytes or megabytes that a user must download,” but as the discussion makes clear, there is no one clear definition.

Martin asserts:

“When you ask people what they think, if this is big or not, you start getting very different answers depending on how they think about page size. And there is no one true definition of it.”

What About Ratio Of Markup To Content?

One of the most interesting distinctions made in the podcast is that a large page is not necessarily inefficient. For example, a 15 MB HTML document is considered acceptable because “pretty much most of these 15 megabytes are actually useful content.” The size reflects the value being delivered.

By contrast, what if the ratio of content to markup were the other way around, where there was a little bit of content but the overwhelming amount of the page weight was markup.

Martin discussed the ratio example:

“…what if the markup is the only overhead? And I mean like what do you mean? It’s like, well, you know, if it’s like five megabytes but it’s only very little content, is that bad? Is that worse as in this case, the 15 megabytes.

And I’m like, that’s tricky because then we come into this weird territory of the ratio between content and markup. Yeah.

And I said, well, but what if a lot of it is markup that is metadata for some third party tool or for some service or for regulatory reasons or licensing reasons or whatever. Then that’s useful content, but not necessarily for the end user, but you still kind of have to have it.

It would be weird to say that that is worse than the page where the weight is mostly content.”

What Martin is doing here is shifting the idea of page weight away from raw size toward what the data actually represents.

Why Pages Include Data Users Never See

A major contributor to page weight is content that users never see.

Gary Illyes points to structured data as an example of content that is specifically meant for machines and not for users. While it can be useful for search engines, it also adds to the overall size of the page. If a publisher adds a lot of structured data to their page in order to take advantage of all the different options that are available, that’s going to add to the page size even though the user will never see it.

This calls attention to a structural reality of the web: pages are not just built for human readers. They are also built for search engines, tools, AI agents, and other systems, all of which add their own requirements to the weight of a web page.

When Overhead Is Justified

Not all non-user-facing content is unnecessary.

Martin talked about how markup may include “metadata” or a tool, regulatory, or licensing purpose, creating a kind of gray area. Even if the additional data does not improve the user experience directly, it does serve a purpose, including helping the user find the page through a search engine.

The point that Martin was getting at is that these considerations of page weight complicate attempts to label page weight as good if it is under this threshold or bad if the page weight exceeds it.

Why Separating Content and Metadata Doesn’t Work

One possible solution that Gary Illyes discussed is separating human-facing content from machine-facing data. While Gary didn’t specifically mention the LLMs.txt proposal, what he discussed kind of resembles it in that it serves content to a machine minus all the other overhead that goes with the user-facing content.

What he actually discussed was a way to separate all of the machine-facing data from what the user will download, thus, in theory, making the user’s version of a web page smaller.

Gary quickly dismisses that idea as “utopic” because there will always be hordes of spammers who will find a way to take advantage of that.

He explained:

“But then unfortunately this is an utopic thing. Because not everyone on the internet is playing nice.

We know how much spam we have to deal with. On our blog we say somewhere that we catch like 40 billion URLs per day that’s spam or some insane number, I don’t remember exactly, but it’s some insane number and definitely billions. That will just exacerbate the amount of spam that search engines receive and other machines receive maybe like I would bet $1 and 5 cents that will actually increase the amount of spam that search engines and LLMs and others ingest.”

Gary also said that Google’s experience is that, historically, when you have separate kinds of content, there will always be differences between the two kinds. He used the example of when websites had mobile and desktop pages, where the two versions of content were generally different, which in turn caused issues for search and also for usability when a site ranks a web page for content on one version of a page, then sends the user to a different version of the page where that content does not exist.

Although he didn’t explicitly mention it, that explanation of Google’s experience may shed more light on why Google will not adopt LLMS.txt.

As a result, search engines have largely settled on a single-document model, even if it is inefficient.

Website Size vs Page Size Is the Real World

The discussion ultimately challenges the original concept of the problem, that heavy web pages are bad.

Gary observes:

“The first question is, are websites getting fat? I think this question is not even meaningful.

Because it does not matter in the context of a website if it’s fat. In the context of a single page, yes.

But in the context of a website, it really doesn’t matter.”

So now Gary and Martin change the focus to web pages that are getting heavier, a more meaningful way to look at the issue of how web pages and websites are evolving.

This moves the discussion from an abstract idea to something more measurable and actionable.

Heavier Pages Still Carry Real Costs

Even with faster connections and better infrastructure, larger pages still have consequences, and smaller weighted pages have positive benefits.

Martin explains:

“I think we are wasting a lot of resources. And I mean we, we had that in another episode where we said that we know that there are studies that show that websites that are faster have better retention and better conversion rates. Yeah. And speed is in part also based on size. Because the more data I ship, the longer it takes for the network to actually transfer that data and the longer it takes for the processor of whatever device you’re on to actually process it and display it to you.”

From a broader perspective, the issue is not just performance but efficiency. As Illyes puts it, “we are wasting a lot of resources.”

The web may be getting heavier, but the more important takeaway is why. Pages are carrying more than just user-facing content, and that design choice shapes both their size and their impact.

Featured Image by Shutterstock/May_Chanikran

Google’s Mueller On SEO Gurus Who Are “Clueless Imposters” via @sejournal, @martinibuster

A search marketing professional from India wrote a blog post about how she feels about seeing the word guru used within the SEO community in a way that’s different from its meaning in India. Several people, including Google’s John Mueller, agreed with her and shared how they felt when people self-identify as SEO gurus.

The Word Guru Is Misused

Preeti Gupta wrote a blog post titled, I don’t like how the word ‘Guru’ is misused in the SEO industry, in which she shared what the word guru actually means and how it’s misused in the SEO industry in a way that trivializes a word that in India holds special meaning.

She wrote that in India the word guru has a deep meaning and that they hold great respect for actual gurus.

Her blog post shared a Sanskrit mantra about it:

“The Guru is like Brahma (the creator). They create the desire for knowledge.
The Guru is like Vishnu (The preserver). They help the student keep and use the knowledge.
The Guru is like Maheshwara (Shiva, the Destroyer). They destroy ignorance and bad habits.
The Guru is the supreme reality itself, standing right before your eyes.
I bow and offer my respects to that great teacher.”

She then contrasted that profound meaning of the word guru with the trivialization of it within the context of self-described SEO gurus, who she regards as shady types who engage in unethical SEO practices. She said that it’s not her intention to tell people what words to use, but she did express the hope that people would use the word in the right context.

The phrase SEO guru is used in both contexts, as a derogatory phrase to paint someone as a false leader with naïve followers and also as someone who is highly regarded. However, I think an argument can be made that using that phrase for oneself is immodest, self-aggrandizing, and simply isn’t a good look.

AlexHarford-TechSEO responded to her post on Bluesky:

“It puts me off when I see an SEO self-describe themselves as a “Guru.” I’ve never come across anyone who does so who is a good and ethical SEO.

A lot of words are losing meaning in today’s world, though there can’t be many that were as special to you as Guru.”

Words are always in a state of change, and the way people speak not only changes from region to region but also from decade to decade. The meaning of words does change, especially when they jump continents and languages.

Self-Declared SEO Gurus

It was at this point that John Mueller responded to share what he thinks about self-described SEO gurus:

“To me, when someone self-declares themselves as an SEO guru, it’s an extremely obvious sign that they’re a clueless imposter. SEO is not belief-based, nobody knows everything, and it changes over time. You have to acknowledge that you were wrong at times, learn, and practice more.”

Mueller is right that nobody knows everything and that SEO changes over time, and for a long time many SEOs didn’t keep up with how Google ranks websites. The industry has largely shed that naivete, and yet nobody really agrees on what to do to rank better in search engines and AI search.

SEO Is A Belief System

Although I know there are some SEOs who firmly believe that SEO is a set of universally agreed upon practices and that that is all there is, unaware that the history of SEO is one of constant change. How SEO is practiced today is quite different from how it was practiced eight years ago. There is no set of practices to be agreed on except Google’s best practices, which are less about do this and you will rank better and more about do this and you may have a chance to rank better.

So yes, to a certain extent, SEO is a belief system and will continue to be a belief system so long as Google’s search ranking algorithms remain a black box algorithm that people can see what goes in and what comes out but not what happens in the middle. That part remains a mystery. So when you don’t know for sure that what you do will guarantee better rankings, the only thing left is to believe, hope, and even have faith that the rankings will happen. Faith, after all, is belief in something that does not provide definitive proof. You don’t need faith to believe in a fact, right?

And that last part, the mystery of what happens in the black box, is why nobody can really call themselves a guru in the sense of being all-knowing. Nobody outside of Google knows everything that’s going on within that part in the middle where the rankings “magic” happens.

Given all that, who can truly call themselves a guru in SEO?

Featured Image by Shutterstock/funstarts33

The Top 6 Search Engines Market Share & The AI Search Engines To Watch via @sejournal, @MattGSouthern

For more than a decade, Google’s share of global search traffic barely moved. It hovered between 91% and 93%, and the SEO industry built its workflow around that reality.

As of March 2026, StatCounter data shows Google at 90.01% of worldwide search traffic, with Bing at 4.98%, Yahoo at 1.39%, Yandex at 1.34%, DuckDuckGo at 0.76%, and Baidu at 0.55%.

Google’s share has fluctuated between roughly 89% and 93% since 2015. It dipped below 90% in each of the final three months of 2024 and again in February 2026 before ticking back above that line in March. At Google’s scale, every tenth of a percent represents millions of searches.

Meanwhile, AI search tools like ChatGPT and Perplexity are growing fast outside of traditional measurement, and platforms like Amazon and TikTok continue to capture search behavior that never touches a traditional engine.

Here we will walk through the top six search engines that account for the measurable search market. I’ll also explain where AI search engines fits into the picture and what all of this means for where you invest resources.

1. Google

Google still leads by a wide margin, with nine out of every 10 searches worldwide running through its engine, according to StatCounter. Alphabet, Google’s parent company, generates most of its revenue from search advertising and continues to invest in AI-powered features.

Google search, March 2026.

The numbers look different by market and device. In the United States, Google holds 84.13%, with Bing at 10.52% and Yahoo at 2.86%. On desktop globally, Google’s share drops to roughly 82%, while Bing picks up over 10%. Mobile is still Google’s strongest position at over 94%.

The biggest change to Google’s search results over the past year has been the expansion of AI Overviews. These AI-generated summaries appear directly on the results page and answer the query before a user clicks anything. Several studies suggest rising rates of zero-click searches on Google. For SEO professionals, this means competition for clicks is intensifying even as Google’s overall share holds steady.

The challenge is familiar to anyone who has optimized for Google over the past few years. The traffic potential is still unmatched, but earning clicks means competing against Google’s own SERP features. Featured Snippets, People Also Ask boxes, local packs, shopping carousels, and AI-generated summaries all sit above or alongside organic results. Paid search costs continue to rise in competitive verticals. And even ranking in the top three organic positions doesn’t guarantee a click the way it once did.

None of that changes the math. Google is still where the traffic is. If you are going to invest in one search engine, it should be this one. But the cost of competing keeps rising, and the share of clicks reaching external sites is being squeezed by more on-SERP answers. Diversification is worth serious consideration.

→ Further reading:

2. Microsoft Bing

Bing has posted the most notable gains in market share among traditional search engines over the past two years. Its 5.01% global share may look small, but in the U.S., it climbs past 10%, and on desktop it’s even higher.

Bing search with Copilot integration, March 2026.

Microsoft has paired Bing’s existing distribution advantages with Copilot integration, adding a conversational AI layer alongside traditional search results.

The volume gap between Google and Bing is obvious. But that gap also means fewer brands competing for the same positions, which is worth considering when you are deciding where to allocate budget.

There’s a strategic reason beyond market share to pay attention to Bing. ChatGPT’s search functionality relies on Bing’s index for retrieving web results. If that connection holds, content that performs well in Bing has an additional path to being surfaced through AI-generated answers in ChatGPT.

Optimizing for Bing overlaps with Google in many areas. The Bing Webmaster Guidelines cover familiar ground, such as content quality, backlinks, and page speed. Where SEO professionals who work across both engines tend to notice differences is in how each engine weighs keyword specificity, how each handles mobile versus desktop indexing, and how social signals factor into results. If you are optimizing for Google already, much of that work carries over to Bing, but reviewing Bing’s own guidelines is worth the time to catch the gaps.

For teams already running Google Ads, testing Microsoft Ads is a low-friction starting point. Microsoft Ads supports direct campaign imports from Google Ads, which reduces the setup time. From there, you can evaluate performance independently and adjust bids and targeting for Bing’s audience.

→ Further reading:

3. Yahoo

Yahoo holds 1.39% of the global search market share and 2.86% in the U.S.

Yahoo’s search results are powered by Bing through a long-standing partnership. The index and ranking algorithms are the same, where Yahoo differs is in its ecosystem. Yahoo Mail, Yahoo Finance, and Yahoo News still draw audiences, and those users often search within the Yahoo environment rather than switching to Google.

Yahoo search, March 2026.

If you are optimizing for Bing, you are also reaching Yahoo’s audience. The combined Bing-Yahoo share in the U.S. is over 13%, which represents a pool of traffic that most SEO strategies ignore. There’s no separate Yahoo optimization playbook. The work you do for Bing carries over.

Where Yahoo becomes strategically relevant is in paid search. Microsoft Ads campaigns serve across both Bing and Yahoo properties. This means a single paid search setup gives you access to the combined audience without additional management overhead.

→ Further reading: New Yahoo Scout AI Search Delivers The Classic Search Flavor People Miss

4. Yandex

Yandex holds 1.34% of the global search market share, but that number undersells its importance in one region. In Russia, Yandex powers roughly 72% of all searches, according to StatCounter. If your business operates in or sells to the Russian market, Yandex is the primary search engine, not an alternative.

Yandex search, March 2026.

A leak of approximately 44 GB of Yandex source code in 2023 exposed over 17,800 ranking factors. The leak gave the SEO community an unprecedented look at how a search engine’s algorithms work. Yandex’s algorithms differ from Google’s, but the leak provided useful insights into ranking signal categories that many SEO professionals have incorporated into their broader thinking.

The leaked code confirmed that Yandex’s algorithms place greater weight on geolocation than Google’s, making local optimization especially important. Domain age, content freshness, and user behavior signals like click-through rate and dwell time all carried weight in the leaked ranking factors. Because of lower competition, paid search on Yandex can be less expensive per click than on Google or Bing. But success requires native-language content and cultural fluency that goes well beyond translation.

Yandex also offers its own webmaster tools, advertising platform (Yandex Direct), and analytics suite (Yandex Metrica), all designed for Russian-language users. Most English-language SEO teams can safely skip Yandex, but international teams working in Russian-speaking markets can’t.

→ Further reading: Yandex Search Ranking Factors Leak: Insights

5. DuckDuckGo

DuckDuckGo holds 0.76% of the global search market share and 1.84% in the U.S. Its growth has been steady rather than explosive, driven by users who prioritize privacy.

DuckDuckGo search, March 2026.

Unlike Google, DuckDuckGo doesn’t track users or build advertising profiles based on search history. This privacy-first positioning has attracted a consistent user base. In Europe, where data privacy regulations like GDPR have raised awareness of tracking practices, DuckDuckGo and similar engines like Ecosia have picked up small but steady gains.

DuckDuckGo’s search results draw from multiple sources, including Bing’s index and its own web crawler. There’s no need to optimize for DuckDuckGo separately. Strong performance on Bing generally translates to visibility here as well.

DuckDuckGo does offer its own advertising platform, which serves keyword-based ads through a partnership with Microsoft Advertising. Brands in privacy-sensitive verticals like healthcare, financial services, or cybersecurity may find that visibility on a privacy-focused engine resonates with their audience.

→ Further reading:

6. Baidu

Baidu holds 0.55% of the global search market share, a number that reflects its almost exclusively Chinese user base rather than a lack of scale. Within China, Baidu commands over 53% of the search market according to StatCounter and processes billions of searches daily.

Baidu search, March 2026.

Baidu has invested in AI through its ERNIE large language model series. The company’s AI assistant reached 200 million monthly active users by January 2026, according to the South China Morning Post, and Baidu unveiled its latest model, ERNIE 5.0, in November 2025. The company also made ERNIE Bot free for individual users starting April 2025. Baidu is integrating AI-generated answers into search results in ways similar to Google’s AI Overviews.

For businesses targeting Chinese consumers, Baidu isn’t optional. But the optimization work looks different than what you do for Google. Content must be in Mandarin, hosted on servers accessible within China, and compliant with Chinese internet regulations. The algorithms also differ. SEO guides for the Chinese market typically emphasize domain age, meta tags, and page load speed as ranking factors that carry more weight on Baidu.

Baidu operates its own advertising platform (Baidu Tuiguang), webmaster tools, and analytics suite. Setting up these tools from outside China adds complexity. Most teams will need a partner or team member with direct experience in the Chinese market. Unless you have a specific China market strategy with native-language resources, Baidu is unlikely to be part of your SEO roadmap.

→ Further reading: Baidu Ranking Factors: A Comprehensive Data Study

AI Search Engines: ChatGPT Search and Perplexity

The market share numbers above don’t capture what may be the most important change in search behavior over the past two years. AI search engines like ChatGPT Search and Perplexity don’t appear in StatCounter’s data. They don’t function as traditional search referral sources that tracking scripts measure in the same way.

The usage numbers are growing fast. ChatGPT has reached 900 million weekly active users, OpenAI reported in late February 2026. That is up from 800 million in October 2025. Perplexity CEO Aravind Srinivas said at Bloomberg’s Tech Summit in June 2025 that the platform processed 780 million search queries in May 2025, up from 230 million less than a year earlier.

In terms of traffic sent to websites, AI platforms are still small. Conductor’s 2026 AEO/GEO Benchmarks Report found that AI referral traffic accounts for 1.08% of total web traffic across 10 industries it studied. ChatGPT drives 87.4% of that AI referral traffic. An SE Ranking study found that AI traffic to websites grew roughly sevenfold between early 2024 and mid-2025.

These tools work differently from traditional search engines. Instead of returning a list of links, they provide synthesized answers with source citations. Users can ask follow-up questions, refine their research, and get summaries without clicking through to individual pages.

For SEO professionals, this creates a new optimization challenge. Appearing as a cited source in ChatGPT or Perplexity answers is becoming a separate discipline from ranking in Google’s organic results. The SEO industry has started calling this “Generative Engine Optimization,” or GEO, though the field is still young and best practices are evolving.

Early data from Conductor’s benchmarks offers some initial signals about Google’s AI Overviews, specifically. In Conductor’s analysis of over 21 million Google searches in September 2025, 25.11% triggered an AI Overview, with rates varying widely by industry. The types of pages cited in those overviews also varied by sector. Whether those patterns hold across ChatGPT, Perplexity, and other AI platforms is less clear, which means tracking your visibility across multiple AI surfaces is becoming necessary.

The first step is to start monitoring whether your content appears in AI responses. Several SEO platforms have added AI visibility tracking tools. From there, the optimization work resembles what has always worked in SEO. Clear, authoritative content with well-organized data and transparent sourcing is the foundation.

For more on the growing AI search landscape, see our guide to the best AI search engines.

→ Further reading:

Search Beyond Traditional Engines

Amazon

Amazon doesn’t appear in StatCounter’s search engine rankings, but it’s one of the most important search platforms for anyone selling products online. Survey-based research has consistently found that a large share of online product searches begin on Amazon (56%, according to Jungle Scout) rather than on a traditional search engine.

Amazon’s search algorithm, originally known as A9 and widely referred to as A10 in the seller community, is built around purchase intent. Amazon has not officially confirmed an algorithm version change, but third-party sellers and agencies have documented differences from earlier versions, including greater weight on external traffic and seller authority.

For ecommerce brands, the key is balancing Amazon marketplace investment with direct-to-consumer search traffic. Amazon gives you access to buyers at the point of purchase, but the platform controls the customer relationship and takes a cut of every sale. Brands that invest only in Amazon search lose the ability to build direct relationships with their customers, while brands that ignore Amazon miss the audience that starts their product research there.

Amazon also offers its own paid advertising through Sponsored Products, Sponsored Brands, and Sponsored Display campaigns. These ads appear within Amazon’s search results and on product detail pages, and they run on a cost-per-click model. For product-focused businesses, Amazon advertising is a separate budget line from Google or Microsoft Ads, and it requires its own keyword research and bid management.

→ Further reading:

TikTok

TikTok has more than 170 million users in the United States and over 1.6 billion worldwide. The platform has become a search and discovery tool for younger audiences looking for recommendations, product reviews, and how-to content.

TikTok’s regulatory path in the United States went through a turbulent period. Congress passed legislation in 2024 requiring ByteDance to divest its U.S. operations or face a ban. TikTok briefly went dark in January 2025. Following multiple executive order extensions, a deal was finalized on January 22, 2026. The resulting TikTok USDS Joint Venture LLC put US operations under majority control of American investors, including Oracle and Silver Lake, with ByteDance retaining a 19.9% minority stake.

With the ownership question resolved for now, TikTok is still a growing discovery platform. Its recommendation algorithm surfaces content based on engagement signals rather than keywords, which means optimization for TikTok search looks nothing like traditional SEO. Content format, hook quality, and engagement metrics matter more than backlinks or domain authority.

For brands, TikTok’s search function is most relevant as a discovery channel. Users search TikTok for product recommendations, restaurant reviews, travel ideas, and tutorials. The content that performs well in TikTok search tends to be short, visually engaging, and structured around a specific question or problem. If your audience skews younger and your content lends itself to video, TikTok search is worth testing as a supplementary channel.

→ Further reading:

What Market Share Changes Mean for Your SEO Strategy

Google is still the leader of search, but its share is dropping while AI search tools are growing fast, and alternative platforms serve specific audiences that Google doesn’t reach as effectively.

Here’s how to think about resource allocation.

Google should get the majority of your SEO investment, because it’s where most of your audience searches. But the expansion of AI Overviews means you need to optimize for visibility within Google’s AI-generated feature as well.

Bing deserves more attention than most teams give it. Between Bing, Yahoo, and the reported connection to ChatGPT’s web search, Bing’s reach is wider than its global share suggests. On the organic side, reviewing your Bing Webmaster Tools data and ensuring your site is properly indexed is a quick win most teams skip.

The traffic volume for AI search engines is small, but the growth rate is steep. Start tracking whether your content appears in ChatGPT and Perplexity responses. If you aren’t being cited, start with the same content fundamentals that drive good SEO.

Vertical platforms matter for specific audiences. If you sell products, Amazon search optimization isn’t optional. If you target younger demographics, TikTok search and discovery behavior is a factor you can’t ignore. Neither of these platforms runs on traditional SEO signals, so they require dedicated strategy and budget.

Regional engines serve their markets. Yandex and Baidu are essential for Russia and China, respectively, but require native-language teams, localized hosting, and familiarity with platform-specific tools and regulations.

The data makes a case for diversifying beyond Google. The harder decision is how much to invest in each channel based on your specific audience, market, and goals. The answer will be different for every business, but the days of building an entire search strategy around a single engine are over.

More Resources:


Featured Image: Accogliente Design/Shutterstock

How AI Is Changing Lead Generation: 3 Key Things SEO & PPC Teams Need To Do Now via @sejournal, @CallRail

1. Identify Which AI Platforms Are Driving Your Visitors

Each LLM and answer engine has different logic, leading to different outputs for the same prompts. It’s important to understand which AI chatbots are aligned with your brand before making decisions that inform a larger AI search or SEO strategy.

Different LLMs Are Driving Leads In Different Industries

Not all AI platforms send leads the same way.

  • ChatGPT = Speed. ChatGPT dominates overall lead volume at 90.1% of AI-referred leads, with especially strong numbers in healthcare and automotive industries, where people want instant options.
  • Perplexity = Research. Perplexity accounts for 6.3%, but it punches well above its weight in high-consideration sectors. In Travel & Hospitality and Manufacturing, nearly one in ten AI leads comes from Perplexity, roughly ten times the rate seen in other industries.
  • Google’s Gemini holds 2.4% of AI-referred leads and is gaining traction in Business Service and Manufacturing, likely because users lean on its Google Workspace integration.
  • Claude, with 1.2% of lead generation, is carving out a niche in both Real Estate verticals and also with Marketing Agencies. Especially in areas where consumers tend to do more specific and detailed research before reaching out.

How To Accurately Track AI Prompt Visibility

AI search isn’t one channel. It’s a set of distinct platforms, each with different behaviors and industry strengths. So, repeat this AI prompt research phase for each LLM.

  1. Identify the LLMs that matter most for your vertical. Use the data above as a starting point. If you’re in healthcare or automotive, prioritize ChatGPT visibility. High-consideration service? Pay attention to Perplexity. B2B or manufacturing? Gemini should be on your radar.
  2. Test how each platform describes your business. Go to ChatGPT, Perplexity, Gemini, and Claude and ask them questions your customers would ask. “Who’s the best [your service] in [your market]?” See if you’re being recommended. If not, note who is and what content those competitors have that you don’t.
  3. Create content that answers the questions AI platforms are fielding. LLMs favor well-structured, authoritative, fact-rich content. Publish service pages, FAQs, comparison guides, and local content that directly answer the kinds of questions consumers ask these platforms.

2. Connect AI Traffic To Actual Conversions

Connecting AI-driven leads to actual revenue in your reporting is key to understanding how to prioritize your marketing activities. Without visibility into AI lead attribution, you’re making decisions in the dark, which is an expensive place to be.

However, if you can identify AI as the source of your best leads, you instantly know how to pivot your SEO strategy.

How To Track AI Traffic & Attribute Conversions Across ChatGPT, Gemini, and Perplexity

As more money flows through AI search, the ability to attribute leads from specific LLMs isn’t a nice-to-have. It’s the difference between knowing what’s working and throwing budget at a black box.

What you need is the ability to trace a lead from the AI platform where it originated, through the call, form, or chat where it converted, all the way to the revenue it generated. That full-funnel visibility is what separates data-driven teams from everyone else.

  1. Implement LLM-specific attribution. Use a platform that can identify which AI model referred each lead. CallRail’s AI search engine attribution, for example, automatically tags whether an inbound call came from ChatGPT, Perplexity, Gemini, or Claude, not just “AI.” That level of granularity is what makes it possible to actually optimize by channel.
  2. Create custom GA4 channel groups for AI traffic. In Google Analytics, go to Admin > Data Display > Channel Groups and create a custom channel group that isolates AI referral traffic by source. This lets you compare AI-driven sessions and conversions against your other channels.
  3. Add “How did you hear about us?” to your intake process. Self-reported attribution (SRA) is a simple but powerful complement to digital tracking. Add it to your intake forms and train front-desk or sales staff to ask on calls. CallRail’s SRA feature lets you capture this data at the conversation level, so you can compare what callers say against what your analytics show. The gaps will reveal exactly where your tracking is falling short.

See what’s changing: The 2026 Outlook for Marketing Agencies

Connect AI Traffic to Calls, Forms & Sales Pipelines

Call tracking lives in one platform. Form submissions in another. Text conversations somewhere else entirely. Sound familiar?

When your lead data is fragmented like that, it’s surprisingly hard to answer basic questions. Which campaigns drive your best leads? Is AI search actually improving results? Where are leads falling off between first contact and conversion?

Make sure you are monitoring every lead interaction for complete funnel visibility. Teams need clear insight into every conversation-whether it comes through calls, forms, texts, or chats. And by channel- Paid Search, Video, SEO, Paid Social, and Content, for example.

Unifying those touchpoints isn’t just a reporting upgrade. It’s the foundation for any AI-ready lead strategy. Without it, every optimization decision you make is based on an incomplete picture. And in a landscape moving this fast, incomplete data leads to costly missteps.

How To Attribute Calls & Form Fills To AI Search

Take a good look at what is happening with your Voice Assistants. Are forms going to a shared inbox and being missed? Are calls not being answered while another line is in use or after business hours? How long is it taking to follow up with leads? Are those leads going to the competition after you miss the first call?

  1. Consolidate your lead tracking into one platform. If calls, forms, texts, and chats are living in separate tools, you’re creating blind spots. CallRail’s unified lead intelligence platform captures every touchpoint in a single dashboard, so you can see the full customer journey from first AI search to closed deal, and finally answer the question: which channels are actually driving revenue?
  2. Map every conversion point to a marketing source. For each way a lead can reach you -phone call, web form, text, live chat- make sure you can trace it back to the campaign, channel, or keyword that drove it. Use dynamic number insertion for calls and hidden fields on forms to capture source data automatically.
  3. Build a weekly reporting cadence around lead quality, not just volume. Don’t just count leads, score them. Review which sources produce leads that actually convert to appointments and revenue. This is the reporting your clients care about, and it’s how you prove the value of your work

Build the foundation: The Agency Roadmap for 2026 and Beyond

3. Respond Faster To High-Intent AI Traffic

28% of business calls go unanswered. Many of those leads never call back.

Take a good look at your Voice Assistants here. Are your forms going to a shared inbox where they sit unread? Are calls going unanswered because another line is busy or it’s after hours? How long does it take your team to follow up with a new lead? And if you miss that first call from an AI-referred prospect who already has high intent and is ready to buy. Are they going straight to your competitor?

Right now, AI search can understand your customers in real time and answer any question they need, making them perfectly ready to convert into a lead.

Now, it’s you who has to be ready.

Dig into the full data: What 20M Leads Reveal About AI Search and High-Intent Calls

AI Leads Convert Faster. Respond Immediately.

Think about how the traditional funnel used to work. Someone searches, browses a few sites, reads some reviews, maybe sleeps on it, then reaches out. There were days, sometimes weeks, of consideration built into the process.

AI has collapsed that timeline dramatically, and AI-directed callers skip the browsing phase entirely.

They’ve already done their research inside the LLM. By the time they call, they’re ready to make a decision. And they expect you to be ready, too. When a prospect has been pre-qualified by an AI recommendation, every minute of delay costs you revenue.

And the stakes go beyond individual calls.

On platforms like Google, answer speed directly impacts your ad rankings. Faster response times earn better placements on Local Service Ads and PPC -meaning slow follow-up doesn’t just lose you a lead, it quietly erodes your visibility and drives up your cost per lead over time. The agencies winning in an AI-search world aren’t just the ones showing up in LLM recommendations. They’re the ones ready to convert the moment the phone rings -day or night.

Get the playbook: 6 Ways To Prepare Your Business for AI in 2026

Apply AI Where Your Team Is Stretched Thinnest: Use AI to Capture & Qualify Leads Automatically

You can’t automate everything. But knowing where to apply AI, specifically, where your agency or internal team is most stretched, is the difference between using it effectively and adding technology for its own sake.

For most agencies and SMBs, the highest-impact bottleneck is follow-up.

If your clients are missing calls, responding slowly, or losing leads somewhere between the first touch and a booked appointment, that’s exactly where AI can deliver immediate, measurable value.

The key to success here is utilizing AI-powered platforms that can answer inbound calls around the clock, qualify leads in real time, capture intake details, and even book appointments automatically. Early adopters have seen answered calls increase by 44%. That’s not a marginal improvement. It’s the kind of shift that directly impacts revenue and client retention.

How To Set Up AI-Assisted Lead Handling

When you can connect your AI-assisted lead handling back to attribution data and revenue outcomes, you’re no longer just reporting on activities. You’re proving ROI. And that’s what earns long-term client trust- and moves agencies from being seen as just a lead source to being a true growth partner.

  1. Deploy an AI voice agent for after-hours and overflow calls. Start with the windows where your team is least available -evenings, weekends, and lunch hours. CallRail’s Voice Assist answers, qualifies, and captures lead details automatically, so no high-intent caller falls through the cracks. Early adopters have seen answered calls increase by 44%.
  2. Automate follow-up texts immediately after missed calls. If a call does go unanswered, trigger an automatic text within seconds: “Hi, we just missed your call -how can we help?” This simple automation recovers a meaningful percentage of leads that would otherwise be lost.
  3. Connect your AI lead handling back to attribution. Make sure the leads captured by AI tools feed into the same reporting dashboard as your other channels. If your AI agent books an appointment at 9 pm on a Saturday, you should be able to trace that back to the Google Ad or AI search referral that started the journey.

Go deeper: Why The Top Marketers Pair Data With Story

Start Tracking & Optimizing AI-Driven Leads Now

The shift isn’t on the horizon. It’s already here.

It’s time to build AI-aware attribution so you can see what’s actually driving leads, unify your data so you can act on it, and respond fast enough to capture the high-intent leads AI search is already sending your way.

MCP, A2A, NLWeb, And AGENTS.md: The Standards Powering The Agentic Web via @sejournal, @slobodanmanic

This is Part 3 in a five-part series on optimizing websites for the agentic web. Part 1 covered the evolution from SEO to AAIO. Part 2 explored how to get your content cited in AI responses. This article goes deeper: the protocols forming the infrastructure layer that make everything else possible.

The early web needed HTTP to transport data, HTML to structure content, and the W3C to keep everyone building on the same foundation. Without those shared standards, we’d have ended up with a fragmented collection of incompatible networks instead of a single web.

The agentic web is at that same inflection point. AI agents need standardized ways to connect to tools, talk to each other, query websites, and understand codebases. Without shared protocols, every AI vendor builds proprietary integrations, and the result is the same fragmentation the early web narrowly avoided.

Four protocols are emerging as the foundational layer. This article covers what each one does, who’s behind it, and what it means for your business. Throughout this series, we draw exclusively from official documentation, research papers, and announcements from the companies building this infrastructure.

Why Standards Matter

Consider how the original web came together. In the early 1990s, competing browser vendors and incompatible standards were fragmenting what should have been a unified network. The W3C brought order by establishing shared protocols. HTTP handled transport. HTML handled structure. Everyone agreed on the rules, and the web took off.

AI is at a similar crossroads. Right now, every major AI company is building agents that need to interact with external tools, data sources, other agents, and websites. Without standards, connecting your business systems to AI means building separate integrations for Claude, ChatGPT, Gemini, Copilot, and whatever comes next. That’s the M x N problem: M different AI models times N different tools equals an unsustainable number of custom connections.

What makes this moment remarkable is who’s building the solution together. On Dec. 9, 2025, the Linux Foundation announced the Agentic AI Foundation (AAIF), a vendor-neutral governance body for agentic AI standards. Eight platinum members anchor it: AWS, Anthropic, Block, Bloomberg, Cloudflare, Google, Microsoft, and OpenAI.

OpenAI, Anthropic, Google, and Microsoft. Competing on AI products, collaborating on AI infrastructure. As Linux Foundation Executive Director Jim Zemlin put it: “We are seeing AI enter a new phase, as conversational systems shift to autonomous agents that can work together.”

This is a bigger deal than most people realize. Competitors building shared infrastructure because they all recognize that proprietary standards would hold back the entire ecosystem, including themselves.

MCP: The Universal Adapter

What it is: The Model Context Protocol (MCP) is an open standard for connecting AI applications to external tools, data sources, and workflows.

The official analogy is apt:

Think of MCP like a USB-C port for AI applications. Just as USB-C provides a standardized way to connect electronic devices, MCP provides a standardized way to connect AI applications to external systems.”

Before MCP, if you wanted your database, CRM, or internal tools accessible to an AI assistant, you had to build a custom integration for each AI platform. MCP replaces that with a single standard interface. Build one MCP server for your data, and every MCP-compatible AI system can connect to it.

The numbers are striking. MCP launched as an open-source project from Anthropic on Nov. 25, 2024. In just over a year, it reached 97 million monthly SDK downloads across Python and TypeScript, with over 10,000 public MCP servers built by the community.

The adoption timeline tells the story. Anthropic’s Claude had native MCP support from day one. In March 2025, OpenAI CEO Sam Altman announced support across OpenAI’s products, stating: “People love MCP and we are excited to add support across our products.” Google followed in April, confirming MCP support in Gemini. Microsoft joined the MCP steering committee at Build 2025 in May, with MCP support in VS Code reaching general availability in July 2025.

From internal experiment to industry standard in 12 months. That pace of adoption signals something real.

What this means for your business: If your data, tools, or services are MCP-accessible, every major AI platform can use them. That’s not a theoretical benefit. It means an AI assistant helping your customer can pull real-time product availability from your inventory system, check order status from your CRM, or retrieve pricing from your database, all through one standardized connection rather than platform-specific integrations.

A2A: How Agents Talk To Each Other

What it is: The Agent2Agent protocol (A2A) enables AI agents from different vendors to discover each other’s capabilities and collaborate on tasks.

If MCP is how agents connect to tools, A2A is how agents connect to each other. The distinction matters. In a world where businesses use AI agents from Salesforce for CRM, ServiceNow for IT, and an internal agent for billing, these agents need a way to discover what each other can do, delegate tasks, and coordinate responses. A2A provides that.

Google launched A2A on April 9, 2025 with over 50 technology partners. By June, Google donated the protocol to the Linux Foundation. By July, version 0.3 shipped with over 150 supporting organizations, including Salesforce, SAP, ServiceNow, PayPal, Atlassian, Microsoft, and AWS.

The core concept is the Agent Card: a JSON metadata document that serves as a digital business card for agents. Each A2A-compatible agent publishes an Agent Card at a standard web address (/.well-known/agent-card.json) describing its identity, capabilities, skills, and authentication requirements. When one agent needs help with a task, it reads another agent’s card to understand what that agent can do, then communicates through A2A to request collaboration.

Google’s own framing of how these pieces fit together is useful: “Build with ADK, equip with MCP, communicate with A2A.” ADK (Agent Development Kit) is Google’s framework for building agents, MCP gives them access to tools, and A2A lets them talk to other agents.

Here’s a practical example. A customer contacts your company with a billing question that requires a refund. Your customer service agent (built on one platform) identifies the issue, passes the context to your billing agent (built on another platform) via A2A, which calculates the refund amount and hands off to your payments agent (yet another platform) to process it. The customer sees one seamless interaction. Behind the scenes, three agents from different vendors collaborated through a shared protocol.

The enterprise adoption signal is strong. When Salesforce, SAP, ServiceNow, and every major consultancy sign on to a protocol within months, it’s because their enterprise clients are already running into the multi-vendor agent coordination problem that A2A solves.

NLWeb: Making Websites Conversational

What it is: NLWeb (Natural Language Web) is an open project from Microsoft that turns any website into a natural language interface, queryable by both humans and AI agents.

Of the four protocols covered here, NLWeb is the most directly relevant to this series’ audience. MCP, A2A, and AGENTS.md are primarily developer concerns. NLWeb is about your website.

NLWeb was introduced at Microsoft Build 2025 on May 19, 2025. It was conceived and developed by R.V. Guha, who joined Microsoft as CVP and Technical Fellow. If that name sounds familiar, it should: Guha is the creator of RSS, RDF, and Schema.org, three standards that fundamentally shaped how the web organizes and syndicates information. When the person behind Schema.org builds a new web protocol, it’s worth paying attention.

The key insight behind NLWeb is that websites already publish structured data. Schema.org markup, RSS feeds, product catalogs, recipe databases. NLWeb leverages these existing formats, combining them with AI to let users and agents query a website’s content using natural language instead of clicking through pages.

Microsoft’s framing is deliberate: “NLWeb can play a similar role to HTML in the emerging agentic web.” The NLWeb README puts it even more directly: “NLWeb is to MCP/A2A what HTML is to HTTP.”

Every NLWeb instance is automatically an MCP server. That means any website running NLWeb immediately becomes accessible to the entire ecosystem of MCP-compatible AI assistants and agents. Your website’s content doesn’t just sit there waiting for visitors. It becomes actively queryable by any AI system that speaks MCP.

Early adopters include Eventbrite, Shopify, Tripadvisor, O’Reilly Media, Common Sense Media, and Hearst. These are content-rich websites that already invest heavily in structured data. NLWeb builds directly on that investment.

Here’s what this looks like in practice. Instead of a user navigating Tripadvisor’s search filters to find family-friendly restaurants in Barcelona with outdoor seating, an AI agent could query Tripadvisor’s NLWeb endpoint: “Find family-friendly restaurants in Barcelona with outdoor seating and good reviews.” The response comes back as structured Schema.org JSON, ready for the agent to present to the user or act on.

If your business has already invested in Schema.org markup (and Part 2 of this series explained why you should), you’re closer to NLWeb readiness than you might think.

AGENTS.md: Instructions For AI Coders

What it is: AGENTS.md is a standardized Markdown file that provides AI coding agents with project-specific guidance, essentially a README written for machines instead of humans.

This protocol is less directly relevant to the marketers and strategists reading this series, but it’s an important piece of the complete picture, especially if your organization has development teams using AI coding tools.

AGENTS.md emerged from a collaboration between OpenAI Codex, Google Jules, Cursor, Amp, and Factory. The problem they were solving: AI coding agents need to understand project conventions, build steps, testing requirements, and architectural decisions before they can contribute useful code. Without explicit guidance, agents make assumptions that lead to inconsistent, buggy output.

Since its release in August 2025, AGENTS.md has been adopted by over 60,000 open-source projects and is supported by tools including GitHub Copilot, Claude Code, Cursor, Gemini CLI, VS Code, and many others. It’s now governed by the Agentic AI Foundation, alongside MCP.

The file itself is simple. Plain Markdown, typically under 150 lines, covering build commands, architectural overview, coding conventions, and testing requirements. Agents read it before making any changes, getting the same tribal knowledge that senior engineers carry in their heads.

GitHub reports that Copilot now generates 46% of code for its users. When nearly half of code is AI-generated, having a standard way to ensure agents follow your conventions, security practices, and architectural patterns isn’t optional. It’s quality control.

Why this matters for your business: If your development teams use AI coding tools (and most do), AGENTS.md ensures those tools produce code that matches your standards. It reduces agent-generated bugs, cuts onboarding time for AI tools on new projects, and provides consistency across teams.

How They Fit Together

These four protocols aren’t competing. They’re complementary layers in the same stack.

Protocol Created By Purpose Web Analogy
MCP Anthropic Connect agents to tools and data USB ports
A2A Google Agent-to-agent communication Email/messaging
NLWeb Microsoft Make websites queryable by agents HTML
AGENTS.md OpenAI + collaborators Guide AI coding agents README files
AAIF Linux Foundation Governance and standards body W3C

The stack works like this: MCP provides the plumbing for agents to access tools and data. A2A enables agents to coordinate with each other. NLWeb makes website content accessible to the entire ecosystem. AGENTS.md ensures AI coding agents build correctly. And the Agentic AI Foundation provides the governance layer, ensuring these protocols remain open, vendor-neutral, and interoperable.

The parallel to the original web is impossible to ignore:

  • HTTP (transport) maps to MCP (tool access) and A2A (agent communication).
  • HTML (content structure) maps to NLWeb (website content for agents).
  • W3C (governance) maps to AAIF (governance).

What’s different this time is the speed. HTTP took years to gain broad adoption. MCP went from launch to universal platform support in 12 months. A2A grew from 50 to 150+ partner organizations in three months. NLWeb shipped with major publisher adoption at launch. AGENTS.md reached 60,000 projects within its first few months.

The infrastructure is being built at internet speed, not standards-committee speed. That’s partly because the companies involved are the same ones building the agents that need these protocols. They’re motivated.

And these four aren’t the only protocols emerging. Commerce-specific standards are building the transaction layer: Shopify and Google co-developed the Universal Commerce Protocol (UCP), launched in January 2026 with support from Etsy, Target, Walmart, and Wayfair. OpenAI and Stripe co-developed the Agentic Commerce Protocol (ACP), which powers Instant Checkout in ChatGPT. CopilotKit’s AG-UI protocol addresses agent-to-frontend communication, with integrations from LangGraph, CrewAI, and Google ADK. We’ll cover the commerce protocols in depth in Part 5.

What This Means For Your Business

You don’t need to implement all four protocols tomorrow. But you need to understand what’s being built, because it shapes what your website, tools, and teams should be ready for.

If you’ve already invested in Schema.org markup, NLWeb is your closest on-ramp. It builds directly on the structured data you already maintain. As NLWeb adoption grows, your Schema.org investment becomes the foundation for making your website conversationally accessible to AI agents. Keep your structured data current and comprehensive.

If you have APIs or internal tools, consider MCP accessibility. Making your services available through MCP means any AI platform can interact with them. For ecommerce, that could mean product catalogs, inventory systems, and order tracking becoming accessible to AI shopping assistants across ChatGPT, Claude, Gemini, and whatever comes next.

If you’re evaluating multi-vendor agent workflows, A2A is the protocol to watch. Enterprise organizations running agents from multiple vendors (Salesforce, ServiceNow, internal tools) will increasingly need these agents to coordinate. A2A is the emerging standard for that coordination.

If your development teams use AI coding tools, adopt AGENTS.md now. It’s the simplest protocol to implement (it’s a single Markdown file) and the one with the most immediate, tangible benefit: fewer bugs, more consistent output, faster onboarding for AI tools on your codebase.

The underlying message across all four protocols is the same: the agentic web is being built on open standards, not proprietary ones. The companies that understand these standards early will be better positioned as AI agents become a primary way users interact with businesses.

These aren’t things you need to implement today. But they are things you need to understand, because Part 4 of this series gets into the technical specifics of making your website agent-ready.

Key Takeaways

  • Four protocols form the agentic web’s infrastructure. MCP (tools), A2A (agent communication), NLWeb (website content), and AGENTS.md (code guidance) are complementary layers, not competitors.
  • The speed of adoption signals real urgency. MCP reached 97 million monthly SDK downloads and universal platform support in 12 months. A2A grew from 50 to 150+ partner organizations in three months. These are not experiments.
  • Competitors are collaborating on infrastructure. OpenAI, Anthropic, Google, and Microsoft are all building shared protocols under the Agentic AI Foundation. This mirrors the W3C moment that unified the early web.
  • NLWeb is potentially the most relevant protocol for website owners. Built by the creator of Schema.org, it turns your existing structured data into a conversational interface for AI agents. Every NLWeb instance is automatically an MCP server.
  • MCP is the universal adapter. Build one MCP connection to your data, and every major AI platform (Claude, ChatGPT, Gemini, Copilot) can access it. No more building separate integrations for each platform.
  • Start with what you have. Schema.org markup readies you for NLWeb. Existing APIs can become MCP servers. AGENTS.md is a single file your dev team can create today. You don’t need to start from scratch.

The original web succeeded because competitors agreed on shared standards. The agentic web is following the same playbook, just faster. The protocols are being established now. The governance is in place. The agents are already using them.

Up next in Part 4: the hands-on technical guide for making your website ready for autonomous AI agents, from semantic HTML to accessibility standards to testing with real agent tools.

More Resources:


This post was originally published on No Hacks.


Featured Image: Collagery/Shutterstock

Why Agentic AI Shopping Feels Unnatural And May Not Threaten SEO via @sejournal, @martinibuster

Google, OpenAI, and Shopify insist that the next revolution in AI is agentic AI shopping agents. Shopping is a lucrative area for AI to burrow into. The thing that I keep thinking is that shopping is a deeply important activity to humans; it’s literally a part of our DNA. Is surrendering the shopping experience something the general public is willing to do?

Agentic AI shopping is like a personal assistant that you tell what you want and maybe why you need it, plus some features and a price range. The AI will go out and do the research and comparison and even make the purchase.

There’s no human performing a search in that scenario. So it’s kind of not necessarily good for SEO unless you’re optimizing shopping sites for agentic AI shoppers.

Shopping Is A Part Of Human Biology

Scientists say that shopping is literally a part of our DNA. Our desire to hunt, to gather, and to flaunt our ability to be successful is a part of the evolutionary competition we participate in (whether we know it or not).

A Wikipedia page on the subject explains:

“Richard Dawkins outlines in The Selfish Gene (1976) that humans are machines made of genes, and genes are the grounding for everything people do.

…Therefore, everything that people do relates to thriving in their environment above competition, including the way people consume as a form of survival in their environment when simply purchasing the basic physiological needs of food, water and warmth. People also consume to thrive above others, for example in conspicuous consumption where a luxury car represents money and high social status…”

What that means is that whether we know it or not, our drive to shop is a part of evolutionary competition with each other. Part of it is to signal our status and attractiveness for reproduction. So when we go shopping for clothes or toilet paper, it’s part of our genetic programming to feel good about it.

Shopping And The Brain’s Chemical Cocktail

And when it comes to feeling good, some of that is triggered by chemicals like dopamine, endorphins, and serotonin firing off to reward you for finding a good deal.

Even scoring a deal on toilet paper can trigger reward signals in the brain.

Another Wikipedia page about the biology of our reward system explains:

“Reward is the attractive and motivational property of a stimulus that induces appetitive behavior, also known as approach behavior, and consummatory behavior. A rewarding stimulus has been described as “any stimulus, object, event, activity, or situation that has the potential to make us approach and consume it.”

A sale sign in a store can act as a reward cue because it signals a lower price or added value, which can drive someone to approach and buy it. The sign itself is just information, but when a person recognizes the discount or deal as beneficial, it can trigger motivation to act. That’s a deeply embedded behavior that we carry with us.

We are like machines that are programmed in our genes to shop.

So that raises the question: Why would anyone delegate that deeply rewarding activity to an AI agent? It’s like delegating the enjoyment of chocolate to a robot.

I suspect that most of you reading this know which supermarkets sell the best produce at the cheapest price, which ones have the yummiest bread, and which markets have the best spices. That’s our programming; it’s biological. It does not make sense to delegate the rewards inherent in discovery or acquisition to an AI shopping agent.

Serendipity And Shopping

Serendipity is when things happen by chance, unplanned, that nonetheless provide a happy outcome or benefit. One of the joys of shopping is stumbling onto something that’s a good deal or beautiful or has some other value. Employing an AI agent will cause humans to miss out on the serendipitous joy of discovering something they hadn’t been looking for that is not just desirable but also something they hadn’t known they needed.

For example, I purchased a birthday gift for my wife. I walked into a gift shop run by a charming new age hippie. We talked about music as I browsed the gifts for sale. I found something, two things, that I hadn’t planned on buying. The two things had a semantic connection to each other that I found to be poetic and therefore extra nice as a gift. The shop owner put the two items into two boxes, then placed the boxes in a lovely mesh gift bag with a ribbon.

That’s serendipity in action. It was a pleasurable moment I enjoyed. I walked out of the store into the sunshine with a fresh cocktail of dopamine, endorphins, and serotonin flooding my brain, and it was a delightful moment. I bought a gift that I was certain my wife would enjoy.

Agentic AI Shopping Is Unnatural

My question is, why does Silicon Valley think it can automate the many things that make us human?

It’s as if Silicon Valley is trying to convert us into teenagers by doing the things adults normally do.

Now they want to take shopping away from us?

I think that the only way that agentic AI has a chance of working is if they build in a sense of serendipity and discovery into the system. I’ve been a part of the technology scene for over 25 years, I lived in the world capital of the Internet in San Francisco and even worked for a time at a leading technology magazine.

So it’s not that I’m a luddite about technology. AI integrated into a shopping site makes a lot of sense. It can make recommendations and answer questions. That’s great. There is still a human who is clicking around and discovering things for themselves in a way that satisfies are natural urge to shop and consume. That’s good for SEO because it means that a store needs to be optimized for search.

AI agents doing the shopping for humans makes less sense because it’s unnatural, it goes against our biology.

Featured Image by Shutterstock/Prostock-studio

Google Core Update, Crawl Limits & Gemini Traffic Data – SEO Pulse via @sejournal, @MattGSouthern

Welcome to the week’s Pulse: updates affect how Google ranks content, how its crawlers handle page size, and where AI referral traffic is heading. Here’s what matters for you and your work.

Google Rolls Out The March 2026 Core Update

Google began rolling out the March core update this week. This is the first broad core update of the year.

Key facts: The rollout may take up to two weeks. Google described it as a regular update designed to surface more relevant, satisfying content from all types of sites. It arrives two days after the March spam update completed in under 20 hours.

Why This Matters

The December core update was the most recent broad core update, finishing on December 29. That’s a three-month gap. The February 2026 update only affected Discover, so Search rankings haven’t been recalibrated since late December.

Ranking changes could appear throughout early April. Google recommends waiting at least a full week after the rollout finishes before analyzing Search Console performance. Compare against a baseline period before March 27.

What SEO Professionals Are Saying

John Mueller, a member of Google’s Search Relations team, wrote on Bluesky when asked whether the two updates overlap:

One is about spam, one is not about spam. If with some experience, you’re not sure whether your site is spam or not, it’s unfortunately probably spam.

Mueller later explained that core updates don’t follow a single deployment mechanism. Different teams and systems contribute changes, and those components can require step-by-step rollouts rather than a single release. That’s why rollouts take weeks and why ranking volatility often appears in waves rather than all at once.

Roger Montti, writing for Search Engine Journal, noted the proximity to the spam update may not be a coincidence. Spam fighting is logically part of the broader quality reassessment in a core update.

Read our full coverage: Google Begins Rolling Out March 2026 Core Update

Read Roger Montti’s coverage: Google Answers Why Core Updates Can Roll Out In Stages

Illyes Explains Googlebot’s Crawling Architecture And Byte Limits

Google’s Gary Illyes, an analyst on Google’s Search team, published a blog post explaining how Googlebot works within Google’s broader crawling systems. The post adds new technical details to the 2 MB crawl limit Google published earlier this year.

Key facts: Illyes described Googlebot as one client of a centralized crawling platform. Google Shopping, AdSense, and other products all route requests through the same system under different crawler names. HTTP request headers count toward the 2 MB limit. External resources like CSS and JavaScript get their own separate byte counters.

Why This Matters

When Googlebot hits 2 MB, it doesn’t reject the page. It stops fetching and passes the truncated content to indexing as if it were the complete file. Anything past 2 MB is never indexed. That matters for pages with large inline base64 images, heavy inline CSS or JavaScript, or oversized navigation menus.

The centralized platform detail also explains why different Google crawlers behave differently in server logs. Each client sets its own configuration, including byte limits. Googlebot’s 2 MB is a Search-specific override of the platform’s 15 MB default.

Google has now covered these limits in documentation updates, a podcast episode, and this blog post within two months. Illyes noted the 2 MB limit is not permanent and may change as the web evolves.

What SEO Professionals Are Saying

Cyrus Shepard, founder of Zyppy SEO, wrote on LinkedIn:

That said, as SEOs we often deal with extreme situations. If you notice certain content not getting indexed on VERY LARGE PAGES, you probably want to check your size.

Read our full coverage: Google Explains Googlebot Byte Limits And Crawling Architecture

Google’s Illyes And Splitt: Pages Are Getting Larger, And It Still Matters

Gary Illyes and Martin Splitt, Developer Advocate at Google, discussed page weight growth and crawling on a recent Search Off the Record podcast episode.

Key facts: Web pages have grown nearly 3x over the past decade. The 15 MB default applies across Google’s broader crawling systems, with individual clients like Googlebot for Search overriding it downward to 2 MB. Illyes raised whether structured data that Google asks websites to add is contributing to page bloat.

Why This Matters

The 2025 Web Almanac reports a median mobile homepage size of 2,362 KB. This indicates pages are getting larger, though it should not be considered safely below Googlebot’s 2 MB fetch limit. However, Illyes’s question about structured data contributing to bloat is worth monitoring. Google encourages sites to add schema markup for rich results, and that markup increases the weight of each page.

Splitt said he plans to address specific techniques for reducing page size in a future episode. Pages with heavy inline content should verify their critical elements load within the first 2 MB of the response.

Read our full coverage: Google: Pages Are Getting Larger & It Still Matters

Gemini Referral Traffic More Than Doubles, Overtakes Perplexity

Google Gemini more than doubled its referral traffic to websites between November 2025 and January 2026. The data comes from SE Ranking’s analysis of more than 101,000 sites with Google Analytics installed.

Key facts: SE Ranking measured a 115% combined increase over two months, with the jump starting around the time Google rolled out Gemini 3. In January, Gemini sent 29% more referral traffic than Perplexity globally and 41% more in the U.S. ChatGPT still generates about 80% of all AI referral traffic. For transparency, SE Ranking sells AI visibility tracking tools.

Why This Matters

In August 2025, Perplexity was sending about 2.9x more referral traffic than Gemini. Gemini’s December-January surge reversed that by January 2026. ChatGPT’s lead over Gemini also narrowed, from roughly 22x in October to about 8x in January.

All AI platforms combined still account for about 0.24% of global internet traffic, up from 0.15% in 2025. That’s measurable growth, but it’s still a small share compared to organic search. Two months of Gemini growth correlates with a known product launch, but it’s too early to call it a sustained pattern.

Gemini is now worth watching alongside ChatGPT and Perplexity in your referral reports.

Read our full coverage: Google Gemini Sends More Traffic To Sites Than Perplexity: Report


Theme Of The Week: Google Is Explaining Its Own Systems

Three of this week’s four stories are Google telling you how its systems work. Illyes published a blog post detailing Googlebot’s architecture. The same week, the Search Off the Record podcast covered page weight and crawl thresholds. Mueller explained why core updates roll out in waves rather than all at once. Each one fills a gap that documentation alone left open.

The Gemini traffic data provides a new perspective. Google is being open about how its crawlers and ranking systems operate. The traffic passing through its AI services is increasing rapidly enough to be reflected in third-party data, and Google isn’t explaining that part.

Top Stories Of The Week:

More Resources: