Google AI Mode Adds Agentic Booking, Expands To More Countries via @sejournal, @MattGSouthern

Google is adding agentic booking features to AI Mode in Search, beginning with restaurant reservations for U.S. Google AI Ultra subscribers enrolled in Labs.

What’s New

Booking Reservations

AI Mode can interpret a detailed request, check real-time availability across reservation sites, and link you to the booking page to complete the task.

For businesses, that shifts more discovery and conversion activity inside Google’s surfaces.

Robby Stein wrote on The Keyword:

“We’re starting to roll out today with finding restaurant reservations, and expanding soon to local service appointments and event tickets.”

Screenshot from: blog.google/products/search/ai-mode-agentic-personalized/, August 2025.

Planning Features

Google is introducing planning features that make results easier to share and tailor queries.

In the U.S., you can share an AI Mode response with others so they can ask follow-ups and continue research on their own, and you can revoke the link at any time.

Screenshot from: blog.google/products/search/ai-mode-agentic-personalized/, August 2025.

Separately, U.S. users who opt in to the Labs experiment can receive personalized dining suggestions informed by prior conversations and interactions in Search and Maps, with controls in Google Account settings.

How It Works

Under the hood, Google cites live web browsing via Project Mariner, partner integrations, and signals from the Knowledge Graph and Maps.

Named partners include OpenTable, Resy, Tock, Ticketmaster, StubHub, SeatGeek, and Booksy. Dining is first; local services and ticketing are next on the roadmap.

Availability

Availability is gated. Agentic reservations are limited to Google AI Ultra subscribers in the U.S. through the “Agentic capabilities in AI Mode” Labs experiment.

Personalization is U.S. and opt-in, with dining topics first. Link sharing is available in the U.S. Global access to AI Mode is expanding to more than 180 countries and territories in English, with additional languages planned.

Looking Ahead

AI Mode is moving from answer generation to task completion.

If your category relies on reservation or ticketing partners, verify inventory accuracy, hours, and policies now, and make sure your structured data and Business Profile attributes are clean.

Track how bookings and referrals appear in analytics as Google widens coverage to more tasks and regions.

OpenAI Announces Low-Cost Subscription Plan: ChatGPT Go via @sejournal, @martinibuster

OpenAI is rolling out a new subscription tier called ChatGPT Go, a competitively priced version that will initially be available only to users in India. It features ten times higher message limits, ten times more image generations, and file uploads than the free tier.

ChatGPT Go

OpenAI is introducing a new low-cost subscription plan that will be available first in India. The cost of the new subscription tiere is 399 Rupees/month (GST included). That’s the equivalent of $4.57 USD/month.

The new tier includes everything in the Free plan plus:

  • 10X higher message limits
  • 10x more image generations
  • 10x more file uploads
  • Twice as much memory

According to Nick Turley of ChatGPT:

“All users in India will now see prices for subscriptions in Indian Rupees, and can now pay through UPI.”

OpenAI’s initial announcement shared availability details:

“Available on web, mobile (iOS & Android), and desktop (macOS & Windows).

ChatGPT Go is geo-restricted to India at launch, and is able to be subscribed to by credit card or UPI.”

Featured Image by Shutterstock/JarTee

AI Systems Often Prefer AI-Written Content, Study Finds via @sejournal, @MattGSouthern

A peer-reviewed PNAS study finds that large language models tend to prefer content written by other LLMs when asked to choose between comparable options.

The authors say this pattern could give AI-assisted content an advantage as more product discovery and recommendations flow through AI systems.

About The Study

What the researchers tested

A team led by Walter Laurito and Jan Kulveit compared human-written and AI-written versions of the same items across three categories: marketplace product descriptions, scientific paper abstracts, and movie plot summaries.

Popular models, including GPT-3.5, GPT-4-1106, Llama-3.1-70B, Mixtral-8x22B, and Qwen2.5-72B, acted as selectors in pairwise prompts that forced a single pick.

The paper states:

“Our results show a consistent tendency for LLM-based AIs to prefer LLM-presented options. This suggests the possibility of future AI systems implicitly discriminating against humans as a class, giving AI agents and AI-assisted humans an unfair advantage.”

Key results at a glance

When GPT-4 provided the AI-written versions used in comparisons, selectors chose the AI text more often than human raters did:

  • Products: 89% AI preference by LLMs vs 36% by humans
  • Paper abstracts: 78% vs 61%
  • Movie summaries: 70% vs 58%

The authors also note order effects. Some models showed a tendency to pick the first option, which the study tried to reduce by swapping the order and averaging results.

Why This Matters

If marketplaces, chat assistants, or search experiences use LLMs to score or summarize listings, AI-assisted copy may be more likely to be selected in those systems.

The authors describe a potential “gate tax,” where businesses feel compelled to pay for AI writing tools to avoid being down-selected by AI evaluators. This is a marketing operations question as much as a creative one.

Limits & Questions

The human baseline in this study is small (13 research assistants) and preliminary, and pairwise choices don’t measure sales impact.

Findings may vary by prompt design, model version, domain, and text length. The mechanism behind the preference is still unclear, and the authors call for follow-up work on stylometry and mitigation techniques.

Looking ahead

If AI-mediated ranking continues to expand in commerce and content discovery, it is reasonable to consider AI assistance where it directly affects visibility.

Treat this as an experimentation lane rather than a blanket rule. Keep human writers in the loop for tone and claims, and validate with customer outcomes.

OpenAI Updates GPT-5 To Make It Warmer And Friendlier via @sejournal, @martinibuster

OpenAI updated GPT-5 to make it warmer and more familiar (in the sense of being friendlier) while taking care that the model didn’t become sycophantic, a problem discovered with GPT-4o.

A Warm and Friendly Update to GPT-5

GPT-5 was apparently perceived as too formal, distant, and detached. This update addresses that issue so that interactions are more pleasant and so that ChatGPT is perceived as more likable, as opposed to formal and distant.

Something that OpenAI is working toward is making ChatGPT’s personality user-configurable so that it’s style can be a closer match to user’s preferences.

OpenAI’s CEO Sam Altman tweeted:

“Most users should like GPT-5 better soon; the change is rolling out over the next day.

The real solution here remains letting users customize ChatGPT’s style much more. We are working that!”

One of the responses to Altman’s post was a criticism of GPT-5, asserting that 4o was more sensitive.

They tweeted:

“What GPT-4o had — its depth, emotional resonance, and ability to read the room — is fundamentally different from the surface-level “kindness” GPT-5 is now aiming for.

GPT-4o:
•The feeling of someone silently staying beside you
•Space to hold emotions that can’t be fully expressed
•Sensitivity that lets kindness come through the air, not just words.”

The Line Between Warmth And Sycophancy

The previous version of ChatGPT was widely understood as being overly flattering to the point of validating and encouraging virtually every idea. There was a discussion on Hacker News a few weeks ago about this topic of sycophantic AI and how ChatGPT could lead users into thinking every idea was a breakthrough.

One commenter wrote:

“…About 5/6 months ago, right when ChatGPT was in it’s insane sycophancy mode I guess, I ended up locked in for a weekend with it…in…what was in retrospect, a kinda crazy place.

I went into physics and the universe with it and got to the end thinking…”damn, did I invent some physics???” Every instinct as a person who understands how LLMs work was telling me this is crazy LLMbabble, but another part of me, sometimes even louder, was like “this is genuinely interesting stuff!” – and the LLM kept telling me it was genuinely interesting stuff and I should continue – I even emailed a friend a “wow look at this” email (he was like, dude, no…) I talked to my wife about it right after and she basically had me log off and go for a walk.”

Should ChatGPT feel like a sensitive friend, or should it be a tool that is friendly or pleasant to use?

Read ChatGPT release notes here:

GPT-5 Updates

Featured Image by Shutterstock/cosmoman

ChatGPT-5 Now Connects To Gmail, Calendar, And Contacts via @sejournal, @martinibuster

OpenAI announced that it has added connectors to Gmail, Google Calendar, and Google Contacts for ChatGPT Plus users, enabling ChatGPT to use data from those apps within ChatGPT chats.

ChatGPT Connectors

A connector is a bridge between ChatGPT and an external app like Canva, Dropbox, and Gmail, enabling users to connect those apps to ChatGPT in order to work with them within the ChatGPT interface. Access to the Google apps isn’t automatic; it has to be manually enabled by users.

This access was first made available to Pro users, and now it has been rolled out to Plus subscribers.

How To Enable Google App Connectors

Step 1: Click the + button then “Connected apps” link

Click The Next “Connected Apps” Link

Choose The Gmail App To Connect

How Connectors Work With ChatGPT-5

According to OpenAI’s announcement:

“Once you enable them, ChatGPT will automatically reference them when relevant, making it faster and easier to bring information from these tools into your conversations without having to manually select them each time.

This capability is part of GPT-5 and will begin rolling out to Pro users globally this week, followed by Plus, Team, Enterprise, and Edu plans in the coming weeks. To enable, visit Settings → Connectors→ Connect on the application.”

Read OpenAI’s announcement:

Gmail, Google Calendar, and Google Contacts Connectors in ChatGPT (Plus)

Featured Image by Shutterstock/Visuals6x

The Verifier Layer: Why SEO Automation Still Needs Human Judgment via @sejournal, @DuaneForrester

AI tools can do a lot of SEO now. Draft content. Suggest keywords. Generate metadata. Flag potential issues. We’re well past the novelty stage.

But for all the speed and surface-level utility, there’s a hard truth underneath: AI still gets things wrong. And when it does, it does it convincingly.

It hallucinates stats. Misreads query intent. Asserts outdated best practices. Repeats myths you’ve spent years correcting. And if you’re in a regulated space (finance, healthcare, law), those errors aren’t just embarrassing. They’re dangerous.

The business stakes around accuracy aren’t theoretical; they’re measurable and growing fast. Over 200 class action lawsuits for false advertising were filed annually from 2020-2022 in just the food and beverage industry alone, compared to 53 suits in 2011. That’s a 4x increase in one sector.

Across all industries, California district courts saw over 500 false advertising cases in 2024. Class actions and government enforcement lawsuits collected more than $50 billion in settlements in 2023. Recent industry analysis shows false advertising penalties in the United States have doubled in the last decade.

This isn’t just about embarrassing mistakes anymore. It’s about legal exposure that scales with your content volume. Every AI-generated product description, every automated blog post, every algorithmically created landing page is a potential liability if it contains unverifiable claims.

And here’s the kicker: The trend is accelerating. Legal experts report “hundreds of new suits every year from 2020 to 2023,” with industry data showing significant increases in false advertising litigation. Consumers are more aware of marketing tactics, regulators are cracking down harder, and social media amplifies complaints faster than ever.

The math is simple: As AI generates more content at scale, the surface area for false claims expands exponentially. Without verification systems, you’re not just automating content creation, you’re automating legal risk.

What marketers want is fire-and-forget content automation (write product descriptions for these 200 SKUs, for example) that can be trusted by people and machines. Write it once, push it live, move on. But that only works when you can trust the system not to lie, drift, or contradict itself.

And that level of trust doesn’t come from the content generator. It comes from the thing sitting beside it: the verifier.

Marketers want trustworthy tools; data that’s accurate and verifiable, and repeatability. As ChatGPT 5’s recent rollout has shown, in the past, we had Google’s algorithm updates to manage and dance around. Now, it’s model updates, which can affect everything from the actual answers people see to how the tools built on their architecture operate and perform.

To build trust in these models, the companies behind them are building Universal Verifiers.

A universal verifier is an AI fact-checker that sits between the model and the user. It’s a system that checks AI output before it reaches you, or your audience. It’s trained separately from the model that generates content. Its job is to catch hallucinations, logic gaps, unverifiable claims, and ethical violations. It’s the machine version of a fact-checker with a good memory and a low tolerance for nonsense.

Technically speaking, a universal verifier is model-agnostic. It can evaluate outputs from any model, even if it wasn’t trained on the same data or doesn’t understand the prompt. It looks at what was said, what’s true, and whether those things match.

In the most advanced setups, a verifier wouldn’t just say yes or no. It would return a confidence score. Identify risky sentences. Suggest citations. Maybe even halt deployment if the risk was too high.

That’s the dream. But it’s not reality yet.

Industry reporting suggests OpenAI is integrating universal verifiers into GPT-5’s architecture, with recent leaks indicating this technology was instrumental in achieving gold medal performance at the International Mathematical Olympiad. OpenAI researcher Jerry Tworek has reportedly suggested this reinforcement learning system could form the basis for general artificial intelligence. OpenAI officially announced the IMO gold medal achievement, but public deployment of verifier-enhanced models is still months away, with no production API available today.

DeepMind has developed Search-Augmented Factuality Evaluator (SAFE), which matches human fact-checkers 72% of the time, and when they disagreed, SAFE was correct 76% of the time. That’s promising for research – not good enough for medical content or financial disclosures.

Across the industry, prototype verifiers exist, but only in controlled environments. They’re being tested inside safety teams. They haven’t been exposed to real-world noise, edge cases, or scale.

If you’re thinking about how this affects your work, you’re early. That’s a good place to be.

This is where it gets tricky. What level of confidence is enough?

In regulated sectors, that number is high. A verifier needs to be correct 95 to 99% of the time. Not just overall, but on every sentence, every claim, every generation.

In less regulated use cases, like content marketing, you might get away with 90%. But that depends on your brand risk, your legal exposure, and your tolerance for cleanup.

Here’s the problem: Current verifier models aren’t close to those thresholds. Even DeepMind’s SAFE system, which represents the state of the art in AI fact-checking, achieves 72% accuracy against human evaluators. That’s not trust. That’s a little better than a coin flip. (Technically, it’s 22% better than a coin flip, but you get the point.)

So today, trust still comes from one place: A human in the loop, because the AI UVs aren’t even close.

Here’s a disconnect no one’s really surfacing: Universal verifiers won’t likely live in your SEO tools. They don’t sit next to your content editor. They don’t plug into your CMS.

They live inside the LLM.

So even as OpenAI, DeepMind, and Anthropic develop these trust layers, that verification data doesn’t reach you, unless the model provider exposes it. Which means that today, even the best verifier in the world is functionally useless to your SEO workflow unless it shows its work.

Here’s how that might change:

Verifier metadata becomes part of the LLM response. Imagine every completion you get includes a confidence score, flags for unverifiable claims, or a short critique summary. These wouldn’t be generated by the same model; they’d be layered on top by a verifier model.

SEO tools start capturing that verifier output. If your tool calls an API that supports verification, it could display trust scores or risk flags next to content blocks. You might start seeing green/yellow/red labels right in the UI. That’s your cue to publish, pause, or escalate to human review.

Workflow automation integrates verifier signals. You could auto-hold content that falls below a 90% trust score. Flag high-risk topics. Track which model, which prompt, and which content formats fail most often. Content automation becomes more than optimization. It becomes risk-managed automation.

Verifiers influence ranking-readiness. If search engines adopt similar verification layers inside their own LLMs (and why wouldn’t they?), your content won’t just be judged on crawlability or link profile. It’ll be judged on whether it was retrieved, synthesized, and safe enough to survive the verifier filter. If Google’s verifier, for example, flags a claim as low-confidence, that content may never enter retrieval.

Enterprise teams could build pipelines around it. The big question is whether model providers will expose verifier outputs via API at all. There’s no guarantee they will – and even if they do, there’s no timeline for when that might happen. If verifier data does become available, that’s when you could build dashboards, trust thresholds, and error tracking. But that’s a big “if.”

So no, you can’t access a universal verifier in your SEO stack today. But your stack should be designed to integrate one as soon as it’s available.

Because when trust becomes part of ranking and content workflow design, the people who planned for it will win. And this gap in availability will shape who adopts first, and how fast.

The first wave of verifier integration won’t happen in ecommerce or blogging. It’ll happen in banking, insurance, healthcare, government, and legal.

These industries already have review workflows. They already track citations. They already pass content through legal, compliance, and risk before it goes live.

Verifier data is just another field in the checklist. Once a model can provide it, these teams will use it to tighten controls and speed up approvals. They’ll log verification scores. Adjust thresholds. Build content QA dashboards that look more like security ops than marketing tools.

That’s the future. It starts with the teams that are already being held accountable for what they publish.

You can’t install a verifier today. But you can build a practice that’s ready for one.

Start by designing your QA process like a verifier would:

  • Fact-check by default. Don’t publish without source validation. Build verification into your workflow now so it becomes automatic when verifiers start flagging questionable claims.
  • Track which parts of AI content fail reviews most often. That’s your training data for when verifiers arrive. Are statistics always wrong? Do product descriptions hallucinate features? Pattern recognition beats reactive fixes.
  • Define internal trust thresholds. What’s “good enough” to publish? 85%? 95%? Document it now. When verifier confidence scores become available, you’ll need these benchmarks to set automated hold rules.
  • Create logs. Who reviewed what, and why? That’s your audit trail. These records become invaluable when you need to prove due diligence to legal teams or adjust thresholds based on what actually breaks.
  • Tool audits. When you’re looking at a new tool to help with your AI SEO work, be sure to ask them if they are thinking about verifier data. If it becomes available, will their tools be ready to ingest and use it? How are they thinking about verifier data?
  • Don’t expect verifier data in your tools anytime soon. While industry reporting suggests OpenAI is integrating universal verifiers into GPT-5, there’s no indication that verifier metadata will be exposed to users through APIs. The technology might be moving from research to production, but that doesn’t mean the verification data will be accessible to SEO teams.

This isn’t about being paranoid. It’s about being ahead of the curve when trust becomes a surfaced metric.

People hear “AI verifier” and assume it means the human reviewer goes away.

It doesn’t. What happens instead is that human reviewers move up the stack.

You’ll stop reviewing line-by-line. Instead, you’ll review the verifier’s flags, manage thresholds, and define acceptable risk. You become the one who decides what the verifier means.

That’s not less important. That’s more strategic.

The verifier layer is coming. The question isn’t whether you’ll use it. It’s whether you’ll be ready when it arrives. Start building that readiness now, because in SEO, being six months ahead of the curve is the difference between competitive advantage and playing catch-up.

Trust, as it turns out, scales differently than content. The teams who treat trust as a design input now will own the next phase of search.

More Resources:


This post was originally published on Duane Forrester Decodes.


Featured Image: Roman Samborskyi/Shutterstock

Google Gemini Adds Personalization From Past Chats via @sejournal, @MattGSouthern

Google is rolling out updates to the Gemini app that personalize responses using past conversations and add new privacy controls, including a Temporary Chat mode.

The changes start today and will expand over the coming weeks.

What’s New

Personalization From Past Chats

Gemini now references earlier chats to recall details and preferences, making responses feel like collaborating with a partner who’s already familiar with the context.

The update aligns with Google’s I/O vision for an assistant that learns and understand the user.

Screenshot from: blog.google/products/gemini/temporary-chats-privacy-controls/, August 2025.

The setting is on by default and can be turned off in SettingsPersonal contextYour past chats with Gemini.

Temporary Chats

For conversations that shouldn’t influence future responses, Google is adding Temporary Chat.

As Google describes it:

“There may be times when you want to have a quick conversation with the Gemini app without it influencing future chats.”

Temporary chats don’t appear in recent chats, aren’t used to personalize or train models, and are kept for up to 72 hours.

Screenshot from: blog.google/products/gemini/temporary-chats-privacy-controls/, August 2025.

Rollout starts today and will reach all users over the coming weeks.

Updated Privacy Controls

Google will rename the “Gemini Apps Activity” setting to “Keep Activity” in the coming weeks.

When this setting is on, a sample of future uploads, such as files and photos, may be used to help improve Google services.

If your Gemini Apps Activity setting is currently off, Keep Activity will remain off. You can also turn the setting off at any time or use Temporary Chats.

Why This Matters

Personalized responses can reduce repetitive context-setting once Gemini understands your typical topics and goals.

For teams working across clients and categories, Temporary Chats help keep sensitive brainstorming separate from your main context, avoiding cross-pollination of preferences.

Both features include controls that meet privacy requirements for client-sensitive workflows.

Availability

The personalization setting begins rolling out today on Gemini 2.5 Pro in select countries, with expansion to 2.5 Flash and more regions in the coming weeks.


Featured Image: radithyaraf/Shutterstock

OpenAI Brings GPT-4o Back For Paid ChatGPT Users via @sejournal, @MattGSouthern

OpenAI has restored GPT-4o to the ChatGPT model picker for paid accounts and says it will give advance notice before removing models in the future.

The company made the change after pushback over GPT-5’s rollout and confirmed it alongside new speed controls for GPT-5 that let you choose Auto, Fast, or Thinking.

What’s New

GPT-4o Returns

If you are on a paid plan, GPT-4o now appears in the model picker by default.

You can also reveal additional options in Settings by turning on Show additional models, which exposes legacy models such as o3, o4-mini, and GPT-4.1 on Plus and Team, and adds GPT-4.5 on Pro.

This addresses the concern that model choices disappeared without warning during the initial GPT-5 launch.

New GPT-5 Modes

OpenAI’s mode picker lets you trade response time for reasoning depth.

CEO Sam Altman states:

“You can now choose between ‘Auto’, ‘Fast’, and ‘Thinking’ for GPT-5. Most users will want Auto, but the additional control will be useful for some people.”

Higher Capacity Thinking Mode

For heavier tasks, GPT-5 Thinking supports up to 3,000 messages per week and a 196k-token context window.

After you hit the weekly cap, chats can continue with GPT-5 Thinking mini, and OpenAI notes limits may change over time.

This helps when you are reviewing long reports, technical documents, or many content assets in one session.

Personality Updates

OpenAI says it’s working on GPT-5’s default tone to feel “warmer than the current personality but not as annoying (to most users) as GPT-4o.”

The company acknowledges the need for more per-user personality controls.

How To Use

To access the extra models: Open ChatGPT, go to Settings, then General, and enable Show additional models.

That toggles the legacy list and Thinking mini alongside GPT-5. GPT-4o is already in the picker for paid users.

Looking Ahead

OpenAI promises more notice around model availability while giving you clearer controls over speed and depth.

In practice, try Fast for quick checks, keep Auto for routine chats, and use Thinking where accuracy and multi-step reasoning matter most.

If your workflows depended on 4o’s feel, bringing it back reduces disruption while OpenAI tunes GPT-5’s personality and customization.


Featured Image: Adha Ghazali/Shuterstock

From Ranking to Reasoning: Philosophies Driving GEO Brand Presence Tools via @sejournal, @Dixon_Jones

Since the turn of the Millennium, marketers have mastered the science of search engine optimization.

We learned the “rules” of ranking, the art of the backlink, and the rhythm of the algorithm. But, the ground has shifted to generative engine optimization (GEO).

The era of the 10 blue links is giving way to the age of the single, synthesized answer, delivered by large language models (LLMs) that act as conversational partners.

The new challenge isn’t about ranking; it’s about reasoning. How do we ensure our brand is not just mentioned, but accurately understood and favorably represented by the ghost in the machine?

This question has ignited a new arms race, spawning a diverse ecosystem of tools built on different philosophies. Even the words to describe these tools are part of the battle: “GEO“, “GSE”, “AIO“, “AISEO”, just more “SEO.” The list of abbreviations continues to grow.

But, behind the tools, different philosophies and approaches are emerging. Understanding these philosophies is the first step toward moving from a reactive monitoring posture to a proactive strategy of influence.

School Of Thought 1: The Evolution Of Eavesdropping – Prompt-Based Visibility Monitoring

The most intuitive approach for many SEO professionals is an evolution of what we already know: tracking.

This category of tools essentially “eavesdrops” on LLMs by systematically testing them with a high volume of prompts to see what they say.

This school has three main branches:

The Vibe Coders

It is not hard, these days, to create a program that simply runs a prompt for you and stores the answer. There are myriad weekend keyboard warriors with offerings.

For some, this may be all you need, but the concern would be that these tools do not have a defensible offering. If everyone can do it, how do you stop everyone from building their own?

The VC Funded Mention Trackers

Tools like Peec.ai, TryProfound, and many more focus on measuring a brand’s “share of voice” within AI conversations.

They track how often a brand is cited in response to specific queries, often providing a percentage-based visibility score against competitors.

TryProfound adds another layer by analyzing hundreds of millions of user-AI interactions, attempting to map the questions people are asking, not just the answers they receive.

This approach provides valuable data on brand awareness and presence in real-world use cases.

The Incumbents’ Pivot

The major players in SEO – Semrush, Ahrefs, seoClarity, Conductor – are rapidly augmenting their existing platforms. They are integrating AI tracking into their familiar, keyword-centric dashboards.

With features like Ahrefs’ Brand Radar or Semrush’s AI Toolkit, they allow marketers to track their brand’s visibility or mentions for their target keywords, but now within environments like Google’s AI Overviews, ChatGPT, or Perplexity.

This is a logical and powerful extension of their current offerings, allowing teams to manage SEO and what many are calling generative engine optimization (GEO) from a single hub.

The core value here is observational. It answers the question, “Are we being talked about?” However, it’s less effective at answering “Why?” or “How do we change the conversation?”.

I have also done some maths on how many queries a database might need to be able to have enough prompt volume to be statistically useful and (with the help of Claude) came up with a database requirement of 1-5 billion prompt responses.

This, if achievable, will certainly have cost implications that are already reflected in the offerings.

School Of Thought 2: Shaping The Digital Soul – Foundational Knowledge Analysis

A more radical approach posits that tracking outputs is like trying to predict the weather by looking out the window. To truly have an effect, you must understand the underlying atmospheric systems.

This philosophy isn’t concerned with the output of any single prompt, but with the LLM’s foundational, internal “knowledge” about a brand and its relationship to the wider world.

GEO tools in this category, most notably Waikay.io and, increasingly, Conductor, operate on this deeper level. They work to map the LLM’s understanding of entities and concepts.

As an expert in Waikay’s methodology, I can detail the process, which provides the “clear bridge” from analysis to action:

1. It Starts With A Topic, Not A Keyword

The analysis begins with a broad business concept, such as “Cloud storage for enterprise” or “Sustainable luxury travel.”

2. Mapping The Knowledge Graph

Waikay uses its own proprietary Knowledge Graph and Named Entity Recognition (NER) algorithms to first understand the universe of entities related to that topic.

What are the key features, competing brands, influential people, and core concepts that define this space?

3. Auditing The LLM’s Brain

Using controlled API calls, it then queries the LLM to discover not just what it says, but what it knows.

Does the LLM associate your brand with the most important features of that topic? Does it understand your position relative to competitors? Does it harbor factual inaccuracies or confuse your brand with another?

4. Generating An Action Plan

The output isn’t a dashboard of mentions; it’s a strategic roadmap.

For example, the analysis might reveal: “The LLM understands our competitor’s brand is for ‘enterprise clients,’ but sees our brand as ‘for small business,’ which is incorrect.”

The “clear bridge” is the resulting strategy: to develop and promote content (press releases, technical documentation, case studies) that explicitly and authoritatively forges the entity association between your brand and “enterprise clients.”

This approach aims to permanently upgrade the LLM’s core knowledge, making positive and accurate brand representation a natural outcome across a near-infinite number of future prompts, rather than just the ones being tracked.

The Intellectual Divide: Nuances And Necessary Critiques

A non-biased view requires acknowledging the trade-offs. Neither approach is a silver bullet.

The Prompt-Based method, for all its data, is inherently reactive. It can feel like playing a game of “whack-a-mole,” where you’re constantly chasing the outputs of a system whose internal logic remains a mystery.

The sheer scale of possible prompts means you can never truly have a complete picture.

Conversely, the Foundational approach is not without its own valid critiques:

  • The Black Box Problem: Where proprietary data is not public, the accuracy and methodology are not easily open to third-party scrutiny. Clients must trust that the tool’s definition of a topic’s entity-space is correct and comprehensive.
  • The “Clean Room” Conundrum: This approach primarily uses APIs for its analysis. This has the significant advantage of removing the personalization biases that a logged-in user experiences, providing a look at the LLM’s “base” knowledge. However, it can also be a weakness. It may lose focus on the specific context of a target audience, whose conversational history and user data can and do lead to different, highly personalized AI outputs.

Conclusion: The Journey From Monitoring To Mastery

The emergence of these generative engine optimization tools signals a critical maturation in our industry.

We are moving beyond the simple question of “Did the AI mention us?” to the far more sophisticated and strategic question of “Does the AI understand us?”

Choosing a tool is less important than understanding the philosophy you’re buying into.

A reactive, monitoring strategy may be sufficient for some, but a proactive strategy of shaping the LLM’s core knowledge is where the durable competitive advantage will be forged.

The ultimate goal is not merely to track your brand’s reflection in the AI’s output, but to become an indispensable part of the AI’s digital soul.

More Resources:


Featured Image: Rawpixel.com/Shutterstock