What Are Google’s Core Topicality Systems? via @sejournal, @martinibuster

Topicality in relation to search ranking algorithms has become of interest for SEO after a recent Google Search Off The Record podcast mentioned the existence of Core Topicality Systems as a part of the ranking algorithms, so it may be useful to think about what those systems could be and what it means for SEO.

Not much is known about what could be a part of those core topicality systems but it is possible to infer what those systems are. Google’s documentation for their commercial cloud search offers a definition of topicality that while it’s not in the context of their own search engine it still provides a useful idea of what Google might mean when it refers to Core Topicality Systems.

This is how that cloud documentation defines topicality:

“Topicality refers to the relevance of a search result to the original query terms.”

That’s a good explanation of the relationship of web pages to search queries in the context of search results. There’s no reason to make it more complicated than that.

How To Achieve Relevance?

A starting point for understanding what might be a component of Google’s Topicality Systems is to start with how search engines understand search queries and represent topics in web page documents.

  • Understanding Search Queries
  • Understanding Topics

Understanding Search Queries

Understanding what users mean can be said to be about understanding the topic a user is interested in. There’s a taxonomic quality to how people search in that a search engine user might use an ambiguous query when they really mean something more specific.

The first AI system Google deployed was RankBrain, which was deployed to better understand the concepts inherent in search queries. The word concept is broader than the word topic because concepts are abstract representations. A system that understands concepts in search queries can then help the search engine return relevant results on the correct topic.

Google explained the job of RankBrain like this:

“RankBrain helps us find information we weren’t able to before by more broadly understanding how words in a search relate to real-world concepts. For example, if you search for “what’s the title of the consumer at the highest level of a food chain,” our systems learn from seeing those words on various pages that the concept of a food chain may have to do with animals, and not human consumers. By understanding and matching these words to their related concepts, RankBrain understands that you’re looking for what’s commonly referred to as an “apex predator.”

BERT is a deep learning model that helps Google understand the context of words in queries to better understand the overall topic the text.

Understanding Topics

I don’t think that modern search engines use Topic Modeling anymore because of deep learning and AI. However, a statistical modeling technique called Topic Modeling was used in the past by search engines to understand what a web page is about and to match it to search queries. Latent Dirichlet Allocation (LDA) was a breakthrough technology around the mid 2000s that helped search engines understand topics.

Around 2015 researchers published papers about the Neural Variational Document Model (NVDM), which was an even more powerful way to represent the underlying topics of documents.

One of the most latest research papers is one called, Beyond Yes and No: Improving Zero-Shot LLM Rankers via Scoring Fine-Grained Relevance Labels. That research paper is about enhancing the use of Large Language Models to rank web pages, a process of relevance scoring. It involves going beyond a binary yes or no ranking to a more precise way using labels like “Highly Relevant”, “Somewhat Relevant” and “Not Relevant”

This research paper states:

“We propose to incorporate fine-grained relevance labels into the prompt for LLM rankers, enabling them to better differentiate among documents with different levels of relevance to the query and thus derive a more accurate ranking.”

Avoid Reductionist Thinking

Search engines are going beyond information retrieval and have been (for a long time) moving in the direction of answering questions, a situation that has accelerated in recent years and months.  This was predicted in 2001 paper that titled,  Rethinking Search: Making Domain Experts out of Dilettantes where they proposed the necessity to engage fully in returning human-level responses.

The paper begins:

“When experiencing an information need, users want to engage with a domain expert, but often turn to an information retrieval system, such as a search engine, instead. Classical information retrieval systems do not answer information needs directly, but instead provide references to (hopefully authoritative) answers. Successful question answering systems offer a limited corpus created on-demand by human experts, which is neither timely nor scalable. Pre-trained language models, by contrast, are capable of directly generating prose that may be responsive to an information need, but at present they are dilettantes rather than domain experts – they do not have a true understanding of the world…”

The major takeaway is that it’s self-defeating to apply reductionist thinking to how Google ranks web pages by doing something like putting an exaggerated emphasis on keywords, on title elements and headings. The underlying technologies are rapidly moving to understanding the world, so if one is to think about Core Topicality Systems then it’s useful to put that into a context that goes beyond the traditional “classical” information retrieval systems.

The methods Google uses to understand topics on web pages that match search queries are increasingly sophisticated and it’s a good idea to get acquainted with the ways Google has done it in the past and how they may be doing it in the present.

Featured Image by Shutterstock/Cookie Studio

Google Expands ‘About This Image’ To More Platforms via @sejournal, @MattGSouthern

Google has announced the expansion of its “About this image” feature to additional platforms, including Circle to Search and Google Lens.

This move gives people more access points to obtain context about images they encounter online.

New Access Points

The “About this image” tool, which offers information about an image’s origins and usage, is now available through:

  1. Circle to Search: A feature on select Android devices
  2. Google Lens: Available in the Google app on both Android and iOS

Functionality & Usage

You can access the feature through different methods depending on the platform:

For Circle to Search:

  • Activate the feature by long-pressing the home button or navigation bar
  • Circle or tap the image on the screen
  • Swipe up on search results and select the “About this image” tab

For Google Lens:

  • Screenshot or download the image
  • Open the Google app and use the Lens icon
  • Select the image and tap the “About this image” tab

Information Provided

The tool offers various details about images, including:

  • How other websites use and describe the image
  • Available metadata
  • Identification of AI-generated images with specific watermarks

Availability & Language Support

“About this image” is available in 40 languages globally, including French, German, Hindi, Italian, Japanese, Korean, Portuguese, Spanish, and Vietnamese.

Broader Context

This expansion comes at a time when digital literacy and the ability to verify online information are increasingly important.

However, it’s worth noting that while such tools can be helpful, they’re not infallible.

Users are still encouraged to critically evaluate information and consult multiple sources when verifying claims or images online.

How Does This Help You?

Here’s how the expansion of Google’s “About this image” feature can help you:

  • Quickly verify claims associated with images.
  • Understand where an image originated and how it’s been used across the web.
  • This tool can help you distinguish between human-created and AI-created visual content.
  • It provides a quick way for students, journalists, and researchers to gather context and potential sources related to an image.
  • Understanding an image’s history and context can help protect you from visual manipulation tactics often used in scams.

Related Algorithm Update: Combating Explicit Deepfakes

Today, Google announced an algorithm update targeting explicit deepfakes in search results.

Key aspects of this update include:

  1. Improved Content Removal: When a removal request is approved, the system will attempt to filter similar explicit results across related searches for the affected individual.
  2. Ranking Adjustments: The search algorithm has been modified to reduce the visibility of explicit fake content in many searches. For queries seeking such content and including people’s names, Google will prioritize non-explicit content, such as news articles.
  3. Site-Wide Impact: Websites with numerous pages removed due to fake explicit imagery may see changes in their overall search rankings.

Google reports that these changes have reduced exposure to explicit image results for specific queries, decreasing over 70% for targeted searches.

Google’s doing two things at once: making it easier to spot fake images and cracking down on deepfakes algorithmically.

These updates demonstrate Google’s commitment to keeping search results safe and trustworthy as the web changes.


Featured Image: Screenshot from blog.google.com, July 2024. 

AI Integration in Marketing: Strategic Insights For SEO & Agency Leaders via @sejournal, @CallRail

This edited extract is from Data Storytelling in Marketing by Caroline Florence ©2024 and is reproduced and adapted with permission from Kogan Page Ltd.

Storytelling is an integral part of the human experience. People have been communicating observations and data to each other for millen­nia using the same principles of persuasion that are being used today.

However, the means by which we can generate data and insights and tell stories has shifted significantly and will continue to do so, as tech­nology plays an ever-greater role in our ability to collect, process, and find meaning from the wealth of information available.

So, what is the future of data storytelling?

I think we’ve all talked about data being the engine that powers business decision-making. And there’s no escaping the role that AI and data are going to play in the future.

So, I think the more data literate and aware you are, the more informed and evidence-led you can be about our decisions, regardless of what field you are in – because that is the future we’re all working towards and going to embrace, right?

It’s about relevance and being at the forefront of cutting-edge technology.

Sanica Menezes, Head of Customer Analytics, Aviva

The Near Future Scenario

Imagine simply applying a generative AI tool to your marketing data dashboards to create audience-ready copy. The tool creates a clear narrative structure, synthesized from the relevant datasets, with actionable and insightful messages relevant to the target audience.

The tool isn’t just producing vague and generic output with question­able accuracy but is sophisticated enough to help you co-author technically robust and compelling content that integrates a level of human insight.

Writing stories from vast and complex datasets will not only drive efficiency and save time, but free up the human co-author to think more creatively about how they deliver the end story to land the message, gain traction with recommendations and influence decisions and actions.

There is still a clear role for the human to play as co-author, including the quality of the prompts given, expert interpretation, nuance of language, and customization for key audiences.

But the human co-author is no longer bogged down by the complex and time-consuming process of gathering different data sources and analysing data for insights. The human co-author can focus on synthesizing findings to make sense of patterns or trends and perfect their insight, judgement, and communication.

In my conversations with expert contributors, the consensus was that AI would have a significant impact on data storytelling but would never replace the need for human intervention.

This vision for the future of storytelling is (almost) here. Tools like this already exist and are being further improved, enhanced, and rolled out to market as I write this book.

But the reality is that the skills involved in leveraging these tools are no different from the skills needed to currently build, create, and deliver great data stories. If anything, the risks involved in not having human co-authors means acquiring the skills covered in this book become even more valuable.

In the AI storytelling exercise WINconducted, the tool came up with “80 per cent of people are healthy” as its key point. Well, it’s just not an interesting fact.

Whereas the humans looking at the same data were able to see a trend of increasing stress, which is far more interesting as a story. AI could analyse the data in seconds, but my feeling is that it needs a lot of really good prompting in order for it to seriously help with the storytelling bit.

I’m much more positive about it being able to create 100 slides for me from the data and that may make it easier for me to pick out what the story is.

Richard Colwell, CEO, Red C Research & Marketing Group

We did a recent experiment with the Inspirient AI platform taking a big, big, big dataset, and in three minutes, it was able to produce 1,000 slides with decent titles and design.

Then you can ask it a question about anything, and it can produce 110 slides, 30 slides, whatever you want. So, there is no reason why people should be wasting time on the data in that way.

AI is going to make a massive difference – and then we bring in the human skill which is contextualization, storytelling, thinking about the impact and the relevance to the strategy and all that stuff the computer is never going to be able to do.

Lucy Davison, Founder And CEO, Keen As Mustard Marketing

Other Innovations Impacting On Data Storytelling

Besides AI, there are a number of other key trends that are likely to have an impact on our approach to data storytelling in the future:

Synthetic Data

Synthetic data is data that has been created artificially through computer simulation to take the place of real-world data. Whilst already used in many data models to supplement real-world data or when real-world data is not available, the incidence of synthetic data is likely to grow in the near future.

According to Gartner (2023), by 2024, 60 per cent of the data used in training AI models will be synthetically generated.

Speaking in Marketing Week (2023), Mark Ritson cites around 90 per cent accuracy for AI-derived consumer data, when triangulated with data generated from primary human sources, in academic studies to date.

This means that it has a huge potential to help create data stories to inform strategies and plans.

Virtual And Augmented Reality

Virtual and augmented reality will enable us to generate more immersive and interactive experiences as part of our data storytelling. Audiences will be able to step into the story world, interact with the data, and influence the narrative outcomes.

This technology is already being used in the world of entertainment to blur the lines between traditional linear television and interactive video games, creating a new form of content consumption.

Within data storytelling we can easily imagine a world with simulated customer conversations, whilst navigating the website or retail environment.

Instead of static visualizations and charts showing data, the audience will be able to overlay data onto their physical environment and embed data from different sources accessed at the touch of a button.

Transmedia Storytelling

Transmedia storytelling will continue to evolve, with narratives spanning multiple platforms and media. Data storytellers will be expected to create interconnected storylines across different media and channels, enabling audiences to engage with the data story in different ways.

We are already seeing these tools being used in data journalism where embedded audio and video, on-the-ground eyewitness content, live-data feeds, data visualization and photography sit alongside more traditional editorial commentary and narrative storytelling.

For a great example of this in practice, look at the Pulitzer Prize-winning “Snow fall: The avalanche at Tunnel Creek (Branch, 2012)” that changed the way The New York Times approached data storytelling.

In the marketing world, some teams are already investing in high-end knowledge share portals or embedding tools alongside their intranet and internet to bring multiple media together in one place to tell the data story.

User-Generated Content

User-generated content will also have a greater influence on data storytelling. With the rise of social media and online communities, audiences will actively participate in creating and sharing stories.

Platforms will emerge that enable collaboration between storytellers and audiences, allowing for the co-creation of narratives and fostering a sense of community around storytelling.

Tailoring narratives to the individual audience member based on their preferences, and even their emotional state, will lead to greater expectations of customization in data storytelling to enhance engagement and impact.

Moving beyond the traditional “You said, so we did” communication with customers to demonstrate how their feedback has been actioned, user-generated content will enable customers to play a more central role in sharing their experiences and expectations

These advanced tools are a complement to, and not a substitution for, the human creativity and critical thinking that great data storytelling requires. If used appropriately, they can enhance your data storytelling, but they cannot do it for you.

Whether you work with Microsoft Excel or access reports from more sophisticated business intelligence tools, such as Microsoft Power BI, Tableau, Looker Studio, or Qlik, you will still need to take those outputs and use your skills as a data storyteller to curate them in ways that are useful for your end audi­ence.

There are some great knowledge-sharing platforms out there that can integrate outputs from existing data storytelling tools and help curate content in one place. Some can be built into existing plat­forms that might be accessible within your business, like Confluence.

Some can be custom-built using external tools for a bespoke need, such as creating a micro-site for your data story using WordPress. And some can be brought in at scale to integrate with existing Microsoft or Google tools.

The list of what is available is extensive but will typically be dependent on what is available IT-wise within your own organization.

The Continuing Role Of The Human In Data Storytelling

In this evolving world, the role of the data storyteller doesn’t disap­pear but becomes ever more critical.

The human data storyteller still has many important roles to still play, and the skills necessary to influence and engage cynical, discerning, and overwhelmed audiences become even more valuable.

Now that white papers, marketing copy, internal presentations, and digital content can all be generated faster than humans could ever manage on their own, the risk of informa­tion overload becomes inevitable without a skilled storyteller to curate the content.

Today, the human data storyteller is crucial for:

  • Ensuring we are not telling “any old story” just because we can and that the story is relevant to the business context and needs.
  • Understanding the inputs being used by the tool, including limitations and potential bias, as well as ensuring data is used ethically and that it is accurate, reliable, and obtained with the appropriate permissions.
  • Framing queries appropriately in the right way to incorporate the relevant context, issues, and target audience needs to inform the knowledge base.
  • Cross-referencing and synthesizing AI-generated insights or synthetic data with human expertise and subject domain knowledge to ensure the relevance and accuracy of recommendations.
  • Leveraging the different VR, AR, and transmedia tools available to ensure the right one for the job.

To read the full book, SEJ readers have an exclusive 25% discount code and free shipping to the US and UK. Use promo code SEJ25 at koganpage.com here.

More resources: 


Featured Image: PopTika/Shutterstock

Search GPT – Can Search GPT Disrupt Google Search? via @sejournal, @Kevin_Indig

Despite initial concerns, Chat GPT has not replaced search. Q2 record earnings show Google Search does better than ever. That’s why OpenAI’s new search engine, Search GPT, makes only sense after a second look.

Boost your skills with Growth Memo’s weekly expert insights. Subscribe for free!

$5b USD

Why would OpenAI launch a search engine if its main product poses one of the biggest threats to Google?

Image Credit: Kevin Indig

Searches for “LLM Search” are growing, but it’s not consumer demand that pulls OpenAI in that direction. There are six good reasons (in order of importance):

1/ OpenAI’s problem is that Chat GPT is not perceived as a search engine despite similar capabilities, so the company positions Search GPT as a direct Google alternative to gain more Search market share.

Rumors about launching a search engine just before Google I/O in 2024 and the impact of the actual announcement on Alphabet’s stock show the ambition to compete directly.

The Information reports that OpenAI loses $5b a year in expenses.1 Just capturing 3% of Google’s $175b Search business would allow OpenAI to recoup expenses.

Image Credit: Kevin Indig

Searches for ChatGPT on Google are growing so much, they get close to searches for “Google”. They’ve already surpassed searches for other search engines by a lot.

To be fair, people search less for “Google” on Google (maybe in their browser bar to get to the Google homepage), and traffic numbers between Google (465b, according to Similarweb) and Chat GPT (660M) are still magnitudes apart.

Image Credit: Kevin Indig

OpenAI has a strategic advantage over Google: Search GPT can provide a very different, maybe less noisy, user experience than Google because it’s not reliant on ad revenue. In any decision regarding Search, Google needs to take ads into account.

2/ OpenAI crawls the web for training data and already has half the ingredients for a search engine on the table. Consumers are already familiar with the concept of a search engine, making adoption more likely.

I have no doubt that OpenAI will see a lot of curious sign-ups for Search GPT but the bigger challenge will be retaining users.

It’s also important to point out that the market hasn’t found the final form of LLMs yet. Chatbots made sense because of their prompting nature, but voice devices will likely become much better devices for LLMs.

3/ Search can deliver better user signals than prompting because it’s a more specific use case.

The beauty of prompting is that it’s an open field. You can do whatever you want. But that’s also a disadvantage because most people have no idea what they want to do and where to start.

As a result, success and failure are harder to measure at scale for chatbots than search engines.

A search engine, despite being versatile, has clearer use cases, which could drive more adoption and deliver better signals for LLMs to learn. In return, those learnings could transfer to chatbot answers, which are a big part of Search GPT.

4/ OpenAI wants to throw publishers a lifeline to secure a content pipeline. LLM developers need fresh content to train models and serve timely answers.

Search is the biggest source of publisher traffic2, but publishers are growing more frustrated with Google due to Algorithm updates, site reputation abuse penalties and AI Overviews.

It’s good timing for OpenAI to offer another source of revenue and get publishers “on their side”, especially after OpenAI itself has received a lot of criticism from publishers and a lawsuit from the NY Times.

The launch of SearchGPT follows a long list of publisher licensing deals:

  1. News Corp (+$250 million over five years): WSJ, New York Post, The Times, The Sun
  2. Associated Press (AP)
  3. Axel Springer: Bild, Politico, Business Insider
  4. Financial Times
  5. Le Monde
  6. Reuters
  7. Prisa Media
  8. Dotdash Meredith
  9. Time magazine
  10. Vox media
  11. Wiley (one-time fee of $23 million for previously published academic articles and books)

But even the best deals don’t help if publishers cannot sustain the creation of fresh content. If Search GPT can become a new traffic and revenue source for publishers, it would be a way to keep the critical ecosystem alive and get on the good side of publishers.

5/ Perplexity is a small challenger to OpenAI, but even a small challenger can take away mind share, and you never want to underestimate the competition. A search engine would conveniently fence in their growth. Why use Perplexity when Search GPT, which looks very similar, can do the same thing?

6/ OpenAI might bet on regulators breaking up Google’s exclusive search engine deal with Apple and hope to become part of a search engine choice set on Apple devices.

Granted, we’re talking about a very small chance, and certainly not the decisive factor for building a search engine, but it could be a small factor nonetheless.

Publisher GPT

Search GPT is clearly the sibling of Chat GPT. Besides SERP Features like weather charts and table stakes features like auto-suggest, the experience feels like Chat GPT.

The differences are hard to spot at first but meaningful in their potential to drive revenue, compete with Google and strengthen OpenAI’s data mining.

But one change stands out: Search GPT has more pronounced links to web results, a clear hat tip to publishers.

The Search GPT landing page mentions the word publisher 14 times and underlines how important publishers are for the open web and how dedicated OpenAI is to working with them.

OpenAI uses a different user agent to crawl websites for its search engine than for LLM training and strongly separates the two.

Importantly, SearchGPT is about search and is separate from training OpenAI’s generative AI foundation models. Sites can be surfaced in search results even if they opt out of generative AI training.

It’s not an accident that OpenAI tries to regain its grip on the web. A recently published study3 found that 25% of words (tokens) in Common Crawl stem from domains that have now excluded AI crawlers, with OpenAI at the top of the list, in their robots.txt or ToS.

SEO Implications

The two questions every SEO is asking themselves is whether they should care about Search GPT and how it might work.

Search GPT has a chance to become relevant for SEO quickly, given Chat GPT’s adoption. The Apple Intelligence integration and a potential phone would spur adoption even more.

However, OpenAI might integrate Search GPT into Chat GPT, which could change the relevance as a traffic source.

We cannot yet know how Search GPT works because it’s not live, but one big differentiator will be whether Search GPT includes results from the broad web or only from publishers OpenAI made a deal with.

If it’s the broad web, Search GPT has a high chance of being relevant. If it’s limited to partnering publishers, SEO won’t make sense for anyone not a partner because the answer set is limited.

If Search GPT uses RAG and ranks results similar to Google’s AI Overviews, we could use AIO performance as indicator and predictor for SearchGPT performance.

There is also chance that an answer from Chat GPT for queries that don’t require QDF (query deserves freshness) is the same on Search GPT, which would give us a way to understand what works before Search GPT launches publicly. Hard to validate without access Search GPT, though.

Search GPT could gain the web’s favor by sending relevant traffic, making it easy for sites to submit content, for example, through XML sitemaps, and providing some sort of webmaster console. As a result, Search GPT would position itself even stronger against Google.

A New Way To Search

If the main benefit or Search GPT for OpenAI is a revenue stream and access to more user data, the next logical step for OpenAI is to build a (AI-powered) browser.

Browser data is incredibly valuable for understanding user behavior, personalization and LLM training. Best of all, it’s app-agnostic, so OpenAI could learn from users even when they use Perplexity or Google.

We’ve seen the power of browser data in the Google lawsuit, where it turned out Google relied on Chrome data all along for ranking. The only layer that’s more powerful is the operating system and device layer.

Image Credit: Kevin Indig

There is already news that Sam Altman is working with Jon Ivy on building a phone. No wonder since Apple holds immense power over other ecosystems and platforms.

Remember when Apple blew a $10b hole into Meta’s annual revenue? Apple could develop its own models and surface them on the OS level—a critical threat to OpenAI. A browser could alleviate at least some of that threat.

Bing released its own update to Search, giving us an idea of what Search GPT could look like. The new Bing prominently features AI answers at the top and search results on the side. A fitting metaphor for classic blue links.

Image Credit: Kevin Indig

Why OpenAI Could Lose $5 Billion This Year

Who Sends Traffic on the Web and How Much? New Research from Datos & SparkToro

Consent in Crisis: The Rapid Decline of the AI Data Commons

Cracking Open The Local SEO Bucket: Expert Strategies To Shape Success via @sejournal, @hethr_campbell

Are your local SEO efforts yielding the results you expected?

This is the perfect time to get a closer look at the strategies that drive real impact in local search rankings and user experience for the big brands. 

Local SEO isn’t just about appearing in local search results; it’s about ensuring your brand is visible, relevant, and engaging to your community. In our upcoming webinar, we’ll break down five key strategies that can significantly boost your local visibility and drive more foot traffic to your business.

On July 24, join industry expert Matt Coghlan, Director of SEO Partnerships at Uberall as he shares practical insights and proven tactics that have led to success for big brands like KFC.

Get actionable tips to help navigate the dynamic landscape of local SEO in 5 core areas:

  1. Discovery: Learn how to ensure your business is easily found by local customers.
  2. Relevance: Discover how to make your business stand out by personalizing your content and SEO strategies to meet local search goals.
  3. Experience: Improve user experience to keep visitors engaged and satisfied.
  4. Engagement: Boost engagement through reviews, social media, and local content.
  5. Conversions: Convert local traffic into paying customers with effective CTAs and optimized landing pages.

We will also delve into:

  • KFC’s Proven Strategies: Gain insights into how KFC enhanced their local SEO to drive significant results.
  • Quick Visibility Wins: Learn tactics that can immediately boost your online presence.
  • Adapting to Changes: Understand the recent shifts in local search and how to adjust your strategies accordingly.

Whether you’re a seasoned SEO professional or just starting out, this webinar is packed with info that will help you master the dynamics of local SEO.

Stay until the end for a live Q&A session where you can get your specific questions answered by our expert.

Can’t make it to the live session? No worries! Register now, and we’ll send you the recording so you won’t miss out.

Strengthen your local SEO game, enhance your digital presence, and drive impactful results for your business. Sign up today!

Google Hints Lowering SEO Value Of Country Code Top-Level Domains via @sejournal, @MattGSouthern

In a recent episode of Google’s Search Off The Record podcast, the company’s Search Relations team hinted at potential changes in how country-code top-level domains (ccTLDs) are valued for SEO.

This revelation came during a discussion on internationalization and hreflang implementation.

The Fading Importance Of ccTLDs

Gary Illyes, a senior member of Google’s Search Relations team, suggested that the localization boost traditionally associated with ccTLDs may soon be over.

Illyes stated:

“I think eventually, like in years’ time, that [ccTLD benefit] will also fade away.”

He explained that ccTLDs are becoming less reliable indicators of a website’s geographic target audience.

Creative Use Of ccTLDs For Branding

According to Illyes, the primary reason for this shift is the creative use of ccTLDs for branding purposes rather than geographic targeting.

He elaborated:

“Think about the all the funny domain names that you can buy nowadays like the .ai. I think that’s Antigua or something… It doesn’t say anything anymore about the country… it doesn’t mean that the content is for the country.”

Illyes further explained the historical context and why this change is occurring:

“One of the main algorithms that do the whole localization thing… is called something like LDCP – language demotion country promotion. So basically if you have like a .de, then for users in Germany you would get like a slight boost with your .de domain name. But nowadays, with .co or whatever .de, which doesn’t relate to Germany anymore, it doesn’t really make sense for us to like automatically apply that little boost because it’s ambiguous what the target is.”

The Impact On SEO Strategies

This change in perspective could have implications for international SEO strategies.

Traditionally, many businesses have invested in ccTLDs to gain a perceived advantage in local search results.

If Google stops using ccTLDs as a strong signal for geographic relevance, this could alter how companies approach their domain strategy for different markets.

Marketing Value Of ccTLDs

However, Illyes also noted that from a marketing perspective, there might still be some value in purchasing ccTLDs:

“I think from a marketing perspective there’s still some value in buying the ccTLDs and if I… if I were to run some… like a new business, then I would try to buy the country TLDs when I can, when like it’s monetarily feasible, but I would not worry too much about it.”

What This Means For You

As search engines become more capable of understanding content and context, traditional signals like ccTLDs may carry less weight.

This could lead to a more level playing field for websites, regardless of their domain extension.

Here are some top takeaways:

  1. If you’ve invested heavily in country-specific domains for SEO purposes, it may be time to reassess this strategy.
  2. Should the importance of ccTLDs decrease, proper implementation of hreflang tags becomes crucial for indicating language and regional targeting.
  3. While the SEO benefits may diminish, ccTLDs can still have branding and marketing value.
  4. Watch for official announcements or changes in Google’s documentation regarding using ccTLDs and international SEO best practices.

While no immediate changes were announced, this discussion provides valuable insight into the potential future direction of international SEO.

Listen to the full podcast episode below:

Google Advises Caution With AI Generated Answers via @sejournal, @martinibuster

Google’s Gary Illyes cautioned about the use of Large Language Models (LLMs), affirming the importance of checking authoritative sources before accepting any answers from an LLM. His answer was given in the context of a question, but curiously, he didn’t publish what that question was.

LLM Answer Engines

Based on what Gary Illyes said, it’s clear that the context of his recommendation is the use of AI for answering queries. The statement comes in the wake of OpenAI’s announcement of SearchGPT that they are testing an AI Search Engine prototype. It may be that his statement is not related to that announcement and is just a coincidence.

Gary first explained how LLMs craft answers to questions and mentions how a technique called “grounding” can improve the accuracy of the AI generated answers but that it’s not 100% perfect, that mistakes still slip through. Grounding is a way to connect a database of facts, knowledge, and web pages to an LLM. The goal is to ground the AI generated answers to authoritative facts.

This is what Gary posted:

“Based on their training data LLMs find the most suitable words, phrases, and sentences that align with a prompt’s context and meaning.

This allows them to generate relevant and coherent responses. But not necessarily factually correct ones. YOU, the user of these LLMs, still need to validate the answers based on what you know about the topic you asked the LLM about or based on additional reading on resources that are authoritative for your query.

Grounding can help create more factually correct responses, sure, but it’s not perfect; it doesn’t replace your brain. The internet is full of intended and unintended misinformation, and you wouldn’t believe everything you read online, so why would you LLM responses?

Alas. This post is also online and I might be an LLM. Eh, you do you.”

AI Generated Content And Answers

Gary’s LinkedIn post is a reminder that LLMs generate answers that are contextually relevant to the questions that are asked but that contextual relevance isn’t necessarily factually accurate.

Authoritativeness and trustworthiness is an important quality of the kind of content Google tries to rank. Therefore it is in publishers best interest to consistently fact check content, especially AI generated content, in order to avoid inadvertently becoming less authoritative. The need to verify facts also holds true for those who use generative AI for answers.

Read Gary’s LinkedIn Post:

Answering something from my inbox here

Featured Image by Shutterstock/Roman Samborskyi

5 Automated And AI-Driven Workflows To Scale Enterprise SEO via @sejournal, @seomeetsdesign

That’s where Ahrefs’ in-built AI translator may be a better fit for your project, solving both problems in one go:

GIF from Ahrefs Keywords Explorer, July 2024

It offers automatic translations for 40+ languages and dialects in 180+ countries, with more coming soon.

However, the biggest benefit is that you’ll get a handful of alternative translations to select from, giving you greater insight into the nuances of how people search in local markets.

For example, there are over a dozen ways to say ‘popcorn’ across all Spanish-speaking countries and dialects. The AI translator is able to detect the most popular variation in each country.

Screenshot from Ahrefs Keywords Explorer, July 2024

This, my friends, is quality international SEO on steroids.

2.   Identify The Dominant Search Intent Of Any Keyword

Search intent is the internal motivator that leads someone to look for something online. It’s the reason why they’re looking and the expectations they have about what they’d like to find.

The intent behind many keywords is often obvious. For example, it’s not rocket science to infer that people expect to purchase a product when searching any of these terms:

Screenshot from Ahrefs Keywords Explorer, July 2024

However, there are many keywords where the intent isn’t quite so clear-cut.

For instance, take the keyword “waterbed.” We could try to guess its intent, or we could use AI to analyze the top-ranking pages and give us a breakdown of the type of content most users seem to be looking for.

Gif from Ahrefs Keywords Explorer, July 2024

For this particular keyword, 89% of results skew toward purchase intent. So, it makes sense to create or optimize a product page for this term.

For the keyword “arrow fletchings,” there is a mix of different types of content ranking, like informational posts, product pages, and how-to guides.

Screenshot from Ahrefs Identify Intents, July 2024

If your brand or product lent itself to one of the popular content types, that’s what you could plan in your content calendar.

Or, you could use the data here to outline a piece of content that covers all the dominant intents in a similar proportion to what’s already ranking:

  • ~40% providing information and answers to common questions.
  • ~30% providing information on fletching products and where to buy them.
  • ~20% providing a process for a reader to make their own fletchings.
  • And so on.

For enterprises, the value of outsourcing this to AI is simple. If you guess and get it wrong, you’ll have to allocate your limited SEO funds toward fixing the mistake instead of working on new content.

It’s better to have data on your side confirming the intent of any keyword before you publish content with an intent misalignment, let alone rolling it out over multiple websites or languages!

3.   Easily Identify Missing Topics Within Your Content

Topical gap analysis is very important in modern SEO. We’ve evolved well beyond the times when simply adding keywords to your content was enough to make it rank.

However, it’s not always quick or easy to identify missing topics within your content. Generative AI can help plug gaps beyond what most content-scoring tools can identify.

For example, ChatGPT can analyze your text against competitors’ to find missing topics you can include. You could prompt it to do something like the following:

Screenshot from ChatGPT, July 2024

SIDENOTE. You’ll need to add your content and competitors’ content to complete the prompt.

Here’s an example of the list of topics it identifies:

Screenshot from ChatGPT, July 2024

And the scores and analysis it can provide for your content:

Screenshot from ChatGPT, July 2024

This goes well beyond adding words and entities, like what most content scoring tools suggest.

The scores on many of these tools can easily be manipulated, providing higher scores the more you add certain terms; even if, from a conceptual standpoint, your content doesn’t do a good job of covering a topic.

If you want the detailed analysis offered by ChatGPT but available in bulk and near-instantly… then good news. We’re working on Content Master, a content grading solution that automates topic gap analysis.

I can’t reveal too much about this yet, but it has a big USP compared to most existing content optimization tools: its content score is based on topic coverage—not just keywords.

Screenshot from Ahrefs Content Master, July 2024

You can’t just lazily copy and paste related keywords or entities into the content to improve the score.

If you rely on a pool of freelancers to create content at scale for your enterprise company, this tool will provide you with peace of mind that they aren’t taking any shortcuts.

4.   Update Search Engines With Changes On Your Website As They Happen

Have you ever made a critical change on your website, but search engines haven’t picked up on it for ages? There’s now a fix for that.

If you aren’t already aware of IndexNow, it’s time to check it out.

It tells participating search engines when a change, any change, has been made on a website. If you add, update, remove, or redirect pages, participating search engines can pick up on the changes faster.

Not all search engines have adopted this yet, including Google. However, Microsoft Bing, Yandex, Naver, Seznam.cz, and Yep all have. Once one partner is pinged, all the information is shared with the other partners making it very valuable for international organizations:

Most content management systems and delivery networks already use IndexNow and will ping search engines automatically for you. However, since many enterprise websites are built on custom ERP platforms or tech stacks, it’s worth looking into whether this is happening for the website you’re managing or not.

You could partner with the dev team to implement the free IndexNow API. Ask them to try these steps as shared by Bing if your website tech stack doesn’t already use IndexNow:

  1. Get your free IndexNow API key
  2. Place the key in your site’s root directory as a .txt file
  3. Submit your key as a URL parameter
  4. Track URL discoveries by search engines

You could also use Ahrefs instead of involving developers. You can easily connect your IndexNow API directly within Site Audit and configure your desired settings.

Here’s a quick snapshot of how IndexNow works with Ahrefs:

In short, it’s an actual real-time monitoring and alerting system, a dream come true for technical SEOs worldwide. Check out Patrick Stox’s update for all the details.

Paired with our always-on crawler, no matter what changes you’re making, you can trust search engines will be notified of any changes you want, automatically. It’s the indexing shortcut you’ve been looking for.

5.   Automatically Fix Common Technical SEO Issues

Creative SEO professionals get stuff done with or without support from other departments. Unfortunately, in many enterprise organizations, relationships between the SEO team and devs can be tenuous, affecting how many technical fixes are implemented on a website.

If you’re a savvy in-house SEO, you’ll love this new enterprise feature we’re about to drop. It’s called Patches.

It’s designed to automatically fix common technical issues with the click of a button. You will be able to launch these fixes directly from our platform using Cloudflare workers or JavaScript snippets.

Picture this:

  1. You run a technical SEO crawl.
  2. You identify key issues to fix across one page, a subset of pages, or all affected pages.
  3. With the click of a button, you fix the issue across your selected pages.
  4. Then you instantly re-crawl these pages to check the fixes are working as expected.

For example, you can make page-level fixes for pesky issues like re-writing page titles, descriptions, and headings:

Screenshot from Ahrefs Site Audit, July 2024

You can also make site-wide fixes. For example, fixing internal links to broken pages can be challenging without support from developers on large sites. With Patches, you’ll be able to roll out automatic fixes for issues like this yourself:

Screenshot from Ahrefs Site Audit, July 2024

As we grow this tool, we plan to automate over 95% of technical fixes via JavaScript snippets or Cloudflare workers, so you don’t have to rely on developers as much as you may right now. We’re also integrating AI to help you speed up the process of fixing fiddly tasks even more.

Get More Buy-In For Enterprise SEO With These Workflows

Now, as exciting and helpful as these workflows may be for you, the key is to get your boss and your boss’ boss on board.

If you’re ever having trouble getting buy-in for SEO projects or budgets for new initiatives, try using the cost savings you can pass as leverage.

For instance, you can show how, usually, three engineers would dedicate five sprints to fixing a particular issue, costing the company illions of dollars—millions, billions, bajillions, whatever it is. But with your proposed solution, you can reduce costs and free up the engineers’ time to work on high-value tasks.

You can also share the Ultimate Enterprise SEO Playbook with them. It’s designed to show executives how your team is strategically valuable and can solve many other challenges within the organization.