SEO Reinvented: Responding To Algorithm Shifts via @sejournal, @pageonepower

A lot has been said about the remarkable opportunities of Generative AI (GenAI), and some of us have also been extremely vocal about the risks associated with using this transformative technology.

The rise of GenAI presents significant challenges to the quality of information, public discourse, and the general open web. GenAI’s power to predict and personalize content can be easily misused to manipulate what we see and engage with.

Generative AI search engines are contributing to the overall noise, and rather than helping people find the truth and forge unbiased opinions, they tend (at least in their present implementation) to promote efficiency over accuracy, as highlighted by a recent study by Jigsaw, a unit inside Google.

Despite the hype surrounding SEO alligator parties and content goblins, our generation of marketers and SEO professionals has spent years working towards a more positive web environment.

We’ve shifted the marketing focus from manipulating audiences to empowering them with knowledge, ultimately aiding stakeholders in making informed decisions.

Creating an ontology for SEO is a community-led effort that aligns perfectly with our ongoing mission to shape, improve, and provide directions that truly advance human-GenAI interaction while preserving content creators and the Web as a shared resource for knowledge and prosperity.

Traditional SEO practices in the early 2010s focused heavily on keyword optimization. This included tactics like keyword stuffing, link schemes, and creating low-quality content primarily intended for search engines.

Since then, SEO has shifted towards a more user-centric approach. The Hummingbird update (2013) marked Google’s transition towards semantic search, which aims to understand the context and intent behind search queries rather than just the keywords.

This evolution has led SEO pros to focus more on topic clusters and entities than individual keywords, improving content’s ability to answer multiple user queries.

Entities are distinct items like people, places, or things that search engines recognize and understand as individual concepts.

By building content that clearly defines and relates to these entities, organizations can enhance their visibility across various platforms, not just traditional web searches.

This approach ties into the broader concept of entity-based SEO, which ensures that the entity associated with a business is well-defined across the web.

Fast-forward to today, static content that aims to rank well in search engines is constantly transformed and enriched by semantic data.

This involves structuring information so that it is understandable not only by humans but also by machines.

This transition is crucial for powering Knowledge Graphs and AI-generated responses like those offered by Google’s AIO or Bing Copilot, which provide users with direct answers and links to relevant websites.

As we move forward, the importance of aligning content with semantic search and entity understanding is growing.

Businesses are encouraged to structure their content in ways that are easily understood and indexed by search engines, thus improving visibility across multiple digital surfaces, such as voice and visual searches.

The use of AI and automation in these processes is increasing, enabling more dynamic interactions with content and personalized user experiences.

Whether we like it or not, AI will help us compare options faster, run deep searches effortlessly, and make transactions without passing through a website.

The future of SEO is promising. The SEO service market size is expected to grow from $75.13 billion in 2023 to $88.91 billion in 2024 – a staggering CAGR of 18.3% (according to The Business Research Company) – as it adapts to incorporate reliable AI and semantic technologies.

These innovations support the creation of more dynamic and responsive web environments that adeptly cater to user needs and behaviors.

However, the journey hasn’t been without challenges, especially in large enterprise settings. Implementing AI solutions that are both explainable and strategically aligned with organizational goals has been a complex task.

Building effective AI involves aggregating relevant data and transforming it into actionable knowledge.

This differentiates an organization from competitors using similar language models or development patterns, such as conversational agents or retrieval-augmented generation copilots and enhances its unique value proposition.

Imagine an ontology as a giant instruction manual for describing specific concepts. In the world of SEO, we deal with a lot of jargon, right? Topicality, backlinks, E-E-A-T, structured data – it can get confusing!

An ontology for SEO is a giant agreement on what all those terms mean. It’s like a shared dictionary, but even better. This dictionary doesn’t just define each word. It also shows how they all connect and work together. So, “queries” might be linked to “search intent” and “web pages,” explaining how they all play a role in a successful SEO strategy.

Imagine it as untangling a big knot of SEO practices and terms and turning them into a clear, organized map – that’s the power of ontology!

While Schema.org is a fantastic example of a linked vocabulary, it focuses on defining specific attributes of a web page, like content type or author. It excels at helping search engines understand our content. But what about how we craft links between web pages?

What about the query a web page is most often searched for? These are crucial elements in our day-to-day work, and an ontology can be a shared framework for them as well. Think of it as a playground where everyone is welcome to contribute on GitHub similar to how the Schema.org vocabulary evolves.

The idea of an ontology for SEO is to augment Schema.org with an extension similar to what GS1 did by creating its vocabulary. So, is it a database? A collaboration framework or what? It is all of these things together. SEO ontology operates like a collaborative knowledge base.

It acts as a central hub where everyone can contribute their expertise to define key SEO concepts and how they interrelate. By establishing a shared understanding of these concepts, the SEO community plays a crucial role in shaping the future of human-centered AI experiences.

SEOntology snapshot
Screenshot from WebVowl, August 2024SEOntology – a snapshot (see an interactive visualization here).

The Data Interoperability Challenge In The SEO Industry

Let’s start small and review the benefits of a shared ontology with a practical example (here is a slide taken from Emilija Gjorgjevska’s presentation at this year’s ZagrebSEOSummit)

Data Interoperability ChallengeImage from Emilija Gjorgjevska’s, ZagrebSEOSummit, August 2024

Imagine your colleague Valentina uses a Chrome extension to export data from Google Search Console (GSC) into Google Sheets. The data includes columns like “ID,” “Query,” and “Impressions” (as shown on the left). But Valentina collaborates with Jan, who’s building a business layer using the same GSC data. Here’s the problem: Jan uses a different naming convention (“UID,” “Name,” “Impressionen,” and “Klicks”).

Now, scale this scenario up. Imagine working with n different data partners, tools, and team members, all using various languages. The effort to constantly translate and reconcile these different naming conventions becomes a major obstacle to effective data collaboration.

Significant value gets lost in just trying to make everything work together. This is where an SEO ontology comes in. It is a common language, providing a shared name for the same concept across different tools, partners, and languages.

By eliminating the need for constant translation and reconciliation, an SEO ontology streamlines data collaboration and unlocks the true value of your data.

The Genesis Of SEOntology

In the last year, we have witnessed the proliferation of AI Agents and the wide adoption of Retrieval Augmented Generation (RAG) in all its different forms (Modular, Graph RAG, and so on).

RAG represents an important leap forward in AI technology, addressing a key limitation of traditional large language models (LLMs) by letting them access external knowledge.

Traditionally, LLMs are like libraries with one book – limited by their training data. RAG unlocks a vast network of resources, allowing LLMs to provide more comprehensive and accurate responses.

RAGs improve factual accuracy, and context understanding, potentially reducing bias. While promising, RAG faces challenges in data security, accuracy, scalability, and integration, especially in the enterprise sector.

For successful implementation, RAG requires high-quality, structured data that can be easily accessed and scaled.

We’ve been among the first to experiment with AI Agents and RAG powered by the Knowledge Graph in the context of content creation and SEO automation.

Agent WordLiftScreenshot from Agent WordLift, August 2023

Knowledge Graphs (KGs) Are Indeed Gaining Momentum In RAG Development

Microsoft’s GraphRAG and solutions like LlamaIndex demonstrate this. Baseline RAG struggles to connect information across disparate sources, hindering tasks requiring a holistic understanding of large datasets.

KG-powered RAG approaches like the one offered by LlamaIndex in conjunction with WordLift address this by creating a knowledge graph from website data and using it alongside the LLM to improve response accuracy, particularly for complex questions.

LlamaIndex in conjunction with WordLiftImage from author, August 2024

We have tested workflows with clients in different verticals for over a year.

From keyword research for large editorial teams to the generation of question and answers for ecommerce websites, from content bucketing to drafting the outline of a newsletter or revamping existing articles, we’ve been testing different strategies and learned a few things along the way:

1. RAG Is Overhyped

It is simply one of many development patterns that achieve a goal of higher complexity. A RAG (or Graph RAG) is meant to help you save time finding an answer. It’s brilliant but doesn’t solve any marketing tasks a team must handle daily. You need to focus on the data and the data model.

While there are good RAGs and bad RAGs, the key differentiation is often represented by the “R” part of the equation: the Retrieval. Primarily, the retrieval differentiates a fancy demo from a real-world application, and behind a good RAG, there is always good data. Data, though, is not just any type of data (or graph data).

It is built around a coherent data model that makes sense for your use case. If you build a search engine for wines, you need to get the best dataset and model the data around the features a user will rely on when looking for information.

So, data is important, but the data model is even more important. If you are building an AI Agent that has to do things in your marketing ecosystem, you must model the data accordingly. You want to represent the essence of web pages and content assets.

Only some data vs Good dataImage from author, August 2024

2. Not Everyone Is Great At Prompting

Expressing a task in written form is hard. Prompt engineering is going at full speed towards automation (here is my article on going from prompting to prompt programming for SEO) as only a few experts can write the prompt that brings us to the expected outcome.

This poses several challenges for the design of the user experience of autonomous agents. Jakon Nielsen has been very vocal about the negative impact of prompting on the usability of AI applications:

“One major usability downside is that users must be highly articulate to write the required prose text for the prompts.”

Even in rich Western countries, statistics provided by Nielsen tell us that only 10% of the population can fully utilize AI! 

Simple Prompt Using Chain-of-Thought (CoT) More Sophisticated Prompt Combining Graph-of-Thought (GoT) and Chain-of-Knowledge (CoK)
“Explain step-by-step how to calculate the area of a circle with a radius of 5 units.” “Using the Graph-of-Thought (GoT) and Chain-of-Knowledge (CoK) techniques, provide a comprehensive explanation of how to calculate the area of a circle with a radius of 5 units. Your response should: Start with a GoT diagram that visually represents the key concepts and their relationships, including: Circle Radius Area Pi (π) Formula for circle area Follow the GoT diagram with a CoK breakdown that: a) Defines each concept in the diagram b) Explains the relationships between these concepts c) Provides the historical context for the development of the circle area formula Present a step-by-step calculation process, including: a) Stating the formula for the area of a circle b) Explaining the role of each component in the formula c) Showing the substitution of values d) Performing the calculation e) Rounding the result to an appropriate number of decimal places Conclude with practical applications of this calculation in real-world scenarios. Throughout your explanation, ensure that each step logically follows the previous one, creating a clear chain of reasoning from basic concepts to the final result.” This improved prompt incorporates GoT by requesting a visual representation of the concepts and their relationships. It also employs CoK by asking for definitions, historical context, and connections between ideas. The step-by-step breakdown and real-world applications further enhance the depth and practicality of the explanation.”

3. You Shall Build Workflows To Guide The User

The lesson learned is that we must build detailed standard operating procedures (SOP) and written protocols that outline the steps and processes to ensure consistency, quality, and efficiency in executing particular optimization tasks.

We can see empirical evidence of the rise of prompt libraries like the one offered to users of Anthropic models or the incredible success of projects like AIPRM.

In reality, we learned that what creates business value is a series of ci steps that help the user translate the context he/she is navigating in into a consistent task definition.

We can start to envision marketing tasks like conducting keyword research as a Standard Operating Procedure that can guide the user across multiple steps (here is how we intend the SOP for keyword discovery using Agent WordLift)

4. The Great Shift To Just-in-Time UX 

In traditional UX design, information is pre-determined and can be organized in hierarchies, taxonomies, and pre-defined UI patterns. As AI becomes the interface to the complex world of information, we’re witnessing a paradigm shift.

UI topologies tend to disappear, and the interaction between humans and AI remains predominantly dialogic. Just-in-time assisted workflows can help the user contextualize and improve a workflow.

  • You need to think in terms of business value creation, focus on the user’s interactive journey, and facilitate the interaction by creating a UX on the fly. Taxonomies remain a strategic asset, but they operate behind the scenes as the user is teleported from one task to another, as recently brilliantly described by Yannis Paniaras from Microsoft.
The Shift to Just-In-Time UX: How AI is Reshaping User Experiences”Image from “The Shift to Just-In-Time UX: How AI is Reshaping User Experiences” by Yannis Paniaras, August 2024

5. From Agents To RAG (And GraphRAG) To Reporting

Because the user needs a business impact and RAG is only part of the solution, the focus quickly shifts from more generic questions and answering user patterns to advanced multi-step workflows.

The biggest issue, though, is what outcome the user needs. If we increase the complexity to capture the highest business goals, it is not enough to, let’s say, “query your data” or “chat with your website.”

A client wants a report, for example, of what is the thematic consistency of content within the entire website (this is a concept that we recently discovered as SiteRadus in Google’s massive data leak), the overview of the seasonal trends across hundreds of paid campaigns, or the ultimate review of the optimization opportunities related to the optimization of Google Merchant Feed.

You must understand how the business operates and what deliverables you will pay for. What concrete actions could boost the business? What questions need to be answered?

This is the start of creating a tremendous AI-assisted reporting tool.

How Can A Knowledge Graph (KG) Be Coupled With An Ontology For AI Alignment, Long-term Memory, And Content Validation?

The three guiding principles behind SEOntology:

  • Making SEO data interoperable to facilitate the creation of knowledge graphs while reducing unneeded crawls and vendor locked-in;
  • Infusing SEO know-how into AI agents using a domain-specific language.
  • Collaboratively sharing knowledge and tactics to improve findability and prevent misuse of Generative AI.

When you deal with at least two data sources in your SEO automation task, you will already see the advantage of using SEOntology.

SEOntology As “The USB-C Of SEO/Crawling Data”

Standardizing data about content assets, products, user search behavior, and SEO insights is strategic. The goal is to have a “shared representation” of the Web as a communication channel.

Let’s take a step backward. How does a Search Engine represent a web page? This is our starting point here. Can we standardize how a crawler would represent data extracted from a website? What are the advantages of adopting standards?

Practical Use Cases

Integration With Botify And Dynamic Internal Linking

Over the past few months, we’ve been working closely with the Botify team to create something exciting: a Knowledge Graph powered by Botify’s crawl data and enhanced by SEOntology. This collaboration is opening up new possibilities for SEO automation and optimization.

Leveraging Existing Data With SEOntology

Here’s the cool part: If you’re already using Botify, we can tap into that goldmine of data you’ve collected. No need for additional crawls or extra work on your part. We use the Botify Query Language (BQL) to extract and transform the needed data using SEOntology.

Think of SEOntology as a universal translator for SEO data. It takes the complex information from Botify and turns it into a format that’s not just machine-readable but machine-understandable. This allows us to create a rich, interconnected Knowledge Graph filled with valuable SEO insights.

What This Means for You

Once we have this Knowledge Graph, we can do some pretty amazing things:

  • Automated Structured Data: We can automatically generate structured data markup for your product listing pages (PLPs). This helps search engines better understand your content, potentially improving your visibility in search results.
  • Dynamic Internal Linking: This is where things get really interesting. We use the data in the Knowledge Graph to create smart, dynamic internal links across your site. Let me break down how this works and why it’s so powerful.

In the diagram below, we can also see how data from Botify can be blended with data from Google Search Console.

While in most implementations, Botify already imports this data into its crawl projects, when this is not the case, we can trigger a new API request and import clicks, impressions, and positions from GSC into the graph.

Collaboration With Advertools For Data Interoperability

Similarly, we collaborated with the brilliant Elias Dabbas, creator of Advertools — a favorite Python library among marketers – to automate a wide range of marketing tasks.

Our joint efforts aim to enhance data interoperability, allowing for seamless integration and data exchange across different platforms and tools.

In the first Notebook, available in the SEOntology GitHub repository, Elias showcases how we can effortlessly construct attributes for the WebPage class, including title, meta description, images, and links. This foundation enables us to easily model complex elements, such as internal linking strategies. See here the structure:

  • Internal_Links
    • anchorTextContent
    • NoFollow
    • Link

We can also add a flag if the page is already using schema markup:

  • usesSchema

Formalizing What We Learned From The Analysis Of The Leaked Google Search Documents

While we want to be extremely conscious in deriving tactics or small schemes from Google’s massive leak, and we are well aware that Google will quickly prevent any potential misuse of such information, there is a great level of information that, based on what we learned, can be used to improve how we represent web content and organize marketing data.

Despite these constraints, the leak offers valuable insights into improving web content representation and marketing data organization. To democratize access to these insights, I’ve developed a Google Leak Reporting tool designed to make this information readily available to SEO pros and digital marketers.

For instance, understanding Google’s classification system and its segmentation of websites into various taxonomies has been particularly enlightening. These taxonomies – such as ‘verticals4’, ‘geo’, and ‘products_services’ – play a crucial role in search ranking and relevance, each with unique attributes that influence how websites and content are perceived and ranked in search results.

By leveraging SEOntology, we can adopt some of these attributes to enhance website representation.

Now, pause for a second and imagine transforming the complex SEO data you manage daily through tools like Moz, Ahrefs, Screaming Frog, Semrush, and many others into an interactive graph. Now, envision an Autonomous AI Agent, such as Agent WordLift, at your side.

This agent employs neuro-symbolic AI, a cutting-edge approach that combines neural learning capabilities with symbolic reasoning, to automate SEO tasks like creating and updating internal links. This streamlines your workflow and introduces a level of precision and efficiency previously unattainable.

SEOntology serves as the backbone for this vision, providing a structured framework that enables the seamless exchange and reuse of SEO data across different platforms and tools. By standardizing how SEO data is represented and interconnected, SEOntology ensures that valuable insights derived from one tool can be easily applied and leveraged by others. For instance, data on keyword performance from SEMrush could inform content optimization strategies in WordLift, all within a unified, interoperable environment. This not only maximizes the utility of existing data but also accelerates the automation and optimization processes that are crucial for effective marketing.

Infusing SEO Know-How Into AI Agents

As we develop a new agentic approach to SEO and digital marketing, SEOntology serves as our domain-specific language (DSL) for encoding SEO skills into AI agents. Let’s look at a practical example of how this works.

GraphQL Query Generator and ValidatorScreenshot from WordLift, August 2024

We’ve developed a system that makes AI agents aware of a website’s organic search performance, enabling a new kind of interaction between SEO professionals and AI. Here’s how the prototype works:

System Components

  • Knowledge Graph: Stores Google Search Console (GSC) data, encoded with SEOntology.
  • LLM: Translates natural language queries into GraphQL and analyzes data.
  • AI Agent: Provides insights based on the analyzed data.

Human-Agent Interaction

Human, LLM, Knowledge Graph, AI Agent interactionImage from author, August 2024

The diagram illustrates the flow of a typical interaction. Here’s what makes this approach powerful:

  • Natural Language Interface: SEO professionals can ask questions in plain language without constructing complex queries.
  • Contextual Understanding: The LLM understands SEO concepts, allowing for more nuanced queries and responses.
  • Insightful Analysis: The AI agent doesn’t just retrieve data; it provides actionable insights, such as:
    • Identifying top-performing keywords.
    • Highlighting significant performance changes.
    • Suggesting optimization opportunities.
  • Interactive Exploration: Users can ask follow-up questions, enabling a dynamic exploration of SEO performance.

By encoding SEO knowledge through SEOntology and integrating performance data, we’re creating AI agents that can provide context-aware, nuanced assistance in SEO tasks. This approach bridges the gap between raw data and actionable insights, making advanced SEO analysis more accessible to professionals at all levels.

This example illustrates how an ontology like SEOntology can empower us to build agentic SEO tools that automate complex tasks while maintaining human oversight and ensuring quality outcomes. It’s a glimpse into the future of SEO, where AI augments human expertise rather than replacing it.

Human-In-The-Loop (HTIL) And Collaborative Knowledge Sharing

Let’s be crystal clear: While AI is revolutionizing SEO and Search, humans are the beating heart of our industry. As we dive deeper into the world of SEOntology and AI-assisted workflows, it’s crucial to understand that Human-in-the-Loop (HITL) isn’t just a fancy add-on—it’s the foundation of everything we’re building.

The essence of creating SEOntology is to transfer our collective SEO expertise to machines while ensuring we, as humans, remain firmly in the driver’s seat. It’s not about handing over the keys to AI; it’s about teaching it to be the ultimate co-pilot in our SEO journey.

Human-Led AI: The Irreplaceable Human Element

SEOntology is more than a technical framework – it’s a catalyst for collaborative knowledge sharing that emphasizes human potential in SEO. Our commitment extends beyond code and algorithms to nurturing skills and expanding the capabilities of new-gen marketers and SEO pros.

Why? Because AI’s true power in SEO is unlocked by human insight, diverse perspectives, and real-world experience. After years of working with AI workflows, I’ve realized that agentive SEO is fundamentally human-centric. We’re not replacing expertise; we’re amplifying it.

We deliver more efficient and trustworthy results by blending cutting-edge tech with human creativity, intuition, and ethical judgment. This approach builds trust with clients within our industry and across the web.

Here’s where humans remain irreplaceable:

  • Understanding Business Needs: AI can crunch numbers but can’t replace the nuanced understanding of business objectives that seasoned SEO professionals bring. We need experts who can translate client goals into actionable SEO strategies.
  • Identifying Client Constraints: Every business is unique, with its limitations and opportunities. It takes human insight to navigate these constraints and develop tailored SEO approaches that work within real-world parameters.
  • Developing Cutting-Edge Algorithms: The algorithms powering our AI tools don’t materialize out of thin air. We need brilliant minds to develop state-of-the-art algorithms, learn from human input, and continually improve.
  • Engineering Robust Systems: Behind every smooth-running AI tool is a team of software engineers who ensure our systems are fast, secure, and reliable. This human expertise keeps our AI assistants running like well-oiled machines.
  • Passion for a Better Web: At the heart of SEO is a commitment to making the web a better place. We need people who share Tim Berners’s—Lee’s vision—people who are passionate about developing the web of data and improving the digital ecosystem for everyone.
  • Community Alignment and Resilience: We need to unite to analyze the behavior of search giants and develop resilient strategies. It’s about solving our problems innovatively as individuals and as a collective force. This is what I always loved about the SEO industry!

Extending The Reach Of SEOntology

As we continue to develop SEOntology, we’re not operating in isolation. Instead, we’re building upon and extending existing standards, particularly Schema.org, and following the successful model of the GS1 Web Vocabulary.

SEOntology As An Extension Of Schema.org

Schema.org has become the de facto standard for structured data on the web, providing a shared vocabulary that webmasters can use to markup their pages.

However, while Schema.org covers a broad range of concepts, it doesn’t delve deeply into SEO-specific elements. This is where SEOntology comes in.

An extension of Schema.org, like SEOntology, is essentially a complementary vocabulary that adds new types, properties, and relationships to the core Schema.org vocabulary.

This allows us to maintain compatibility with existing Schema.org implementations while introducing SEO-specific concepts not covered in the core vocabulary.

Learning From GS1 Web Vocabulary

The GS1 Web Vocabulary offers a great model for creating a successful extension that interacts seamlessly with Schema.org. GS1, a global organization that develops and maintains supply chain standards, created its Web Vocabulary to extend Schema.org for e-commerce and product information use cases.

The GS1 Web Vocabulary demonstrates, even recently, how industry-specific extensions can influence and interact with schema markup:

  • Real-world impact: The https://schema.org/Certification property, now officially embraced by Google, originated from GS1’s https://www.gs1.org/voc/CertificationDetails. This showcases how extensions can drive the evolution of Schema.org and search engine capabilities.

We want to follow a similar approach to extend Schema.org and become the standard vocabulary for SEO-related applications, potentially influencing future search engine capabilities, AI-driven workflows, and SEO practices.

Much like GS1 defined their namespace (gs1:) while referencing schema terms, we have defined our namespace (seovoc:) and are integrating the classes within the Schema.org hierarchy when possible.

The Future Of SEOntology

SEOntology is more than just a theoretical framework; it’s a practical tool designed to empower SEO professionals and tool makers in an increasingly AI-driven ecosystem.

Here’s how you can engage with and benefit from SEOntology.

If you’re developing SEO tools:

  • Data Interoperability: Implement SEOntology to export and import data in a standardized format. This ensures your tools can easily interact with other SEOntology-compliant systems.
  • AI-Ready Data: By structuring your data according to SEOntology, you’re making it more accessible for AI-driven automations and analyses.

If you’re an SEO professional:

  • Contribute to Development: Just like with Schema.org, you can contribute to SEOntology’s evolution. Visit its GitHub repository to:
    • Raise issues for new concepts or properties you think should be included.
    • Propose changes to existing definitions.
    • Participate in discussions about the future direction of SEOntology.
  • Implement in Your Work: Start using SEOntology concepts in your structured data.

In Open Source We Trust

SEOntology is an open-source effort, following in the footsteps of successful projects like Schema.org and other shared linked vocabularies.

All discussions and decisions will be public, ensuring the community has a say in SEOntology’s direction. As we gain traction, we’ll establish a committee to steer its development and share regular updates.

Conclusion And Future Work

The future of marketing is human-led, not AI-replaced. SEOntology isn’t just another buzzword – it’s a step towards this future. SEO is strategic for the development of agentive marketing practices.

SEO is no longer about rankings; it’s about creating intelligent, adaptive content and fruitful dialogues with our stakeholders across various channels. Standardizing SEO data and practices is strategic to build a sustainable future and to invest in responsible AI.

Are you ready to join this revolution?

There are three guiding principles behind the work of SEOntology that we need to make clear to the reader:

  • As AI needs semantic data, we need to make SEO data interoperable, facilitating the creation of knowledge graphs for everyone. SEOntology is the USB-C of SEO/crawling data. Standardizing data about content assets and products and how people find content, products, and information in general is important. This is the first objective. Here, we have two practical use cases. We have a connector for WordLift that gets crawl data from the Botify crawler and helps you jump-start a KG that uses SEOntology as a data model. We are also working with Advertools, an open-source crawler and SEO tool, to make data interoperable with SEOntology;
  • As we progress with the development of a new agentic way of doing SEO and digital marketing, we want to infuse the know-how of SEO using SEOntology, a domain-specific language to infuse the SEO mindset to SEO agents (or multi-agent systems like Agent WordLift). In this context, the skill required to create dynamic internal links is encoded as nodes in a knowledge graph, and opportunities become triggers to activate workflows.
  • We expect to work with human-in-the-loop HITL, meaning that the ontology will become a way to collaboratively share knowledge and tactics that help improve findability and prevent the misuse of Generative AI that is polluting the Web today.

Project Overview

This work on SEOntology is the product of collaboration. I extend my sincere thanks to the WordLift team, especially CTO David Riccitelli. I also appreciate our clients for their dedication to innovation in SEO through knowledge graphs. Special thanks to Milos Jovanovik and Emilia Gjorgjevska for their critical expertise. Lastly, I’m grateful to the SEO community and the SEJ editorial team for their support in sharing this work.

More resources: 


Featured Image: tech_BG/Shutterstock

Google Rolls Out AI-Organized Search Results Pages via @sejournal, @MattGSouthern

Google is introducing AI-organized search results pages in the United States.

The new feature, set to launch this week, returns a full page of multi-format results personalized for the searcher.

Google’s announcement states:

“This week, we’re rolling out search results pages organized with AI in the U.S. — beginning with recipes and meal inspiration on mobile. You’ll now see a full-page experience, with relevant results organized just for you. You can easily explore content and perspectives from across the web including articles, videos, forums and more — all in one place.”

Key Features

The AI-organized pages will compile various content types, including articles, videos, and forum discussions.

Google claims this approach will provide users with a more diverse range of information sources and perspectives.

In its announcement, Google adds:

“… with AI-organized search results pages, we’re bringing people more diverse content formats and sites, creating even more opportunities for content to be discovered.”

Industry Implications

While Google touts the benefits of AI-organized search results pages, the update raises several questions:

  1. How will the AI-organized pages affect traffic to individual websites? Keeping users on Google’s results page might reduce clicks to source websites.
  2. With AI determining content organization, there are concerns about potential biases in how information is presented.
  3. The new format may require new strategies to ensure visibility within these AI-organized results.
  4. It’s unclear how this change will impact ad visibility.

This update could alter how we approach SEO. We may need to adapt strategies to ensure content is discoverable and presentable in this new format.

Microsoft’s Bing recently announced an expansion of its generative search capabilities, focusing on handling complex, informational queries. Google’s reorganizing of entire results pages appears to be a unique offering compared to Bing’s.

The initial rollout focusing on mobile devices for recipe and meal-related queries aligns with Google’s mobile-first indexing approach.

It remains to be seen how this feature will translate to desktop searches.

Google’s Response to Industry Concerns

In light of the questions raised by this update, we contacted Google for clarification on several key points.

Impact on Search Console Tracking

Regarding how AI-organized search results will be tracked in Google Search Console, a Google spokesperson stated:

“We do not separate traffic by every feature in Search Console, but publishers will continue to see their traffic from Search reflected there. Check out the supported search appearances in our documentation.”

This suggests that while specific metrics will not be available for AI-organized pages, site owners will still be able to access overall traffic data.

Timeline for Expansion

When asked about the timeline for expanding this feature to other categories and regions, Google responded:

“When we previewed this feature, we mentioned expanding this to additional categories including dining, movies, music, books, hotels, and shopping. No further details to share at this time.”

While this confirms expansion plans, Google has not provided specific timelines for these rollouts.

Guidance for SEO Professionals and Content Creators

On whether new tools or guidance will be provided for optimizing content for AI-organized search results, Google emphasized that no changes are necessary:

“SEO professionals and creators don’t need to do anything differently. Search results pages organized with AI are rooted in our core Search ranking and quality systems, which we have been honing for decades to surface high quality information.”

This response suggests that existing SEO best practices should continue to be effective for visibility in these new result formats.

Looking Ahead

Google’s responses provide some clarity but also leave room for speculation.

The lack of specific tracking for AI-organized pages in Search Console may present challenges for SEO professionals in understanding the direct impact of this new feature on their traffic.

The confirmation of plans to expand to other categories like dining, movies, music, books, hotels, and shopping indicates that this update could have far-reaching effects across various industries.

Despite Google’s assurances, new best practices may emerge as the SEO community adapts to this significant change in search result presentation.

We here at SEJ will closely monitor the rollout and report on its effects and what it means for you in the coming months. Sign up for the SEJ newsletter to stay up to date.


Featured Image: JarTee/Shutterstock

An Introduction To SEO Strategy For A Digital Presence

This edited extract is from Digital and Social Media Marketing: A Results-Driven Approach  edited by Aleksej Heinze, Gordon Fletcher, Ana Cruz, Alex Fenton ©2024 and is reproduced with permission from Routledge. The extract below was taken from the chapter Using Search Engine Optimisation to Build Trust co-authored with Aleksej Heinze, Senior Professor at KEDGE Business School, France.

The key challenge for SEO is that good rankings in SERPs are almost entirely based on each search engine’s private algorithm for identifying high-quality content and results, which is a long-term activity.

The initial formula of PageRank (Page et al. 1999) used by Google, which used links pointing to a page to rank its importance, has evolved significantly and is no longer publicly available.

All search engines regularly update their algorithms to identify high-quality, relevant content to a particular search query. Google implements around 500 – 600 changes to its algorithm each year (Gillespie 2019).

These are product updates, similar to Windows updates. Most of these changes are minor with little impact, but a few critical core updates each year will require careful review on the majority of websites since they can result in major SERP changes.

Search engines are using artificial intelligence to improve their technology to enable them to identify high-quality, relevant content and are constantly testing new ways to present users with relevant content.

The arrival of ChatGPT by Open AI in 2022 presents a rival type of offering that has shaken the foundations of the traditional search engine business model (Poola 2023).

In such a dynamic environment, it is important to keep up to date with algorithm changes.

This can be done by following the Google Search Status dashboard (Google) and SEO-related blog posts and monitoring, including the MOZ algorithm change calendar (Moz).

How Search Engines Work 

In essence, a search engine’s crawler, spider, robot or ‘bot’ discovers web page links, and then internally determines if there is value in analysing the links.

Then, the bot automatically retrieves the content behind each link (including more links). This process is called crawling.

Bots may then add the discovered pages to the search engines’s index to be retrieved when a user searches for something.

The ranking order in which the links appear in SERPs is calculated by the engine’s algorithm, which examines the relevance of the content to the query.

This relevance is determined by a combination of over 200 factors such as the visible text, keywords, the position and relationship of words, links, synonyms and semantic entities (Garg 2022).

When the user of a search engine types in a query, they are presented with a list of links to content that the engine calculates will satisfy the intent of the query – the list of results is the SERP.

Typically, the list of results that are shown in SERPs includes a mix of paid-for and organic results. Each link includes a short URL, title and description, as well as other options such as thumbnail images, videos and other related internal site links.

Search engines are constantly making changes to SERPs to improve the experience for those searching. For example, Bing includes Bing Chat, allowing responses to be offered by their AI bot.

Google introduced a knowledge graph or a summary answer box, found underneath the search box on the right of the organic search results.

The Bing Chat as well as Google knowledge graph provide a direct and relevant summary response to a query without the need for a further click to the source page (and retaining the user at the search engine).

This offering leads to so-called 0-click searches, which cannot be tracked in the data relating to a digital presence and are only seen in data that relates content visibility to SERPs.

Some Google SERP snippets can also appear as a knowledge graph (Figure 12.8) or a search snippet (Figure 12.9).

Figure 12.8: Google SERP for “KEDGE Business School” including a knowledge graph on the right-hand side of the page (Google and the Google logo are trademarks of Google LLC).Figure 12.8: Google SERP for “KEDGE Business School” including a knowledge graph on the right-hand side of the page.
Figure 12.9: Search snippet for Jean Reno (Google and the Google logo are trademarks of Google LLC).Figure 12.9: Search snippet for Jean Reno.

The volatility of the SERPs can be evidenced by the varying results produced by the same search in different locations.

The listing for the US market (Figure 12.10) and carousel for the European market (Figure 12.11) for “best DJs” shows that geolocation increasingly comes into play in the page ranking of SERPs.

Personalisation is also relevant. For example, when a user is logged into a Google product, their browser history influences the organic SERPs. SERPs change depending on what terms are used.

This means a pluralised term produces different SERPs to searches that use the singular term.

Tools, such as those offered by Semrush, include functionality to quickly identify this form of volatility and understand sectors that are being affected by changes.

Figure 12.10: US results for “best DJs” (Google and the Google logo are trademarks of Google LLC).Figure 12.10: US results for “best DJs”
Figure 12.11: European results for “best DJs” (Google and the Google logo are trademarks of Google LLC)Figure 12.11: European results for “best DJs”

Recent innovations by Google include the search generative experience (SGE) currently being tested in the US market. This is a different search experience that is more visual and uses artificial intelligence.

The 2015 introduction of RankBrain and other algorithms means that Google now better understands human language and context.

Industry publications, including Search Engine Roundtable and Search Engine Land, keep pace with this dynamic landscape.

Implementing Search Engine Optimisation 

Identification of the most relevant search terms is the starting point for developing a website map and themes for content.

The search terms will also define the focus for individual pages and blog posts. This approach has a focus on the technical/on-page, content, and off-page aspects of the website.

Any SEO activity begins with prior knowledge of the organisation, including its objectives and targets as well as the persona that has been defined.

The initial phase of optimising a website for Google search involves:

  1. A technical and content audit.
  2. Keyword identification and analysis.
  3. Implementing any changes in the content management system (CMS) and content.
  4. Using the secure HTTPS protocol for the website.
  5. Submitting the website to Google Search Console.
  6. Submitting the website to Bing Webmaster Tools.
  7. Submitting the website to other appropriate search engines.
  8. Adding website tracking code such as Google Analytics, Hotjar or others to the website.

Summary

SEO plays a critical role in enhancing an organisation’s digital presence, and the dynamic nature of search engine algorithms provides a way to address the immediate pain touchpoints of a persona.

This focused around the imperative for organisations to offer content that not only resonates with a persona’s needs but also aligns with the evolving criteria of search engines like Google, Baidu or Bing.

This latter alignment is crucial given the stakeholder tendency to focus only on the first SERP. It is important to adhere to ethical SEO practices employing ‘White Hat SEO’ tactics that comply with search engine guidelines, as opposed to more manipulative techniques.

There is a need for continuous monitoring and reviewing of any SEO activities.

Frequently changing search engine algorithms, which now heavily incorporate AI and machine learning, means that a campaign’s parameters can change quickly. SEO is not a “set and forget” activity.

Staying informed and adapting to these changes is essential for maintaining and improving search engine rankings.

The environmental impact of digital activities should also be a consideration in SEO and wider marketing practices, optimising websites not only aligns with SEO best practices but also contributes to sustainability.

Search engines offer marketers one of the largest big data sets available to refine and target their content creation activities.

Historic search behaviours are good predictors of the future, and the use of these resources helps marketers to optimise and be better placed to offer value to their persona.


To read the book, SEJ readers have an exclusive 20% discount until the end of 2024 using the code DSMM24 at Routledge.

The book officially launches on October 7 2024 and you can attend the event with a chance to hear from some of the authors by registering through this link.

More resources:


Featured Image: Sutthiphong Chandaeng/Shutterstock

Google Rolls Out CrUX Vis Core Web Vitals Tool via @sejournal, @martinibuster

Google rolled out a new Core Web Vitals tool called CrUX Vis that shows you hidden patterns in performance scores and offers guidance on what to improve. The data is sourced from the CrUX dataset which is based on actual user experiences on the URLs and websites that are analyzed and explored in the new tool.

CrUX

The new tool is based on the CrUX dataset which is what the Core Web Vitals scores are based on.

Chrome’s documentation of CrUX explains:

“The Chrome User Experience Report (also known as the Chrome UX Report, or CrUX for short) is a dataset that reflects how real-world Chrome users experience popular destinations on the web.

CrUX is the official dataset of the Web Vitals program. All user-centric Core Web Vitals metrics are represented.

CrUX data is collected from real browsers around the world, based on certain browser options which determine user eligibility. A set of dimensions and metrics are collected which allow site owners to determine how users experience their sites.

The data collected by CrUX is available publicly through a number of Google tools and third-party tools and is used by Google Search to inform the page experience ranking factor.

Not all origins or pages are represented in the dataset. There are separate eligibility criteria for origins and pages, primarily that they must be publicly discoverable and there must be a large enough number of visitors in order to create a statistically significant dataset.”

Debugging Core Web Vitals

Improving website performance scores may not offer the direct ranking benefit that many SEOs and publishers hoped it would but it’s still the same critical factor to get right it’s always been. High performance scores improve earnings, ad clicks, conversions, user experience, website popularity and virtually every goal an SEO and publisher has for a site, including indirect benefits to rankings. A site can still limp along with poor performance scores but it will not be living up to its full earnings potential.

Although tools based on Chrome’s Lighthouse offer performance snapshots and estimated scores those tools were unable to provide a sense of how the site was performing over time or provide a breakout of important performance metrics to gauge whether performance is trending up or down.

CrUX Vis

Chrome’s new tool is called CrUX Vis, a data visualization tool that enables users to visualize the Chrome User Experience data (CrUX). CrUX Vis provides an entirely new way to understand website performance and gain a big picture view of what’s going on at the URL and website level (called origin).

The different variables for what is visualized can be changed in the section at the top of the page called Controls, covering data, device and period.

Screenshot Of CrUX Vis Controls

Segment Data By Multiple Variables

As seen in the screenshot above, the data can be segmented in three ways:

  1. Data
    Performance scores can be viewed by origin (the entire site) or by URL
  2. Device
    Data can be segmented and visualized by mobile, data and a combined view.
  3. Period (Date Range)
    The tool currently allows data visualization by 25 overlapping time periods stretching back about six months. It currently shows performance visualizations from 3/17/2024 through 09/28/2024.

Five Views Of Metrics

There are five ways to analyze the data, covering core web vitals, three categories of metrics and all metrics combined. These variables are accessible on left hand navigation panel on the desktop UI (user interface).

  1. Core Web Vitals
  2. Loading Performance
  3. Interactivity
  4. Visual Stability
  5. All Metrics Combined

Visualizing Data

The visualization for Core Web Vitals shows a time-based trend graph that’s colored with green, yellow, and pink. Green is good and pink is not good.

The three core web vitals are represented by a circle, squate and a triangle:

  • Circle = Largest Contentful Paint (LCP):
  • Square = Interaction to Next Paint (INP)
  • Triangle = Cumulative Layout Shift (CLS)

The desktop UI (user interface) shows the trend graph and a summary on the left and a text explanation on the right.

Screenshot Of User Interface

The graph offers a visual snapshot of which direction the core web vitals are moving and an explanation of the kind of trend for each metric.

The three kinds of trends are:

  1. Good And Improving
  2. Good And Stable
  3. Poor And Regressing

Screenshot Showing CWV Performance

A more comprehensive explanation of the data is to the right of the trend graph, with each metric identified by the circle, square, and triangle icons.

Screenshot Of Data Explanation

Loading Performance

Using the left hand navigation to get to the Loading Performance screen shows another trend graph that offers additional metrics related to how fast the site or URL loads.

It offers the following six visualizations:

  • Largest Contentful Paint (LCP)
  • First Contentful Paint (FCP)
  • Time to First Byte (TTFB)
  • Round Trip Time (RTT)
  • Navigation Types
  • Form Factors

Screenshot Of Six Visualization Choices

There’s a toggle next to each choice:

Clicking the toggle shows the trend graph:

The rest of the choices show similar breakdowns of each kind of metric.

The new CrUX Vis tool should be useful to publishers and digital marketers who want to get an accurate measurement of website performance, visualized as a trend. It’s useful for competitior research and for website audits.

Go check it out at:

CrUX Vis

Featured Image by Shutterstock/Krakenimages.com

Google’s Search Liaison Addresses Brand Bias Concerns via @sejournal, @MattGSouthern

In a recent interview with Aleyda Solis, Google’s Search Liaison, Danny Sullivan, discussed the company’s approach to ranking smaller websites versus larger brands.

This topic has long been a point of contention, with concerns that Google’s ranking systems favor brands over independent sites.

Fairness In Search Results

Sullivan claims that Google doesn’t inherently favor brands, stating:

“Our ranking systems aren’t saying ‘are you a big brand therefore you rank’… The core of it isn’t really whether you’re big or you’re small, the core of it is whether you have the most useful, the most relevant, most satisfying information.”

The Perception Problem

Despite Google’s stance, Sullivan acknowledged the widespread perception that larger, well-established sites have an advantage in search results.

He recognized the frustration of smaller site owners who feel they cannot compete with bigger brands for visibility.

Sullivan states:

“I have looked at cases where people say you don’t like small sites, and I am not taking away from any of the real concerns because they are there… I wish they were doing better, but I can also see co-occurring in some of the same queries that I’m given other independent sites that are doing well.”

Challenges & Improvements

Sullivan admitted that Google’s systems sometimes fail to recognize high-quality content from smaller sites.

He assured that the company is actively improving this aspect of its algorithms.

Sullivan said:

“We don’t want it to be only the very big things rank well and I think in the last update we did talk about how we were taking in a lot of these concerns and trying to understand how we can do more for some of the smaller sites, the so-called independent sites.”

Advice For Smaller Sites

For independent website owners feeling discouraged, Sullivan offered some advice: focus on developing your brand.

He advised:

“If you’re a smaller site that feels like you haven’t really developed your brand, develop it. That’s not because we’re going to rank you because of your brand, but because it’s probably the things that cause people externally to recognize you as a good brand may in turn co-occur or be alongside the kinds of things that our ranking systems are kind of looking to reward.”

On advice for content creators, Sullivan adds:

“Just keep listening to your heart and doing what it is that you think is the right thing to be doing… Our ranking systems are trying to reward great content that’s made for people and if you feel like you’re doing that, then we’re going to try to catch up to you.”

Looking Ahead

Google appears to be taking these concerns seriously.

Sullivan mentioned that recent updates have aimed to do more for smaller sites. However, he maintains that Google’s goal is to show the best content regardless of brand recognition.

While challenges remain, Google’s acknowledgment of the issue and efforts to improve suggests a potential shift with future updates.

Hear Sullivan’s full statements in the video below:


Featured Image: rudall30/Shutterstock

Google’s SEO Tip To Get New Site Picked Up Faster via @sejournal, @martinibuster

Google’s John Mueller offered a useful for technical SEO tip for those launching a new site that will help your site get picked up by Google faster by avoiding this one common mistake.

High Priority For Site Launch

Launching a website is a chance to take everything learned from previous experiences and apply them with the benefit of hindsight. There’s no better teacher for success than failure because lessons learned from mistakes are never forgotten.

Someone who recently registered a new domain started a discussion on Reddit asking what were the top three considerations for launching a successful website before anything else has been done. The person asking the question preemptively ruled out the obvious answer of adding the domain to Google Search Console and set the ground rule that the niche or type of business didn’t matter. What did matter is that the suggestions must be important for scaling traffic within the first six month of the website.

They asked:

“Let’s say you have a brand new domain and you’ve been given a task to build traffic in the next 6 months. The niche, business does not matter, and the basics like ‘adding domain to Google search console’ don’t matter.

Tell me what are the first 3, high-priority things you’ll implement.”

The Most Upvoted Answer

It’s somewhat surprising that the most upvoted answer, with 83 votes, was one that offered the most obvious suggestions.

The top upvoted answer was:

“Create landing pages/content for your lowest funnel keyword opportunities and work the the way up.”

It’s a matter of course that the information architecture of the site should be planned out ahead of time (things like keywords, topics, key pages, a complete org-chart style map of categories with room left for expanding topical coverage, and an interlinking strategy). The upvoted answer is absolutely correct but it’s also fairly obvious.

The rest of that highly upvoted response:

“Claim brand on top social medias.

Build easiest citations and directories that I know get indexed. Plus niche relevant ones.

Start reactive digital PR as main initial link building campaign.”

The obviousness of that upvoted answer is in contrast with the not so obvious quality of Mueller’s response.

John Mueller Advice On SEO Preparation

John Mueller’s advice is excellent and offers an insight into a technical issue that is easy to overlook.

He wrote:

“Just throwing this out there – if you don’t have a site ready, either keep DNS disabled or put up a custom holding page. Don’t use a generic server / CMS holding page. It generally takes longer for a site that’s known to be parked / duplicate to get recognized as a normal site than it does for a site to be initially picked up.”

Keep DNS Disabled

DNS stands for Domain Name System and is a reference to the backend process of converting a domain name to the IP address where the actual content exists. All content exists at an IP address, not at the domain name. The domain name just points to where the content is. By keeping DNS disabled what happens is that Google doesn’t discover the domain pointing to anything so it essentially doesn’t exist.

Don’t Use Generic Server/CMS Holding Page

A generic server holding page is the same as a parked domain, it’s like a false signal to Google that something exists at the IP address that a domain name resolves to.

The effect of Mueller’s advice regarding disabling a DNS and not using a generic holding page is to keep the domain name from resolving to a holding page (assuming that a registrar’s holding page is also turned off). This keeps Google from sniffing out the domain and finding a generic “nothing here” holding page.

Mueller’s advice points to the technical issue that Google will recognize and index a site faster if a generic version is never activated and the domain name essentially doesn’t exist.

So if you want your website to be picked up and indexed quickly then it’s best to not use a generic domain holding page.

Read Mueller’s advice here:

Brand New Domain : What are the first 3 things you’ll do?

Featured Image by Shutterstock/Luis Molinero

Bing Expands Generative Search Capabilities For Complex Queries via @sejournal, @MattGSouthern

Microsoft has announced an expansion of Bing’s generative search capabilities.

The update focuses on handling complex, informational queries.

Bing provides examples such as “how to effectively run a one-on-one” and “how can I remove background noise from my podcast recordings.”

Searchers in the United States can access the new features by typing “Bing generative search” into the search bar. This will present a carousel of sample queries.

Screenshot from: blogs.bing.com, October 2024.

A “Deep search” button on the results page activates the generative search function for other searches.

Screenshot from: blogs.bing.com, October 2024.

Beta Release and Potential Challenges

It’s important to note that this feature is in beta.

Bing acknowledges that you may experience longer loading times as the system works to ensure accuracy and relevance.

The announcement reads:

“While we’re excited to give you this opportunity to explore generative search firsthand, this experience is still being rolled out in beta. You may notice a bit of loading time as we work to ensure generative search results are shown when we’re confident in their accuracy and relevancy, and when it makes sense for the given query. You will generally see generative search results for informational and complex queries, and it will be indicated under the search box with the sentence “Results enhanced with Bing generative search” …”

This is the waiting screen you get after clicking on “Deep search.”

Screenshot from: blogs.bing.com, October 2024.

In practice, I found the wait was long and sometimes the searches would fail before completing.

The ideal way to utilize this search experience is to click on the suggestions provided after entering “Bing generative search” into the search bar.

Potential Impact

Bing’s generative search results include citations and links to original sources.

Screenshot from: blogs.bing.com, October 2024.

This approach is intended to drive traffic to publishers, but it remains to be seen how effective this will be in practice.

Bing encourages users to provide feedback on the new feature using thumbs up/down icons or the dedicated feedback button.

Looking Ahead

This development comes as search engines increasingly use AI to enhance their capabilities.

As Bing rolls out this expanded generative search feature, remember the technology is still in beta, so performance and accuracy may vary.


Featured Image: JarTee/Shutterstock

Making SEO Decisions With Confidence: A Guide To Data-Driven Strategies via @sejournal, @AdamHeitzman

In SEO, making strategic decisions without empirical data is like relying on luck for consistent results.

But how can you effectively harness data to guide your SEO efforts and ensure you’re not just shooting in the dark?

This comprehensive guide will show you how to leverage data for confident, results-driven SEO strategies.

The Power Of Data-Driven SEO: A Case Study

Let’s start with a compelling example. Glassdoor.com, before its sale to Recruit Holdings in 2018, had an impressive 29,500,000 in monthly traffic – almost entirely organic.

Screenshot from Semrush, September 2024

Their success wasn’t by chance; it was the result of a meticulous, data-driven approach to SEO.

In her 2017 presentation at a marketing summit, Dawn Lyon, vice president of corporate affairs, shared how they weaved data from different internet sources.

Weave the biggest web possibleImage from Dawn Lyon/Glassdoor, September 2024

Glassdoor’s strategy involved analyzing data from various internet sources to identify content gaps and create high-value, well-optimized content.

This approach led to over 200,000,000 backlinks from more than 200,000 websites, establishing their influence and authority in the online employment industry.

The takeaway? Glassdoor used data to identify valuable content assets and gaps, creating content that brought them closer to their prospects.

This data-driven strategy significantly influenced their rankings in search results for the online employment industry.

What Types Of Data Are Important In SEO?

Before diving into strategies, it’s crucial to understand the types of data that matter in SEO:

Each of these data types provides unique insights that can inform your SEO strategy.

The Importance Of Data In SEO

Data takes the guesswork out of SEO, allowing you to focus on what works based on empirical evidence. For instance:

  • Keyword research data helps you understand your target audience’s pain points.
  • Bounce rate data can help you address issues affecting user engagement.
  • Engagement metrics show which content resonates with your audience.

How To Use Data In Your SEO Strategy

Now, let’s explore how to implement data-driven strategies in your SEO efforts:

1. Define Clear Objectives For SEO

Start by setting SMART (Specific, Measurable, Achievable, Relevant, Time-bound) SEO objectives. This helps you navigate the volume of available data and prioritize key areas for your campaigns.

SMART goalsScreenshot taken by author, September 2024

2. Establish Baseline Metrics And KPIs

Identify KPIs that align with your objectives and establish baseline metrics to measure current performance. This provides a reference point for evaluating the impact of your SEO strategy.

SEO KPIsScreenshot from Semrush, September 2024

3. Understand User Intent

Analyze search intent behind keywords to effectively optimize your content. Use tools like Google Search Console to track click-through rates (CTR) for individual pages, which can indicate how well your content matches user intent.

4. Choose High-Opportunity Keywords

Identify “low-hanging fruit” keywords with significant search volume and low to moderate competition. Evaluate their business potential before targeting them.

Business Potential: keyword researchScreenshot from Ahrefs, September 2024

5. Gather And Analyze Your Website Data

Use tools like Google Analytics to track your website traffic and user engagement metrics. This data can provide insights into your search performance and help you identify areas for improvement.

6. Conduct Competitor Analysis

Identify your SEO competitors and analyze their strategies. Tools like Semrush can help you find keyword gaps and backlink opportunities.

Screenshot from SemrushScreenshot from Semrush, September 2024

7. Create A Data-Driven SEO Strategy

Based on your gathered data, create an informed SEO strategy. This should include:

  • Creating your ideal customer profile.
  • Targeting the right keywords.
  • Conducting a site-wide audit.
  • Creating a content calendar.

8. Double-Down On High-Performing Keyword Categories

Identify which keywords drive the most organic traffic and conversions on your site. Use Google Search Console to see which terms rank highest and attract the most click-throughs from search results.

If you use a rank-tracking tool, combine this data with Google Analytics to see how pages perform in terms of traffic, engagement, and conversions.

Once you’ve identified your best-performing keywords, expand your content footprint within these high-value areas.

For example, if “beginner yoga poses” is a top performer, consider developing content for related terms like “yoga poses for flexibility,” “yoga routines for beginners,” and “best yoga mats for beginners.”

These “content clusters” around a topic will help you capture more traffic from thematically related keywords and can increase your domain’s overall authority for that topic area.

9. Analyze What Makes Your Best Content Effective

Examine your highest-performing content to identify factors that make it engaging for users. Consider aspects like:

  • Word count: Is longer content performing better, or do users prefer concise information?
  • Tone of voice: Is a casual, conversational tone more effective, or do users respond better to a formal, authoritative voice?
  • Presentation: How does the use of headings, bullet points, images, and other visual elements impact engagement?
  • Originality: Are unique insights or original research driving more engagement?
  • Expertise demonstrated: How does the level of experience, expertise, authoritativeness, and trustworthiness (E-E-A-T) impact performance?
  • Call-to-action (CTA): Analyze the clarity and appeal of your CTAs. Are they driving the desired user actions?

Understanding which of these variables plays a part will guide you in crafting future content that might mirror the same success.

10. Eliminate Friction From Your Conversion Paths

Analyze which user journeys lead to the highest levels of conversions. Look for commonalities in these high-converting paths and aim to replicate these elements across your site.

11. Prioritize Core Web Vitals

Use Google’s PageSpeed Insights to evaluate your site’s performance across Core Web Vitals metrics. Implement recommended fixes to improve your site’s user experience.

12. Enhance Your Site’s Mobile Usability

With mobile accounting for about 63% of organic search traffic in the U.S., optimizing for mobile users is crucial. Use Google’s Lighthouse tool to test your site’s mobile-friendliness and implement necessary improvements.

13. Analyze Backlinks For More Targeted Outreach

Study your site’s backlink data to optimize your link-building strategy. Use this information to tailor your outreach strategy and target high-authority websites that are likely to find your content valuable.

14. Collaborate With Cross-Functional Teams

Communicate the value of SEO to all stakeholders and align it with broader business goals. Integrate feedback from various teams to improve your SEO workflow efficiency.

15. Monitor And Iterate

Remember, SEO isn’t a set-it-and-forget-it strategy. Continuously monitor your progress and be prepared to iterate based on new data and insights.

Tools To Find SEO Data

To implement these strategies effectively, you’ll need the right tools. Here are some essential ones:

  • Google Analytics: For traffic data and user behavior insights. GA4 provides detailed information about your website visitors, including their demographics, interests, and how they interact with your site.
  • Google Search Console: For keyword research and onsite data. Google Search Console shows you how your site appears in Google search results and can help you identify and fix indexing problems.
  • Ahrefs: For backlink data and competitor analysis. It offers comprehensive insights into your backlink profile and helps you identify link-building opportunities.
  • Semrush: For comprehensive competitor data and keyword research. It’s particularly useful for understanding your competitors’ strategies and finding keyword gaps.
  • Screaming Frog SEO Spider: For technical SEO data. This tool crawls your website to identify technical issues that could be impacting your search engine performance.
  • PageSpeed Insights: For Core Web Vitals analysis. It provides both lab and field data about page performance, with suggestions for improvement.
  • Lighthouse: For mobile usability testing. This open-source tool audits performance, accessibility, progressive web apps, and more.

Remember, while these tools provide valuable data, the real power lies in how you interpret and act on this information.

Regularly review your data, look for trends and patterns, and use these insights to continuously refine your SEO strategy.

Leverage Data For Decision Making

Success in SEO isn’t luck or magic. With the right data, you can make informed strategies that cut through the noise and achieve better results on the search engine results pages (SERPs).

Remember, SEO is not just about theory – it’s about implementation. The final step of your data-driven decisions is to put your strategies into action and benchmark against your previous performance.

By leveraging data as a foundation for decision-making, you can create more effective SEO strategies.

From capitalizing on high-performing keywords to enhancing mobile usability and optimizing backlink strategies, each data-driven action you take helps solidify your online presence and improve your rankings.

Stay analytical, stay informed, and let the data illuminate your path to SEO success.

More resources: 


Featured Image: Deemerwha studio/Shutterstock

Is That SEO Course Worth Your Time? How To Tell via @sejournal, @WixStudio

This post was sponsored by Wix. The opinions expressed in this article are the sponsor’s own.

SEO courses seem to be everywhere. In that sense, they’ve become a dime a dozen. That’s generally because they’re both a lead magnet and a solution to a real problem.

To the latter, SEO isn’t an exact science, and it’s not like you can major in SEO at a traditional university. Instead, we have a disparate set of publications (like the one you are reading right now) offering collections of various resources and media from which to learn SEO.

Moreover, most practitioners “somehow” fall into SEO from other disciplines. So you’re really starting from scratch. It’s why a lot of SEOs start off following some pretty bad advice from some pretty questionable (yet highly popular) marketers.

It all makes for a pretty messed up learning curve when trying to “uplevel” (sorry for the marketing buzzword) your SEO know-how.

The solution? A set of organized material from knowledgeable sources built on sound pedagogical tactics—AKA a course.

The problem is, even if a course is free, it’s a serious time commitment. So before you fork over your money and/or your time, how do you know if a course is even worthwhile?

To help answer this question, I am going to take you behind the scenes of how I constructed our SEO course (while of course—pun intended—subtly selling you on the idea of taking our course as this is indeed a sponsored article; I just smashed like three “fourth walls” in one sentence).

Let’s get started. Here’s what I was looking for when I constructed the course and what you should look for when taking or recommending an SEO course.

1. Look For Course Instructors That Align with Specific Expertise

If you see a course and it’s just a bunch of random “big names” you’ve seen all over social media, run.

Yes, you do want a course created by experts in the field. However, it’s not as simple as having a “big name” on board.

To start, popularity on social media doesn’t always equate with actual SEO expertise. What you want to do is sniff out if the course instructors are there for the substance they provide or if this is just some sort of influencer marketing play.

When we went about creating our SEO course, one of the things I was adamant about was that the instructors we hired weren’t just experts in “SEO,” but experts in a specific type of SEO (local SEO, tech SEO, etc.). Meaning, it wasn’t just “Are these folks well-respected,” but are they known in the industry for focusing on that subtopic within wider SEO?

I was also looking for good communicators, not just the folks with the largest social followings. Being a good communicator on social media is different from being able to convey information in a more formal setting that relies on video as the medium.

Aside from areas of specialization, there are all sorts of practitioners who are used to doing SEO in different contexts. There are enterprise SEOs, in-house SEOs, consultants, SEOs who work at agencies, etc.

We thought it was important to have course instructors from all these backgrounds to offer a more complete set of approaches to various SEO considerations.

So if you’re looking into an SEO course, ask yourself:

  • Is there something to the instructors beyond their name recognition?
  • Do the instructors have strong subtopic alignment with their course sections?
  • Are the various subsets of SEO represented? (In-house, agency, etc.)
  • Do the instructors have previous presenting experience or just big social media followings?

1. There Is No ‘Best’ Course

If a course names itself something like “Best SEO Course”—run. There is no such thing as a “best” SEO course. There is a course that is great at dealing with advanced SEO topics, or courses that are geared towards specializing in something like tech SEO or local SEO, and there are courses that offer a wider breadth, etc.

The right course depends on your needs and objectives and where you are in your SEO journey. As such, the “best SEO course” is a fictitious construct.

To pull the curtain back, we were debating what to do with the term “best” for our course. The query “best SEO course” is certainly a lucrative one. Perhaps it would be strategic for our SEO to include the term “best” in the name of the course.

We decided not to do so even if there was an SEO benefit to it.

We had an idea of who we built the course for: We specifically wanted to help people having a hard time piecing together the topic and who are relying on doing so for their business needs and/or careers. (That’s not to say the course doesn’t apply to other demographics, but rather this is the primary pain point and audience we had in mind when designing the course).

This is basically because when we sat down to discuss the course, we all felt that when we were first starting out in SEO, it was a struggle to piece it all together.

This not only shows why there is no “best” SEO course, it also leads me to my next point.

2. Make Sure The People Behind The Course Are Genuine SEO Experts

Instructors are one thing. The people responsible for developing the course overall are another thing entirely. Now, I am not saying that if the people behind the course are not SEOs, they can’t create a wonderful course. What I am saying is that you better hope they gave creative control to the SEOs involved.

That’s not just for SEO accuracy per se (although that is a big part of it), it’s also because SEOs who have themselves struggled to grasp SEO concepts at some point in their career will better understand what to include in an SEO course.

Then, of course, there is the actual accuracy of the content. While the instructors may be a part of the process, they are far from in control of the course and what it ultimately looks like in post-production.

The question is, how can you tell if SEOs were involved in the backend of the course and to what extent?

There is no 100% tell-tale sign. However, I can say that from my experience working with our course, the way the course is structured might provide hints.

What do I mean?

Usually in an SEO course’s “on-page SEO” section, you would discuss concepts related to all things content, from E-E-A-T to strategy.

In our course, I purposefully did not put this course material under “On-Page SEO.” Rather, the deep dive into content, quality, and algorithms were placed under keyword research.

Why?

I felt that, often, the SEO industry thinks about topics like keyword research a bit too linearly and without enough depth and nuance. I wanted to contextualize keyword research by connecting it to a discussion about what creating quality content for the SERP looks like.

This is what I mean by looking at unique course structuring as a sign that there is real SEO expertise going into the material.

If a course follows an overly generic format, this (combined with other signals might) be a sign that folks without genuine SEO experience have too much input. Which is, obviously, not what you want.

3. Look For Signs of Pedagogy

This has nothing to do with SEO itself, but is purely about what it means to create a good course.

Yes, the curriculum needs to be accurate, but it also has to be delivered in a way that is conducive for learning. Our own process involved a lot of back and forth with our educational team to ensure that we structured everything from the assessments to the course scripts in a way that facilitates learning.

To give some context, we borrowed something that I used to do back in my teaching days—backward planning.

At the start of the process, we developed learning goals within the course section. These goals would form the basis for the assessments that we offer at the end of each course section. All of the course sections were built to fulfill those specific learning goals. In this way, the course itself directly aligns with the assessment, which is only fair.

On top of that, we made sure to use the assessments to extend the learning by mixing in scenario-based questions.

There are millions of ways to go about constructing a course that incorporates sound pedagogy. If you’re looking at a course and it all seems very linear, that might be a good indication that the course lacks pedagogical depth. Which is clearly not what you want, no matter how amazing the instructors listed are.

So when looking at an SEO course and deciding to dive in, don’t just look at it from an SEO perspective. Getting the SEO education right is only half the battle. The course also has to effectively communicate that information to you.

Look for signs of pedagogical life when choosing an SEO course.

Learning SEO can be hard. It can be a very informal process that leaves you wondering what gaps you might have and what you still need to learn.

The need for an SEO course can be real. There are a lot of great SEO courses out there. There are also a heap of “grifters” looking to take advantage of people who need a comprehensive way to learn SEO.

When it comes to signing up for an SEO course, if it feels too “markety” or too “salesy,” it probably is.

  • Look past the “certifications” every course offers (ours included). They’re nice, but no one is hiring you or giving you a raise because you have one.
  • Look past the big names a course may have procured.
  • Look past the overpromising (“Our users have improved their organic traffic by 1000000000000000000000000.9% in just 1 day after completing our course”).

Instead, think about what your specific needs are and if the course is suitable and substantial enough to help you fill those needs.

It pays to dig a bit deeper into a course and pull the curtain back a bit before investing money and in the case of a free course, time.

Ready to start optimizing your website? Sign up for Wix and get the data you need to deliver great user experiences.


Image Credits

Featured Image: Image by Wix. Used with permission.

Why Google Lighthouse Doesn’t Include INP, A Core Web Vital via @sejournal, @MattGSouthern

Google’s Lighthouse doesn’t use the Interaction to Next Paint (INP) metric in its standard tests, despite INP being one of the Core Web Vitals.

Barry Pollard, Web Performance Developer Advocate on Google Chrome, explained the reasoning behind this and offered insights into measuring INP.

Lighthouse Measures Page Loads, Not Interactions

Lighthouse measures a simple page load and captures various characteristics during that process.

It can estimate the Largest Contentful Paint (LCP) and Cumulative Layout Shift (CLS) under specific load conditions, identify issues, and advise on improving these metrics.

However, INP is different as it depends on user interactions.

Pollard explained:

“The problem is that Lighthouse, again like many web perf tools, typically just loads the page and does not interact with it. No interactions = No INP to measure!”

Custom User Flows Enable INP Measurement

While Lighthouse can’t measure INP, knowing common user journeys allows you to use “user flows” to measure INP.

Pollard added:

“If you as a site-owner know your common user journeys then you can measure these in Lighthouse using ‘user flows’ which then WILL measure INP.”

These common user journeys can be automated in a continuous integration environment, allowing developers to test INP on each commit and spot potential regressions.

Total Blocking Time As An INP Proxy

Although Lighthouse can’t measure INP without interactions, it can measure likely causes, particularly long, blocking JavaScript tasks.

This is where the Total Blocking Time (TBT) metric comes into play.

According to Pollard:

“TBT (Total Blocking Time) measures the sum time of all tasks greater 50ms. The theory being:

  • Lots of long, blocking tasks = high risk of INP!
  • Few long, blocking tasks = low risk of INP!”

Limitations Of TBT As An INP Substitute

TBT has limitations as an INP substitute.

Pollard noted:

“If you don’t interact during long tasks, then you might not have any INP issues. Also interactions might load MORE JavaScript that is not measure by Lighthouse.”

He adds:

“So it’s a clue, but not a substitute for actually measuring INP.”

Optimizing For Lighthouse Scores vs. User Experience

Some developers optimize for Lighthouse scores without considering the user impact.

Pollard cautions against this, stating:

“A common pattern I see is to delay ALL JS until the user interacts with a page: Great for Lighthouse scores! Often terrible for users 😢:

  • Sometimes nothing loads until you move the mouse.
  • Often your first interaction gets a bigger delay.”

Pollard’s Full Post

Why This Matters

Understanding Lighthouse, INP, and TBT relationships is necessary for optimizing user experience.

Recognizing limitations in measuring INP helps avoid misguided optimizations.

Pollard’s advice for measuring INP is to focus on real user interactions to ensure performance improvements enhance UX.

As INP remains a Core Web Vital, grasping its nuances is essential for keeping it within an acceptable threshold.

Practical Applications

To monitor site performance and INP:

  1. Use Lighthouse’s “user flows” for INP measurement in common journeys.
  2. Automate user flows in CI to monitor INP and catch regressions.
  3. Use TBT as an INP proxy, but understand its limitations.
  4. Prioritize field measurements for accurate INP data.
  5. Balance performance optimizations with UX considerations.

Featured Image: Ye Liew/Shutterstock