What Can Log File Data Tell Me That Tools Can’t? – Ask An SEO via @sejournal, @HelenPollitt1

For today’s Ask An SEO, we answer the question:

As an SEO, should I be using log file data, and what can it tell me that tools can’t?

What Are Log Files

Essentially, log files are the raw record of an interaction with a website. They are reported by the website’s server and typically include information about users and bots, the pages they interact with, and when.

Typically, log files will contain certain information, such as the IP address of the person or bot that interacted with the website, the user agent (i.e., Googlebot, or a browser if it is a human), the time of the interaction, the URL, and the server response code the URL provided.

Example log:

6.249.65.1 - - [19/Feb/2026:14:32:10 +0000] "GET /category/shoes/running-shoes/ HTTP/1.1" 200 15432 "-" "Mozilla/5.0 (Macintosh; Intel Mac OS X 14_2) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/121.0.0.0 Safari/537.36" 
  • 6.249.65.1This is the IP address of the user agent that hit the website.
  • 19/Feb/2026:14:32:10 +0000 – This is the timestamp of the hit.
  • GET /category/shoes/running-shoes/ HTTP/1.1 – The HTTP method, the requested URL, and the protocol version.
  • 200 – The HTTP status code.
  • 15432 – The response size in bytes.
  • Mozilla/5.0 (Macintosh; Intel Mac OS X 14_2) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/121.0.0.0 Safari/537.36 – The user agent (i.e., the bot or browser that requested the file)

What Log Files Can Be Used For

Log files are the most accurate recording of how a user or a bot has navigated around your website. They are often considered the most authoritative record of interactions with your website, though CDN caching and infrastructure configuration can affect completeness.

What Search Engines Crawl

One of the most important uses of log files for SEO is to understand what pages on our site search engine bots are crawling.

Log files allow us to see which pages are getting crawled and at what frequency. They can help us validate if important pages are being crawled and whether often-changing pages are being crawled with an increased frequency compared to static pages.

Log files can be used to see if there is crawl waste, i.e., pages that you don’t want to have crawled, or with any real frequency, are taking up crawling time when a bot visits a site. For example, by looking at log files, you may identify that parameterized URLs or paginated pages are getting too much crawl attention compared to your core pages.

This information can be critical in identifying issues with page discovery and crawling.

True Crawl Budget Allocation

Log file analysis can give a true picture of crawl budget. It can help with the identification of which sections of a site are getting the most attention, and which are being neglected by the bots.

This can be critical in seeing if there are poorly linked pages on a site, or if they are being given less crawl priority than those sections of the site with less importance.

Log files can also be helpful after the completion of highly technical SEO work. For example, when a website has been migrated, viewing the log files can aid in identifying how quickly the changes to the site are being discovered.

Through log files, it’s also possible to determine if changes to a website’s structure have actually aided in crawl optimization.

When carrying out SEO experiments, it is necessary to know if a page that is a part of the experiment has been crawled by the bots or not, as this can determine whether the test experience has been seen by them. Log files can give that insight.

Crawl Behavior During Technical Issues

Log files can also be useful in detecting technical issues on a website. For example, there are instances where the status code reported by a crawling tool will not necessarily be the status code that a bot will receive when hitting a page. In that instance, log files would be the only way of identifying that with certainty.

Log files will enable you to see if bots are encountering temporary outages on the site, but also how long it takes them to re-encounter those same pages with the correct status once the issue has been fixed.

Bot Verification

One very helpful feature of log file analysis is in distinguishing between real bots and spoofed bots. This is how you can identify if bots are accessing your site under the guise of being from Google or Microsoft, but are actually from another company. This is important because bots may be getting around your site’s security measures by claiming to be a Googlebot, whereas, in fact, they are looking to carry out nefarious actions on your site, like scraping data.

By using log files, it’s possible to identify the IP range that a bot came from and check it against the known IP ranges of legitimate bots, like Googlebot. This can aid IT teams in providing security for a website without inadvertently blocking genuine search bots that need access to the website for SEO to be effective.

Orphan Pages Discovery

Log files can be used to identify internal pages that tools didn’t detect. For example, Googlebot may know of a page through an external link to it, whereas a crawling tool would only be able to discover it through internal linking or through sitemaps.

Looking through log files can be useful for diagnosing orphan pages on your site that you were simply not aware of. This is also very helpful in identifying legacy URLs that should no longer be accessible via the site but may still be crawled. For example, HTTP URLs or subdomains that have not been migrated properly.

What Other Tools Can’t Tell Us That Log Files Can

If you are currently not using log files, you may well be using other SEO tools to get you partway to the insight that log files can provide.

Analytics Software

Analytics software like Google Analytics can give you an indication of what pages exist on a website, even if bots aren’t necessarily able to access them.

Analytics platforms also give a lot of detail on user behavior across the website. They can give context as to which pages matter most for commercial goals and which are not performing.

They don’t, however, show information about non-user behavior. In fact, most analytics programs are designed to filter out bot behavior to ensure the data provided reflects human users only.

Although they are useful in determining the journey of users, they do not give any indication of the journey of bots. There is no way to determine which sequence of pages a search bot has visited or how often.

Google Search Console/Bing Webmaster Tools

The search engines’ search consoles will often give an overview of the technical health of a website, like crawl issues encountered and when pages were last crawled. However, crawl stats are aggregated and performance data is sampled for large sites. This means you may not be able to get information on specific pages you are interested in.

They also only give information about their bots. This means it can be difficult to bring bot crawl information together, and indeed to see the behavior of bots from companies that do not offer a tool like a search console.

Website Crawlers

Website crawling software can help with mimicking how a search bot might interact with your site, including what it can technically access and what it can’t. However, they do not show you what the bot actually accesses. They can give information on whether, in theory, a page could be crawled by a search bot, but do not give any real-time or historical data on whether the bot has accessed a page, when, or how frequently.

Website crawlers are also mimicking bot behavior in the conditions you are setting them, not necessarily the conditions the search bots are actually encountering. For example, without log files, it is difficult to determine how search bots navigated a site during a DDoS attack or a server outage.

Why You Might Not Use Log Files

There are many reasons why SEOs might not be using log files already.

Difficulty In Obtaining Them

Oftentimes, log files are not straightforward to get to. You may need to speak with your development team. Depending on whether that team is in-house or not, this may literally mean trying to track down who has access to the log files first.

For teams working agency-side, there is an added complexity of companies needing to transfer potentially sensitive information outside of the organization. Log files can include personally identifiable information, for example, IP addresses. For those subject to rules like GDPR, there may be some concern around sending these files to a third party. There may be a need to sanitize the data before sharing it. This can be a material cost of time and resources that a client may not want to spend simply to share their log files with their SEO agency.

User Interface Needs

Once you have access to log files, it isn’t all smooth sailing from there. You will need to understand what you are looking at. Log files in their raw form are simply text files containing string after string of data.

It isn’t something that is easily parsed. To truly make sense of log files, there is usually a need to invest in a program to help decipher them. These can range in price depending on whether they are programs designed to let you run a file through on an ad-hoc basis, or whether you are connecting your log files to them so they stream into the program continuously.

Storage Requirements

There is also a need to store log files. Alongside being secure for the reasons mentioned above, like GDPR, they can be very difficult to store for long periods due to how quickly they grow in size.

For a large ecommerce website, you might see log files reach hundreds of gigabytes over the course of a month. In those instances, it becomes a technical infrastructure issue to store them. Compressing the files can help with this. However, given that issues with search bots can take several months of data to diagnose, or require comparison over long time periods, these files can start to get too big to store cost-effectively.

Perceived Technical Complexity

Once you have your log files in a decipherable format, cleaned and ready to use, you actually need to know what to do with them.

Many SEOs have a big barrier to using log files simply based on the fact they seem too technical to use. They are, after all, just strings of information about hits on the website. This can feel overwhelming.

Should SEOs Use Log Files?

Yes, if you can.

As mentioned above, there are many reasons why you may not be able to get hold of your log files and transform them into a usable data source. However, once you can, it will open up a whole new level of understanding of the technical health of your website and how bots interact with it.

There will be discoveries made that simply could not be achieved without log file data. The tools you are currently using may well get you part of the way there. They will never give you the full picture, however.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal

Should I Optimize My Content Differently For Each Platform? – Ask An SEO via @sejournal, @rollerblader

This week’s Ask an SEO question is from an anonymous reader who asks:

Should I be optimizing content differently for LinkedIn, Reddit, and traditional search engines? I’m seeing these platforms rank highly in Google results, but I’m not sure how to create a cohesive multi-platform SEO approach.”

Yes, you should absolutely be optimizing your content differently based on where you publish it, where you want to reach the audience, and the way they engage. This includes what you put out, what goes on your website, and what exists in your metadata. Each platform has a different user experience, and the people there go for different reasons, so your job with your content is to meet their needs where they are.

Metadata

For SEO purposes, you’re limited to a certain amount of pixels for meta titles and descriptions in a search result, whereas on social media platforms, you’re limited to a different number of characters. This means your titles and descriptions need to be modified to fit the pixel or character lengths defined by the platforms, including Open Graph, rich pins, etc.  The people on the platform may also be at different stages of their journey and be different audience demographics.

If the audience on one platform that has its own metadata elements, and it is younger or skews towards one gender, cater the text and imagery in your metadata towards them. It’s worth seeing if that resonates better, but only if that is the majority from that platform. For search engines, it can be anyone and any demographic, so make it a strong sales pitch that is all-inclusive. Use your customer service and review data to find out what matters to them and use it in your messaging. The same goes for the images you use.

What fits on Pinterest won’t look good on LinkedIn, and an image for Google Discover may not work great on Instagram. Pinterest can display a vertical infographic and make it look great, but it will be illegible on platforms that have squares and landscape-oriented images. Resize, change the wording, and ensure the focal point on the image matches the platform it’ll be used on via your metadata.

Search engines and social algorithms look for different things as well. A search engine may allow some clickbait and salesy types of titles and metadata, but social media algorithms may penalize sites that do this. And each platform will be using and looking for different signals.

This is why you want to speak to the audiences on the platforms and focus on what the platform rewards, not just a search engine. Your customers on TikTok may be younger and use different wording than your customers on Facebook, but both will need a balance of the wording on your webpage. This is where using unique metadata by platform and purpose matters.

Content On Your Own Pages

Not every page on your website has to be for SEO, AIO, or GEO, and neither does the user experience. If the page is for an email blast or remarketing where you have strong calls to action, less text, and more conversion, you can noindex it or use a canonical link to the detailed new customer experience page. The same goes for SEO vs. social media visitors.

Someone from social media may need more of an education when buying a product because they didn’t set out that day to buy it; they were on social media to have fun. Someone looking for a product, product + review, or a comparison has a background on the product and wants a solution, so they went to a search engine to find one. This is where an educational vs. a conversion option can happen, and both can exist without competing, even though they’re optimized for the same keyword phrase.

The schema, the way wording is used, and elements on the page like an “add to cart” button above the fold help search engines to know the page is for conversions, while an H1, H2, and text with internal links to product and content pages mean it’s for educational purposes. Now apply this to the goal of what you want to the person to do on the page to the page, and keep in mind where they are coming from before they reach it.

You may want a more visual approach with video demonstrations or reviews, and options to shop and learn more for some experiences, vs. giving them the product and a buy now button. Both are optimized for the same keywords, but both are there for different visitors. This is where you deduplicate them using your SEO skills.

The keywords and phrases will be similar in your title tag, H1 tag, and compete directly against the product or collection page, but the page is how people from Snapchat and Reddit engage vs. someone from an email blast that knows how your brand behaves. So, set the canonical link to the main product page and/or add a meta robots noindex,nofollow. When you’re pushing your content out, share the version of the page to the platform it is designed for. Your site structure and robots.txt guide the search engines and AI to the pages meant for them, helping to eliminate the cannibalization.

It is the same content, the same purpose, and the same goal, just a unique format for the platform you want traffic from. I wouldn’t recommend this for everything because it is a ton of work, but for important pages, products, and services, it can make a difference to provide a better UX based on what the person and platform prefer.

What You Post To The Platform

Last is the content you post to the platform. Some allow hashtags; others prefer a lot of words, and platforms like X or Bluesky restrict the number of words you can use unless you pay. The audiences on these platforms pay attention to and use different words, and the algorithms may reward or penalize content differently.

On LinkedIn and Reddit, you may want to share a portion of the post and a summary of what the person will learn, then encourage engagement and a click through to your website or app. On Facebook, you may do a snippet of text and a stronger call-to-action, as people aren’t there for networking and learning like they are on LinkedIn.

Reddit may also benefit from examples and trust builders, where YouTube Shorts is about a quick message that entices an interaction and ideally a click through. The written description on a YouTube Short may go ignored as it is hidden, so the video is more important message-wise. Reddit can also be people looking for real human experiences, reviews, and comparisons from real customers. So, if you engage and publish your content, look at the topic of the forum and meet the user on the page at that specific stage of their journey.

The description still matters on most of these platforms because they are algorithm-based, and so are the search engines that feature their content. The content here acts like food for the algorithms along with user signals, so make sure you write something that properly matches the video’s content and follows best practices. If you’re publishing to Medium or Reddit and want to get the comparison queries, focus on unbiased and fair comparisons or reviews so Google surfaces it (disclosing you are one of the brands if you are). Then focus your own pages on conversion copy so as the person is ready to buy a blue t-shirt, they see your conversion page.

You should change the content based on the platform, and even your own website, when the goal is to bring users in from a specific traffic source. Someone from social media may like a video, while someone from a search engine wants text. Just make sure you code and structure your pages correctly, and you cater the experience to the right platform so the users reach their correct experience.

This is not practical for every page, so do your best, and at a minimum, customize what you share publicly and what is in your metadata. Those are easy enough and fast enough to be able to be done at scale, then pay attention to the UX on the page and make adjustments as needed.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal

Is Your Internal Linking Helping Or Hurting Topical Authority? Ask An SEO via @sejournal, @HelenPollitt1

Today’s question is about understanding internal linking and how it can help or hinder a search engine’s perception of a page’s topical relevance and authority.

“How do you technically assess whether a site’s internal linking is diluting topical authority rather than strengthening it?”

What Is Topical Authority

In essence, topical authority is the concept of how a search engine may view a website’s ability to provide an authoritative answer for a topic, inferred from how consistently it covers that topic and how signals reinforce that coverage.

Although there is no single standard defined metric for topical authority, it is, in essence, a measure of a page or a whole website’s relevance to a specific knowledge area, and trustworthiness as a source of information.

How Is It Affected By Internal Links

Internal links are crucial in shaping topical authority. They influence how authority, relevance, and intent signals are distributed across a website or folder. If we think of backlinks as bringing topical authority into a website, internal linking then helps to disperse it across the site. Internal linking determines where that authority accumulates and aids search engines in interpreting a page’s topical focus.

Links that connect topically relevant pages together help to strengthen the perception of the destination page’s authority on a subject. Lots of links from pages that aren’t seemingly relevant to each other can dilute the destination’s topical authority.

Something that is central to understanding the role of internal links in shaping topical authority is PageRank. PageRank is an algorithmic system developed in the late ’90s by Google founders Larry Page and Sergey Brin. It was used to measure the importance of a page based on the nature and volume of the links pointing to it. We need to keep this concept in mind when considering the use of internal links to shape the perception of a page’s topical authority.

How Important Are Internal Links In Regard To Topical Authority?

There are several factors of internal links that can affect how beneficial they are in strengthening a page’s topical authority.

Does The Link Pass Authority?

The first aspect is whether the link is followable, or if it is marked as “rel=nofollow.” This also applies to other variations of the “nofollow” tag, like “rel=sponsored.” Note, these tags are hints and not absolutes and Google might ignore them in some cases.

The URL that the link is on, and the page it is pointing to, also need to be crawlable. If those pages are disallowed via the robots.txt, then the value of the authority will not pass, as the page will not be crawled for the internal link to be picked up by the search bots.

Where Is It Placed On The Page?

Where a link is on the page could affect its authority. For example, links placed in the footer of every page on the site, get weighted differently than those that sit within the page’s main content. Google’s Martin Splitt has explained that Google does treat content in different parts of the page differently when trying to understand the topic of a page, and its content that is perceived to be main content that is used most to help with that.

Google’s John Muller recently answered a question about how links are valued in these different areas of a page. He said, “I don’t think there is anything quantifiably different about internal links in different parts of the page.” Although that may seem to contradict Splitt’s comments, remember that Muller is addressing how the value of a link may be affected by its location on a page, whereas Splitt is discussing how location of content affects how it is weighted to determine topic.

Following this logic, links appearing in the main content of a page may affect how that link passes topical relevancy.

What Is The Anchor Text?

The anchor text, or alt-text in cases where an image is linked, will help to inform the search engines of the nature of the page being linked to. The words that form the link are critical in helping the user and search engines know what to expect when they land on the page it takes them to. This context is another signal to the bots of the link destination’s relevancy to a subject.

What Is The Link Pointing From And To?

Similarly, if a link is on a page that is topically similar to the page being linked to, that also reinforces the topical authority of the destination page. If Page A on my fictitious hobby ecommerce site is about different craft hobbies, and Page B is about textile craft hobbies, it will help to reinforce Page B’s relevance to those seeking information about craft hobbies.

How To Assess Your Internal Linking Structure’s Effect On Topical Authority

Internal links can help a site’s topical authority by reinforcing the destination URLs’ topical relevance. They also help to ensure that any external authority signals are being passed to the correct internal pages.

There are calculations that could factor in the flow of link equity and authority through pages to assess the full impact of internal linking on a page’s topical authority. Calculations required include assigning value for position of link placement, click-depth from a topically relevant and authoritative page and topical authority of the links to the page where the link is coming from.

It’s a lot of math.

Instead, I’m in favor of keeping it simple, and defining a process that will allow you to get enough of an understanding of your website’s topical authority to make decisions from.

By looking at a sample of pages from your site across different topics, or if you are particularly focused, just one area of your topical authority, you can get an idea of any issues.

1. Identify Where Your Pages Are Getting Their Internal Links From

First of all, crawl your site, taking a sample of URLs. Export all of the internal links pointing to those pages, including their anchor text and URL the link is on.

2. Classify The URLs In Topic Clusters

Group all the pages into topical themes, i.e., for an ecommerce site that sells hobby equipment, “knitting, crochet, embroidery, and weaving” would all sit within “crafts” and the sub-category of “textile arts.” “Die cutting, digital cutting, laser cutting” would all sit within “crafts” and the sub-category of “cutting and engraving.”

3. Analyze What Proportion Of Each URL’s Followable Internal Links Are From Within The Same Topic And  Outside Of The Topic

Using the exported links, for follow links only, match them against the URLs and mark them as “within” or “outside” their topical family

Divide the volume of links that are from the same topic by the volume of links in total. For example, for “examplehobbyshop.com/crafts/embroidery/intro-to-embroidery/, if the total number of internal links is 100 and the volume of internal links from categories that are within the “craft” family is 60, then it would be 60/100 = 60%

The rule I apply is, if the URL internal links from the same family are around 75% or higher, that suggests that internal links are helping solidify topical authority. If it is less than 74%, that suggests that there could be some improvement.

How To Assess How Your Links’ Anchor Text Is Contributing To Your Topical Authority

1. Extract The Anchor Text Of Links Pointing To Your URLs

When gathering the links pointing to a page, remove common links like static header navigation and footer links that stay the same on each page. Then, extract the anchor text or alt text for linked images.

2. Categorize The Relevance Of The Anchor Text Of Links

Next, you want to look at how on-topic the anchor text of the links is for the page they are linking to.

Classify each anchor text as “topically relevant,” “topically irrelevant,” or “generic.” Topically relevant anchor text will have great alignment with the subject of the linked-to page. Topically irrelevant anchor text will not show any useful reinforcement of the topic. “Generic” anchor text includes “click here” or pagination links.

For the URL, examplehobbyshop.com/crafts/embroidery/intro-to-embroidery/, the following internal links’ anchor text could be grouped as follows:

Topically relevant Topically irrelevant Generic
“get started with embroidery”

“learn the tools needed to pick up embroidery”

“want to try another fibre craft?”

“beginners’ guide”

“start a new hobby”

“try something new”

“click here”

“next”

“page 2”

The goal is to have a lot of links from topically relevant pages pointing to the URL using topically relevant anchor text.

Measure the relevance of the anchor text against the total volume of anchor text.

For example, if that page had 30 topically relevant anchor texts, 20 topically irrelevant, and 50 generic, of the total 100 internal links pointing to it, it would have a topically relevant anchor text score of 30%. So despite there being a high volume (60%) of relevant internal links pointing to it, only 30% of the links have topically relevant anchor text.

3. Identify The Intent Mix Of The Anchor Text

Next, you want to identify the intent of the anchor text.

When grouping the anchor text by topical relevancy, also consider the intent behind the anchor text. For example, is it suggesting the page you will go to after clicking on it is informational, commercial, or transactional?

This matters because it can lead to dilution of the page intent. If there is a wide spread of intent shown through the anchor text, it can lead to confusion as to the purpose of the page being linked to.

Following on from the previous example, if some of the internal links had the anchor text “learn more about embroidery,” but others were more akin to “buy all the tools you need for your first embroidery project,” it’s not clear if examplehobbyshop.com/crafts/embroidery/intro-to-embroidery/ is an informational, commercial, or transactional page. This suggests the anchor text has a high intent mix, which is not ideal. If the majority of the anchor text were aligned with informational intent, it would have low intent mix.

Together, you want the anchor text to show high topical relevance, and low intent mix.

Final Thoughts

By the end of your analysis, you should have an idea of the topical relevance of the source pages of the internal links and how their anchor text aligns to both the topic and intent of the page being linked to.

Scaling this across a larger volume of URLs means you can start to see how topical relevance and authority are being strengthened or diluted via internal linking.

Once you have an idea of weaker areas of your site, you can begin to optimize anchor text and link sources to reinforce the value of the linked-to page as a source of authority on a subject.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal

Ask An SEO: Can AI Systems & LLMs Render JavaScript To Read ‘Hidden’ Content? via @sejournal, @HelenPollitt1

For this week’s Ask An SEO, a reader asked:

“Is there any difference between how AI systems handle JavaScript-rendered or interactively hidden content compared to traditional Google indexing? What technical checks can SEOs do to confirm that all page critical information is available to machines?”

This is a great question because beyond the hype of LLM-optimization sits a very real technical challenge: ensuring your content can actually be found and read by the LLMs.

For several years now, SEOs have been fairly encouraged by Googlebot’s improvements in being able to crawl and render JavaScript-heavy pages. However, with the new AI crawlers, this might not be the case.

In this article, we’ll look at the differences between the two crawler types, and how to ensure your critical webpage content is accessible to both.

How Does Googlebot Render JavaScript Content?

Googlebot processes JavaScript in three main stages: crawling, rendering, and indexing. In a basic and simple explanation, this is how each stage works:

Crawling

Googlebot will queue pages to be crawled when it discovers them on the web. Not every page that gets queued will be crawled, however, as Googlebot will check to see if crawling is allowed. For example, it will see if the page is blocked from crawling via a disallow command in the robots.txt.

If the page is not eligible to be crawled, then Googlebot will skip it, forgoing an HTTP request. If a page is eligible to be crawled, it will move to render the content.

Rendering

Googlebot will check if the page is eligible to be indexed by ensuring there are no requests to keep it from the index, for example, via a noindex meta tag. Googlebot will queue the page to be rendered. The rendering may happen within seconds, or it may remain in the queue for a longer period of time. Rendering is a resource-intensive process, and as such, it may not be instantaneous.

In the meantime, the bot will receive the DOM response; this is the content that is rendered before JavaScript is executed. This typically is the page HTML, which will be available as soon as the page is crawled.

Once the JavaScript is executed, Googlebot will receive the fully constructed page, the “browser render.”

Indexing

Eligible pages and information will be stored in the Google index and made available to serve as search results at the point of user query.

How Does Googlebot Handle Interactively Hidden Content?

Not all content is available to users when they first land on a page. For example, you may need to click through tabs to find supplementary content, or expand an accordion to see all of the information.

Googlebot doesn’t have the ability to switch between tabs, or to click open an accordion. So, making sure it can parse all the page’s information is important.

The way to do this is to make sure that the information is contained within the DOM on the first load of the page. Meaning, content may be “hidden from view” on the front end before clicking a button, but it’s not hidden in the code.

Think of it like this: The HTML content is “hidden in a box”; the JavaScript is the key to open the box. If Googlebot has to open the box, it may not see that content straightaway. However, if the server has opened the box before Googlebot requests it, then it should be able to get to that content via the DOM.

How To Improve The Likelihood That Googlebot Will Be Able To Read Your Content

The key to ensuring that content can be parsed by Googlebot is making it accessible without the need for the bot to render the JavaScript. One way of doing this is by forcing the rendering to happen on the server itself.

Server-side rendering is the process by which a webpage is rendered on the server rather than by the browser. This means an HTML file is prepared and sent to the user’s browser (or the search engine bot), and the content of the page is accessible to them without waiting for the JavaScript to load. This is because the server has essentially created a file that has rendered content in it already; the HTML and CSS are accessible immediately. Meanwhile, JavaScript files that are stored on the server can be downloaded by the browser.

This is opposed to client-side rendering, which requires the browser to fetch and compile the JavaScript before content is accessible on the webpage. This is a much lower lift for the server, which is why it is often favored by website developers, but it does mean that bots struggle to see the content on the page without rendering the JavaScript first.

How Do LLM Bots Render JavaScript?

Given what we now know about how Googlebot renders JavaScript, how does that differ from AI bots?

The most important element to understand about the following is that, unlike Googlebot, there is no “one” governing body that represents all the bots that might be encompassed under “LLM bots.” That is, what one bot might be capable of doing won’t necessarily be the standard for all.

The bots that scrape the web to power the knowledge bases of the LLMs are not the same as the bots that visit a page to bring back timely information to a user via a search engine.

And Claude’s bots do not have the same capability as OpenAI’s.

When we are considering how to ensure that AI bots can access our content, we have to cater to the lowest-capability bots.

Less is known about how LLM bots render JavaScript, mainly because, unlike Google, the AI bots are not sharing that information. However, some very smart people have been running tests to identify how each of the main LLM bots handles it.

Back in 2024, Vercel published an investigation into the JavaScript rendering capabilities of the main LLM bots, including OpenAI’s, Anthropic’s, Meta’s, ByteDance’s, and Perplexity’s. According to their study, none of those bots were able to render JavaScript. The only ones that were, were Gemini (leveraging Googlebot’s infrastructure), Applebot, and CommonCrawl’s CCbot.

More recently, Glenn Gabe reconfirmed Vercel’s findings through his own in-depth analysis of how ChatGPT, Perplexity, and Claude handle JavaScript. He also runs through how to test your own website in the LLMs to see how they handle your content.

These are the most well-known bots, from some of the most heavily funded AI companies in this space. It stands to reason that if they are struggling with JavaScript, lesser-funded or more niche ones will be also.

How Do AI Bots Handle Interactively Hidden Content?

Not well. That is, if the interactive content requires some execution of JavaScript, they may struggle to parse it.

To ensure the bots are able to see content hidden behind tabs, or in accordions, it is prudent to ensure the content loads fully in the DOM without the need to execute JavaScript. Human visitors can still interact with the content to reveal it, but the bots won’t need to.

How To Check For JavaScript Rendering Issues

There are two very easy ways to check if Googlebot is able to render all the content on your page:

Check The DOM Through Developer Tools

The DOM (Document Object Model) is an interface for a webpage that represents the HTML page as a series of “nodes” and “objects.” It essentially links a webpage’s HTML source code to JavaScript, which enables the functionality of the webpage to work. In simple terms, think of a webpage as a family tree. Each element on a webpage is a “node” on the tree. So, a header tag

, a paragraph

, and the body of the page itself are all nodes on the family tree.

When a browser loads a webpage, it reads the HTML and turns it into the family tree (the DOM).

How To Check It

I’ll take you through this using Chrome’s Developer Tools as an example.

You can check the DOM of a page by going to your browser. Using Chrome, right-click and select “Inspect.” From there, make sure you’re in the “Elements” tab.

To see if content is visible on your webpage without having to execute JavaScript, you can search for it here. If you find the content fully within the DOM when you first load the page (and don’t interact with it further), then it should be visible to Googlebot and LLM bots.

Use Google Search Console

To check if the content is visible specifically to Googlebot, you can use Google Search Console.

Choose the page you want to test and paste it into the “Inspect any URL” field. Search Console will then take you to another page where you can “Test live URL.” When you test a live page, you will be presented with another screen where you can opt to “View tested page.”

How To Check If An LLM Bot Can See Your Content

As per Glenn Gabe’s experiments, you can ask the LLMs themselves what they can read from a specific webpage. For example, you can prompt them to read the text of an article. They will respond with an explanation if they cannot due to JavaScript.

Viewing The Source HTML

If we are working to the lowest common denominator, it is prudent to assume, at this point, LLMs can’t read content in JavaScript. To make sure that your content is available in the HTML of a webpage so that the bots can definitely access it, be absolutely sure that the content of your page is readable to these bots. Make sure it is in the source HTML. To check this, you can go to Chrome and right click on the page. From the menu, select “View page source.” If you can “find” the text in this code, you know it’s in the source HTML of the page.

What Does This Mean For Your Website?

Essentially, Googlebot has been developed over the years to be much better at handling JavaScript than the newer LLM bots. However, it’s really important to understand that the LLM bots are not trying to crawl and render the web in the same way as Googlebot. Don’t assume that they will ever try to mimic Googlebot’s behavior. Don’t consider them “behind” Googlebot. They are a different beast altogether.

For your website, this means you need to check if your page loads all the pertinent information in the DOM on the first load of the page to satisfy Googlebot’s needs. For the LLM bots, to be very sure the content is available to them, check your static HTML.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal

Ask An SEO: What Is The Threshold Between Keyword Stuffing & Being Optimized? via @sejournal, @rollerblader

In this week’s Ask An SEO, Bre asks:

“What is the threshold between keyword stuffing and being optimized? Is there a magic rule for how often to use your main keyword and related keywords in a 2,000-word page? Should the main keyword be in the Headers AND the body in the same section?”

Great question!

There is no such thing as “being optimized” when it comes to keywords and repetitions. This is similar to looking at “authority” scores for domains. The optimization scores you get are measurements based on what an SEO tool thinks gives a domain trust, and not the actual search engines or LLM and AI systems. The idea of a keyword needing to be repeated is from an SEO concept called keyword density, which is a result of SEO tools.

Each tool would have a different way to say if you repeated a word or phrase enough for it to be “SEO friendly,” and because people trust the tools, they trust that this is a valid ranking factor or signal for a search engine. It is not because the search engines do not pay attention to how many times a word is on a page or in a paragraph, as that doesn’t produce a good experience.

Panda reduced the effectiveness of low-quality, keyword-stuffed content, and Google’s later advancements, BERT and MUM, allowed better understanding of context, relationships between terms, and the overall structure of a page. Google is now far better at interpreting meaning without relying on repeated exact-match keywords.

With that said, keywords are important.

Keywords help to send a signal to a search engine about the topic of the page. And they can be used in headers, within text, as internal links, within title tags, schema, and the URL structure. But worrying about using the keyword for SEO purposes can lead to trouble. So, let’s define keyword stuffing for the sake of this post.

Keyword stuffing is when you force a keyword or keyword phrase into content, headers, and URLs for the sole purpose of SEO.  

By forcing a keyword into a post, or forcing it into headers, you hurt the user experience. Although the search engine will know what you want to rank for, the language won’t feel natural. Instead of worrying about how many times you say the keyword, think about synonyms and other ways to say things that are easy to understand. Many search engines are getting better and better at understanding how topics, words, sentences, and phrases relate to one another. You don’t have to repeat the same words over and over anymore.

If you Google the word “swimsuit,” you’ll likely see it in a couple of title tags, but also see “swimwear.” Now type “bathing suits” in, you’ll likely not see it in a ton of the title tags, but the title tags will say “swimwear” and other synonyms, even though “bathing suits” is a popular name for the same product.

Now try “hairdresser near me,” and you’ll likely not see “hairdresser” in a lot of the results, but you will see “hair salon” and similar types of businesses. This is because search engines produce solutions to problems, and if they understand the page has the solution, you don’t need to keep repeating keywords.

For example, instead of saying “keyword stuffing” in this post, I could say “overusing phrases for SEO.” It means the same thing. Readers on this column will get bored pretty fast if I keep saying keyword stuffing, and by mixing it up, I can keep their interest, and search engines are still able to determine it is one-in-the-same. This also applies to header tags.

I don’t have any solid proof of this, but it seems to work well for our clients and the content we create, and it has worked for more than 10 years. If the main keyword phrase is in the H1 tag, whether it is a menu item or a blog post, we don’t worry about placing it in H2, H3, etc. I won’t be upset if the keyword shows up naturally, as that creates a good UX.

The theory here is that headers carry the theme and topic through the sections below. If the top-level header has the word “blue” in it, I make the assumption that theme “blue” carries through the page and applies to the H2 tag as the H2 is a sub-topic of “blue.” “H2’s” for blue could be “t-shirts” and “shorts.”

If this is true, by having the H1 be “blue” and the H2 be “shorts,” a search engine will know they are “blue shorts,” and I feel very confident users will too. They clicked blue or found a SERP for blue clothing, and they clicked shorts from the menu or found them from scrolling.

If you stuff “blue” into each link and header, it is annoying for the user to see it over and over. But many sites that get penalized will have “blue cargo shorts,” “blue chino shorts,” “blue workout shorts,” etc. It looks nicer to just say the styles of shorts like “cargo” or “chino,” and search engines likely already know they’re blue because you had it in the H tag one level up. You also likely have the “blue” part in breadcrumbs, site structure, product descriptions, etc.

One thing you definitely do not want to do is have a million footer links that match the navigation or are keyword-stuffed. This worked a long time ago, but now it is just spam. It doesn’t benefit the user; it is obvious to search engines you’re doing it for SEO. Sites that stuff keywords tend to use these outdated tactics too, so I want to include it here.

I hope this helps answer your question about overusing specific topics or phrases. Doing this only makes the tool happy; it does not mean you’ll be creating a good UX for users or search engines. If you focus on writing for your consumer and incorporate a keyword or phrase naturally, you’ll likely be rewarded.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal

Ask An SEO: Digital PR Or Traditional Link Building, Which Is Better? via @sejournal, @rollerblader

This week’s ask an SEO question is:

“Should SEOs be focusing more on digital PR than traditional link building?”

Digital PR is synonymous with link building at this point as SEO’s needed a new way to package and resell the same service. Actual PR work will always be more valuable than link building because PR, whether digital or traditional, focuses on a core audience of customers and reaching specific demographics. This adds value to a business and drives revenue.

With that said, here’s how I’d define digital PR vs. link building if a client asked what the difference is.

  • Digital PR: Getting brand coverage and citations in media outlets, niche publications, trade journals, niche blogs, and websites that do not allow guest posting, paid links, or unvetted contributors with the goal of building brand awareness and driving traffic from the content.
  • Link Building: Getting links from websites as a way to try and increase SERP rankings. Traffic from the links, sales from the links, etc., are not being tracked, and the quality of the website can be questionable.

Digital PR is always going to be better than link building because you’re treating the technique as a business and not a scheme to try and game the rankings. Link building became a bad practice years ago as links became less relevant, they are still important, so I want to ensure that isn’t taken out of context, and we stopped doing link building completely. Quality content attracts links naturally, including media mentions. When this happens in a natural way, the website will begin rising as the site has a lot of value for users, and search engines can tell when the site is quality.

If you’re building links without evaluating the impact they have traffic and sales-wise, you’re likely setting your site up for failure. Getting a ton of links, just like creating content in mass with AI/LLMs or article spinners, can grow a site quickly. That URL/domain can then burn to the ground equally as fast.

That’s why when we purchase a link, an advertorial, or we’re doing a partnership, we always ask ourselves the following questions:

  • Is there an active audience on this website that is also coming back to the website via branded search for information?
  • Is the audience on this website part of our customer base?
  • Will the article we’re pitching or being featured in be helpful to the user, and is our product or service something that is part of the post naturally vs. being forced?
  • Are we ok with the link being nofollow or sponsored if we’re paying for the inclusion?

If the answer is yes to these four, then we’re good to go with the link. The active audience on the website and people returning by brand name means there is an audience that trusts them for information. If the readership, visitors, or customers are similar or the same demographics as our user base, then it makes sense we’d want to be in front of them where they go for information.

We may have knowledge that is helpful to the user, but if it is not on topic within the post, there is no reason for them to come through and use our services, buy our products, or subscribe to our newsletters. Instead, we’ll wait until there is a fit, so there is a direct “link” between the content we’re contributing, or being an expert on, and our website.

For the last question, our goal is always traffic and customer acquisition, not getting a link. The website owner controls this, and if they want to follow Google’s best practices (which we obviously recommend doing), we will still be happy if they mark it as sponsored or nofollow. This is the most important of the questions. Building links to game the SERPs is a bad idea; building a brand that people search for by name will overpower any link any day of the week. This is always our goal when it comes to Digital PR and link building. Driving that branded search.

So, that begs the question, where do we go for digital PR?

Sources To Get Digital PR Mentions And Links

When we’re about to start a Digital PR campaign, we create lists of the following targets to reach out to.

  • Mass Media: Household names like magazines, news websites, and local media, where everyone in the area, the customers, or the country or world knows them by name. The only stipulation we apply is if they have an active category vs. only a few articles here and there. The active category means it is something interesting enough to their reader base that they’re investing in it, so our customers may be there.
  • Trade Publications: Conferences, associations, and non-profits, as well as industry insiders will have websites and print publications that go out to members. Search Engine Journal could be considered a trade publication for the SEO and PPC industry, same with SEO Roundtable, and some of the communities like Webmaster World. They publish directly relevant content for search engine marketers and have active users, so if I was an SEO service provider or tool, this is where I’d be looking to get featured and ideally links from.
  • Niche Sites and Bloggers: There is no shortage of niche sites and content producers out there. The trick is finding ones that do not publicly allow guest contributions, advertorials, etc., and that do not link out to non-niche websites and content. This includes sites that got hacked and had link injections. Even if their “authority” is zero, there is value if they quality control and all links and mentions are earned.
  • Influencers: Whether it is YouTube, Facebook group leaders, LinkedIn that is crawlable, or other channels, getting coverage from people with subscribers and an active audience can let search engines crawl the link back to your website. It may not boost your rankings, but it drives customers to you and helps with page discoverability if the link gets crawled. LLMs are also citing their content as sources, so there could be value for AIO, too.

Link building is not dead by any means; links still matter. You just don’t need to build them anymore. Focus on quality where an active audience is and where you have a chance at getting traffic and revenue. This is what will move the needle for the long run and help you grow in SERPs that matter.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal

Ask an SEO: Is An XML Or HTML Sitemap Better For SEO? via @sejournal, @HelenPollitt1

In this edition of Ask An SEO, we break down a common point of confusion for site owners and technical SEOs:

Do I need both an XML sitemap and an HTML one, and which one is better to use for SEO?

It can be a bit confusing to know whether it’s better to use an XML sitemap or an HTML one for your site. In some instances, neither is needed, and in some, both are helpful. Let’s dive into what they are, what they do, and when to use them.

What Is An XML sitemap?

An XML sitemap is essentially a list of URLs for pages and files on your website that you want the search bots to be able to find and crawl. You can also use the XML sitemap to detail information about the files, like the length of run-time for the video file specified, or the publication date of an article.

It is primarily used for bots. There is little reason why you would want a human visitor to use an XML sitemap. Well, unless they are debugging an SEO issue!

What Is The XML Sitemap Used For?

The purpose of the XML sitemap is to help search bots understand which pages on your website should be crawled, as well as giving them extra information about those pages.

The XML sitemap can help bots identify pages on the site that would otherwise be difficult to find. This can be orphaned pages, those with low internal links, or even pages that have changed recently that you may want to encourage the bots to recrawl.

Best Practices For XML Sitemaps

Most search bots will understand XML sitemaps that follow the sitemaps.org protocol. This protocol defines the necessary location of the XML sitemap on a site, schema it needs to use to be understood by bots, and how to prove ownership of domains in the instance of cross-domain references.

There is typically a limit on the size an XML sitemap can be, and still be parsed by the search bots. This means when building an XML sitemap, you should ensure it is under 50 MB uncompressed, and no more than 50,000 URLs. If your website is larger, you may need multiple XML sitemaps to cover all of the URLs. In that instance, you can use a sitemap index file to help organize your sitemaps into one location.

As the purpose of the XML sitemap is typically to help bots find your crawlable, indexable pages, it is usually necessary to ensure the file references it contains all lead to URLs with 200 server response codes. In most instances, the URLs should be the canonical version, and not contain any crawl or index restrictions.

Things To Be Aware Of With XML Sitemaps

There may be good reasons to go against “best practice” for XML sitemaps. For example, if you are instigating a lot of redirects, you may wish to include the old URLs in an XML sitemap even though they will return a 301 server response code. Adding a new XML sitemap containing those altered URLs can encourage the bots to recrawl them and pick up the redirects sooner than if they were just left to find them via crawling the site. This is especially the case if you have gone to the trouble of removing links to the 301 redirects on the site itself.

What Is An HTML Sitemap?

The HTML sitemap is a set of links to pages within your website. It is usually linked to from somewhere on the site, like the footer, that is easily accessed by users if they are specifically looking for it. However, it doesn’t form part of the main navigation of the site, but more as an accompaniment to it.

What Is An HTML Sitemap Used For?

The idea of the HTML sitemap is to serve as a catch-all for navigation. If a user is struggling to find a page on your site through your main navigation elements, or search, they can go to the HTML sitemap and find links to the most important pages on your site. If your website isn’t that large, you may be able to include links to all of the pages on your site.

The HTML sitemap pulls double duty. Not only does it work as a mega-navigation for humans, but it can also help bots find pages. As bots will follow links on a website (as long as they are followable), it can aid in helping them to find pages that are otherwise not linked to, or are poorly linked to, on the site.

Best Practices For HTML Sitemaps

Unlike the XML sitemap, there is no specific format that an HTML sitemap needs to follow. As the name suggests, it tends to be a simple HTML page that contains hyperlinks to the pages you want users to find through it.

In order to make it usable for bots too, it is important that the links are followable, i.e., they do not have a nofollow attribute on them. It is also prudent to make sure the URLs they link to aren’t disallowed through the robots.txt. It won’t cause you any serious issues if the links aren’t followable for bots; it just stops the sitemap from being useful for bots.

Things To Be Aware Of With HTML Sitemaps

Most users are not going to go to the HTML sitemap as their first port of call on a site. It is important to realize that if a user is going to your HTML sitemap to find a page, it suggests that your primary navigation on the site has failed them. It really should be seen as a last resort to support navigation.

Which Is Better To Use For SEO?

So, which is more important for SEO? Well, neither. That is, it really is dependent on your website and its needs.

For example, a small website with fewer than 20 pages may not have a need for either an XML sitemap or an HTML sitemap. In this instance, if all the pages are linked to well from the main navigation system, the chances are high that users and search bots alike will easily be able to find each of the site’s pages without additional help from sitemaps.

However, if your website has millions of pages, and has a main navigation system that buries links several sub-menus deep, an XML sitemap and an HTML sitemap may be useful.

They both serve different purposes and audiences.

When To Use The XML Sitemap

In practice, having an XML sitemap, or several, can help combat crawl issues. It gives a clear list of all the pages that you want a search bot to crawl and index. An XML sitemap can also be very helpful for debugging crawling issues, as when you upload it to Google Search Console, you will get an alert if there are issues with it or the URLs it contains. It can allow you to narrow in on the indexing status of URLs within the XML sitemap. This can be very useful for large websites that have millions of pages.

Essentially, there isn’t really a reason not to use an XML sitemap, apart from the time and cost of creating and maintaining them. Many content management systems will automatically generate them, which can take away some of the hassle.

Really, if you can have an XML sitemap, you might as well. If, however, it will be too costly or developer-resource intensive, it is not critical if your site is fairly small and the search engines already do a good job of crawling and indexing it.

When To Use The HTML Sitemap

The HTML sitemap is more useful when a website’s navigation isn’t very intuitive, or the search functionality isn’t comprehensive. It serves as a backstop to ensure users can find deeply buried pages. An HTML sitemap is particularly useful for larger sites that have a more complicated internal linking structure. It can also show the relationship between different pages well, depending on the structure of the sitemap. Overall, it is helpful to both users and bots, but is only really needed when the website is suffering from architectural problems or is just exceedingly large.

So, in summary, there is no right or wrong answer to which is more important. It is, however, very dependent on your website. Overall, there’s no harm in including both, but it might not be critical to do so.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal

Ask An SEO: Do I Need To Rethink My Content Strategy For LLMs? via @sejournal, @MordyOberstein

For this week’s Ask An SEO, the question asked was:

“Do I need to rethink my content strategy for LLMs and how do I get started with that?”

To answer, I’m going to explain the non-linear journey down the customer journey funnel and where large language models (LLMs) show up.

From rethinking traffic expectations to conducting an audit on sentiment picked up by LLMs, I will talk about why brand identity matters in building the kind of reputation that both users and machines recognize as authoritative.

You can watch this week’s Ask An SEO video and read the full transcript below.

Editor’s note: The following transcript has been edited for clarity, brevity, and adherence to our editorial guidelines.

Don’t Rush Into Overhauling Your Strategy

Off the bat, I strongly advise not to rush into this. I know there’s an extreme amount of noise and buzz and advice out there on social media that you need to rethink your strategy because of LLMs, but this thing is very, very far from settled.

For example, or most notably, AI Mode is still not in traditional search results. When that happens, when Google moves the AI Mode tab from being a tab into the main search results, the whole ecosystem is set for another upheaval, whatever that looks like, because we don’t actually know what that will look like.

I personally think that Google’s Gemini demo (the one they did way, way back, where they showed customized results for certain types of queries with certain answer formats) might be what AI Mode ends up resembling more than what it does right now, which is purely a text-based output that sort of aligns with ChatGPT.

I think Google will differentiate those two products once it moves AI Mode over from the tab into the main search results. So, things are not settled yet. And if you think they’re not. They are not settled yet.

Rethinking Traffic Expectations From LLMs

The other thing I want you to rethink is the traffic expectations from LLMs.

There’s been a lot of talk about citations and traffic – citations and traffic, citations and traffic. I don’t think citations, and therefore traffic, are the main diamond within the LLM ecosystem. I believe mentions are. And I don’t think that’s anything really new, by the way.

Traditionally, the funnel has been messy, and Google’s been talking about that for a long time. Now, you have an LLM that may be a starting point or a step in that messy funnel, but I don’t believe it’s fundamentally different.

I’ll give you an example. If I’m looking for a pair of shoes, I might go to Google and search, [Are these Nike shoes any good?]. I might look at a website, then go to Amazon and look at the actual product.

Then I might go to YouTube, see a review of the product, maybe watch a different one, go back to Amazon, have a look, check Google Shopping to see if it’s cheaper there, and then head back to Amazon to buy it.

Now, you have an LLM thrown into the mix, and that’s really the main difference. Maybe now, the LLM gives me the answer. Or maybe Google gives me the answer. Then I go to Amazon, look at the product, go to Google Shopping to see if it’s cheaper, watch a YouTube review, maybe switch things up a bit, go back to ChatGPT, see if it recommends something different this time, go through the whole process, and eventually buy on Amazon. That’s just me, personally.

It’s important to realize that the paradigm has been around for a while. But if you’re thinking of LLMs as a source of traffic, I highly recommend you don’t. They are not necessarily built for that.

ChatGPT, specifically, is not built for citations or to offer traffic. It’s built to provide answers and to be interactive. You’ll notice you usually don’t get a citation in ChatGPT until the third, fourth, or fifth prompt, whatever it is.

Other LLMs, like AI Mode or Perplexity, are a little bit more citation or link-based, but still, their main commodity is the output, giving you the answer and the ability to explore further.

So, I’m a big believer that the brand mention is far more important than the actual citation, per se. Also, the citation might just be the source of information. If I’m asking, “Are Nike shoes good?” I might get a review from a third-party website, say, the CNET of shoes, and even if I click there, that’s not where I’m going to buy the actual shoe.

So, the traffic in that case isn’t even the desirable outcome for the brand. You want users to end up where they can buy the shoe, not just read a review of it.

The Importance Of Synergy And Context With Content

The next thing is the importance of synergy and context with your content. In order to be successful with LLMs, it’s not about (and I’ve heard this before from people) that the top citations are just the ones that already do well on Google. Not necessarily.

There might be a correlation, but not causation. LLMs are trying to do something different than search engines. They’re trying to synthesize the web to serve as a proxy for the entire web. So, what happens with your content across the web matters way more: How your content is talked about, where it’s talked about, who’s talking about it, and how often it’s mentioned.

That doesn’t mean what’s on your site doesn’t factor in, but it’s weighted differently than with traditional search engines. You need to give the LLM the brand context to realize that you have a digital presence in this area, that you’re someone worth mentioning or citing.

Again, I’d focus more on mentions. That’s not to say citations aren’t important (they are), but mentions tend to carry more weight in this context.

Conducting An Audit

The way to go about this, in my opinion, is to conduct an audit. You need to see how the LLM is talking about the topic.

LLMs are notoriously positive and tend to loop in tiny bits of negative sentiment within otherwise positive answers. I was looking at a recent dataset. I don’t have the formal numbers, but I can tell you they’re built to lean neutral or net positive.

For example, if I ask, “Are the Dodgers good?” LLMs, in this case, I was looking at AI Mode, which will say, “Yes, the Dodgers are good…” and go on about that. If I ask, “Are the Yankees good?” and let’s say two or three weeks ago they weren’t doing well, it won’t say, “Yes, the Yankees are good.” It’ll say, “Well, if you look at this and you look at that, overall you might say the Yankees are good.”

Those are two very different answers. They’re both trying to be positive, but you have to read between the lines to understand how the LLM is actually perceiving the brand and what possible user hesitancies or skepticism are bound up in that. Or where are the gaps?

For instance, if I ask, “Is Gatorade a great drink?” and it answers one way, and then I ask, “Is Powerade a good drink?” and it answers slightly differently, you have to notice why that’s happening. Why does it say, “Gatorade is great,” but “Powerade is loved by many”? You have to dig in and understand the difference.

Running an audit helps you see how the LLM is treating your brand and your market. Is it consistently bringing up the same user points of skepticism or hesitation? If I ask, “What’s a good alternative to Folgers coffee?” AI Mode might say, “If you’re looking for a low-cost coffee, Folgers is an option. But if you want something that tastes better at a similar price, consider Brand X.”

That tells you something: There’s a negative sentiment around Folgers and its taste. That’s something you should be picking up on for your content and brand strategy. The only way to know that is to conduct an audit, read between the lines, and understand what the LLM is saying.

Shaping What LLMs Say About Your Brand

The way to get LLMs to say what you want about your brand is to start with a conscious point of view: What do you want LLMs to say about your brand? Which really comes down to: what do you want people to say about your brand?

And the only way to do that is to have a very strong, focused, and conscious brand identity. Who are you? What are you trying to do? Why is that meaningful? Who are you doing it for? And who is interested in you because of it?

Your brand identity is what gives your brand focus. It gives your content marketing focus, your SEO strategy focus, your audience targeting focus, and your everything focus.

If this is who you are, and that is not who you are, then you’re not going to write content that’s misaligned with who you are and what you’re trying to do. You’re not going to dilute your brand identity by creating content that’s tangential or inconsistent.

If you want third-party sites and people around the web to pick up who you are and what you’re about, to build that presence, you need a very conscious and meaningful understanding of who you are and what you do.

That way, you know where to focus, where not to, what content to create, what not to, and how to reinforce the idea around the web that you are X and relevant for X.

It sounds simple, but developing all of that, making sure it’s aligned, and auditing all the way through to ensure it’s actually happening … that’s easier said than done.

Final Thoughts

LLMs may shift how your customers find information about your brands, but chasing citations and clicks isn’t a solid strategy.

Despite the chaos in AI and search in the age of LLMs, marketers need to stick to the fundamentals: brand identity, trust, and relevance still matter.

Focus on brand identity to build your reputation, ensuring that both users and search engines recognize your brand as an authority in your niche.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal

Ask An SEO: How To Manage Stakeholders When An Algorithm Update Hits via @sejournal, @HelenPollitt1

In this edition of Ask An SEO, we address a familiar challenge for marketers:

How do you keep stakeholders from abandoning SEO when algorithm updates cause traffic drops?

This is an all-too-common issue that SEOs will encounter. They have strong plans in place, the buy-in from their leadership, and are making great strides in their organic performance.

When disaster strikes – or, more specifically, a Google algorithm update – all of that goodwill and great results are lost overnight.

What’s worse is, rather than doubling down and trying to recoup lost visibility through data-led SEO work, leadership starts questioning if there is a faster way.

Check The Cause Of The Decline In Traffic

First of all, I would say the most critical step to take when you see a drastic traffic drop is to check that it is definitely the result of an algorithm update.

It’s very easy to ascribe the blame to an update, when it could be caused by a myriad of things. The timing might be suspicious, but before anything, you need to rule out other causes.

Is It Definitely The Result Of The Algorithm Update?

This means checking if there have been any development rollouts, SEO fixes set live, or changes in the SERPs themselves recently. Make sure that the traffic loss is genuine, and not a missing Google Analytics 4 tag. Check that you aren’t seeing the same seasonal dip that you saw this time last year.

Essentially, you need to run down every other possible cause before concluding that it is definitely the result of the algorithm update.

This is important. If it’s not the algorithm update, the loss could be reversible.

Identify Exactly What Has Been Impacted

You are unlikely to have seen rankings and traffic decimated across your entire site. Instead, there are probably certain pages, or topics that you have seen a decline in.

Begin your investigation with an in-depth look into which areas of your site have been impacted.

Look at the webpages that were favored in place of yours. Have they got substantially different content? Are they more topically aligned to the searcher’s intent than yours? Or has the entire SERP changed to favor a different type of SERP feature, or content type?

Why Are These Specific Pages Affected?

What is the commonality between the pages on your site that have seen the rankings and traffic drops? Look for similarities in the templates used, or the technical features of the pages. Investigate if they are all suffering from slow-loading or poor-quality content. If you can spot the common thread between the affected pages, it will help you to identify what needs to be done to recover their rankings.

Is The Impact As Disastrous As It First Appears?

Also, ask yourself if the affected pages are actually important to your business. The impulse might be to remedy what’s gone wrong with them to recover their rankings, but is that the best use of your time? Sometimes, we jump to trying to fix the impact of an algorithm update when, actually, the work would be better spent further improving the pages that are still performing well, because they are the ones that actually make money. If the pages that have lost rankings and traffic were not high-converting ones in the first place, stop and assess. Are the issues they have symptomatic of a wider problem that might affect your revenue-driving pages? If not, maybe don’t worry too much about their visibility loss.

This is good context to have when speaking to your stakeholders about the algorithm impact. Yes, you may have seen traffic go down, but that doesn’t necessarily mean you will see a revenue loss alongside it.

Educate Stakeholders On The Fluctuations In SEO

SEO success is rarely linear. We’ve all seen the fluctuations on the Google Search Console graphs. Do your stakeholders know that, too?

Take time to educate them on how algorithm updates, seasonality, and changing user behavior can affect SEO traffic. Remind them that traffic is not the end goal of SEO; conversions are. Explain to them how algorithm updates are not the end of the world, and just mean there is room for further improvement.

The Best Time To Talk About Algorithm Updates

Of course, this is a lot easier to do before the algorithm update decimates your traffic.

Before you get to the point where panic is ensuing, make sure you have a good process in place to identify the impact of an algorithm update and explain it to your stakeholders. This means that you will take a methodical approach to diagnosing the issues, and not a reactive one.

Let your stakeholders know a reasonable timeframe for that analysis, and that they can’t expect answers on day one of the update announcement. Remind them that the algorithm updates are not stable as they first begin to roll out. They can cause temporary fluctuations that may resolve. You need time and space to consider the cause and remedies of any suspected algorithm update generated traffic loss.

If you have seen this type of impact before, it would be prudent to show your stakeholders where recovery has happened and how. Help them to see that now is the time for further SEO investment, not less.

Reframe The Conversation Back To Long-Term Strategy

There is a very understandable tendency for SEOs to panic in the wake of an algorithm update and try to make quick changes to revert the traffic loss. This isn’t a good idea.

Instead, you need to look at your overarching SEO strategy and locate changes that might have a positive impact over time. For example, if you know that you have a problem with low-quality and duplicate content on your site that you had intended to fix through your SEO strategy, don’t abandon that plan now. Chances are, working to improve the quality of your content on the site will help with regaining that lost traffic.

Resist The Urge To Make Impulsive Changes And Instead Be Methodical About Your Recovery Plans

Don’t throw away your existing plans. You may need to modify them to address specific areas of the site that have been impacted negatively by the update. Carry out intensive investigations into exactly what has happened and to which keywords/topics/pages on your site. Using this information, you can refine your existing strategy.

Any work that is carried out without much thought to the long-term impacts will be unlikely to stand the test of time. You may see a temporary boost, which will placate your stakeholders for a period, but that traffic growth may only be short-lived. For example, buying links to point to the areas of the site most negatively affected by the algorithm update might give you the boost in authority needed to see rankings recover. Over time, though, they are unlikely to carry the same weight, and at worst, may see you further penalized in future algorithm updates or through manual actions.

In Summary

The best time to talk to your stakeholders about the steps to resolve a negative impact from an algorithm update is before it happens. Don’t wait until disaster strikes before communicating your investigation and recovery plans. Instead, let them know ahead of time what to expect and why it isn’t worth a panicked and reactive response.

If you do find your site on the receiving end of a ferocious algorithm update, then take a deep breath. Let your analytical head prevail. Spend time assessing the breadth and depth of the damage, and formulate a plan that yields dividends for the long-term and not just to placate a worried leadership team.

SEO is about the long game. Don’t let your stakeholders lose their nerve just because an algorithm update has happened.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal

Ask An SEO: Is It Better To Refresh Content Or Create New Pages? via @sejournal, @rollerblader

This week’s Ask An SEO asks a classic content conundrum:

“Are content refreshes still an effective tactic, or is it better to create new pages altogether?”

Yes, content refreshes are still an effective tactic in cases such as:

  • Product releases where you only continue to sell the new product (new colors or sizes and other variants, but the same product).
  • Data is released and should be updated for the content to be helpful or accurate.
  • New customer or reader questions that are something readers are considering and thinking about.
  • New brands enter the space and others close down, making shopping lists non-helpful if there’s nowhere to shop.
  • New ways to present the content, such as adding bullet lists or tables, or a new video.

With that said, not every page needs to be refreshed. If there is a similar topic that will help the reader but isn’t directly related to an existing header or sub-header, refreshing the page to include the new content could take your page off-topic. This can make it somewhat irrelevant or less helpful for users, which makes it bad for SEO, too. In this case, you’ll want to create a new page.

Once you have the new page created, look for where it can tie into the page you initially wanted to refresh and add an internal link to the new page. This gives the visitor on the page the opportunity to learn more or find the alternative, and then click back to finish reading or shopping. It also helps search engines and crawlers find their way to the new content.

New pages could be a good solution for:

  • Articles and guides where you want to define a topic, strategy, or theory in more detail.
  • Ecommerce experience to bring users to a sub-collection or sub-category, or a product alternative for things that are better for specific needs like size, fit, make, or model, etc.
  • Lead gen pages where you have a few service options and want the person to find the more relevant funnel for their specific needs.

For example, a recipe site that offers a regular, gluten-free, and vegetarian option doesn’t need to stuff all three recipe versions into the main recipe page. They can use an internal link at the top of the main recipe that says, “Click here for the gluten free version,” which helps the user and lets the search engines know they have this solution, too. Clothing brands can talk about tighter or looser fits and recommend a complementary brand if a customer complains about the same thing for a specific product or brand; this can go on product or category and collection pages.

If a client asks if they should refresh or create a new page, we:

  • Recommend refreshing pages when the content begins to slip, does not recover, and we realize that the content is no longer as helpful as it could be. If refreshing the content can keep it on topic and provide a more accurate solution, or a better way for visitors to absorb it.
  • Add new pages when the solution a visitor needs is relevant to the page that we thought about refreshing, but is unique enough from the main topic to justify having its own page. SEO pages aren’t about the keywords; they are about the solution the page provides and how you can uncomplicate it.

Complicated pages are ones with:

  • Tons of jargon that regular consumers won’t understand without doing another search.
  • Multiple sections where the content is hard to scan through and has solutions that are difficult to find.
  • Large bulky paragraphs and no visual breaks, or short choppy paragraphs that don’t have actual solutions, just general statements.
  • Sentences that should instead be lists, headers, tables, and formatted in easier-to-absorb formats.

But knowing what you could do or try doing doesn’t mean anything if you aren’t measuring the results.

How To Measure The Effectiveness

Depending on which one you choose, you’ll have different ways to measure the effectiveness. Here are a few tests we do with clients in these same situations:

The first option is to have a control group with a couple of pages or topics, and we leave them alone as a control group. We then either expand with an equal amount of new content or refresh the same amount. The control group should be about as competitive to rank as the test groups, and from there, we watch over a few months to see if the test group begins climbing or gaining traffic while the control group remains the same.

The second test you can run, assuming you have a reasonably reliable rank tracking tool, is to monitor how many new keywords the content group has in the top 100 positions, top 20 positions, and top 10 positions after a couple of months. If the keywords and phrases have the same user intent as the topic (i.e., shopping vs. how to do something vs. informative and educational), then it looks like you made a good decision. On top of this, look for rich results like increases in People Also Ask and AI overview appearances. This is a sign the new content may be high quality and that you made the right decision.

Summary

I hope this helps answer your question. Refresh when the content is outdated, could be formatted better, or because it is fluffy and doesn’t provide value. Add new pages when there is a solution for a problem or an answer for a question, and it is unique enough from an existing page to justify the page’s existence. SEO keywords and search volumes do not justify this; an actual unique solution does.

More Resources:


Featured Image: Paulo Bobita/Search Engine Journal