SEOs Are Recommending Structured Data For AI Search… Why? via @sejournal, @martinibuster

A post on LinkedIn questioned the idea that Schema.org structured data has an impact on what a large language model outputs. Apparently there are some SEOs who are recommending structured data to rank better in AI search engines.

Patrick Stox wrote the following post on LinkedIn:

“Did I miss something? Why do SEOs think schema markup will impact LLM output?”

Patrick said “LLM output” in the context of an SEO recommendation so it’s likely that it’s a reference to ChatGPT Search and other AI search engines. So do AI search engines get their data from structured data?

LLMs are trained on web text, books, government records, legal documents and other text data (as well as other forms of media, too) which is then used to produce summaries and answers but without plagiarizing the training data.  What that means is that it’s pointless to think that optimizing your web content will result in the LLM itself sending referrals to that website.

AI search engines are grounded on search indexes (and knowledge graphs) through Retrieval Augmented Generation (RAG). Search engine indexes themselves are created from crawled data, not Schema structured data.

Perplexity AI ranks web-crawled content using a modified version of PageRank on their search index, for example. Google and Bing crawl text data and do things like remove duplicate content, remove stop words, and other manipulation of the text extracted from the HTML, plus not every page has structured data on it.

In fact, Google only uses a fraction of the available Schema.org structured data for specific kinds of search experiences and rich results, which in turn limits the kind of structured data that publishers use.

Then there’s the fact that both Bing and Google’s crawlers render the HTML, identify the headers, footers and main content (from which they extract the text for ranking purposes). Why would they do that if they’re going to rely on Schema structured data, right?

The idea that it’s good to use Schema.org structured data to rank better in an AI search engine is not based on facts, it’s just fanciful speculation. Or it could be from a “game of telephone” effect where one person says something and then twenty people later it’s transformed into something completely different.

For example, Jono Alderson proposed that structured data could be a standard that AI search engines could use to understand the web better. He wasn’t saying that AI search engines currently use it, he was just proposing that AI search engines should consider adopting it and maybe that post got telephoned into a full-blown theory twenty SEOs later.

Unfortunately, there’s a lot of unfounded ideas floating around in SEO circles. The other day I saw an SEO assert in social media that Google Local Search doesn’t use IP addresses in response to search “near me” search queries. All anyone had to do to test that idea is to sign into a VPN, choose a geographic location for their IP address and do a “near me” search query and they will see that the IP address used by the VPN influenced the “near me” search results.

Screenshot Of Near Me Query Influenced By IP Address

Google even publishes a support page that says they use IP address to personalize search results yet there are people who believe otherwise because some SEO did a correlation study and when questioned we’re back to someone bellowing that Google lies.

Will You Believe Your Lying Eyes?

Schema.Org Structured Data And AI Search Results

“SEOs” recommending that publishers use Schema.org structured data for LLM training data also makes no sense because training data isn’t cited in LLM output, just for output that is sourced from the web, which itself is sourced from a search index that’s from a crawler. As mentioned earlier, publishers only use a fraction of available Schema.org structured data because Google itself only uses a tiny fraction of it. So it makes no sense for an AI search engine to rely on structured data for their output.

Search marketing expert Christopher Shin (LinkedIn profile) commented:

“Thinking the same thing after reading your post Patrick. This is how I interpret it currently. I thought LLM’s typically do not generate responses from search engines serps but rather from data interpretation. Right? But schema data markup would be used by SER{s to show rich snippets etc. no? I think the key nuance with schema and LLMs is that search engines use schema for SERPs whereas LLM’s use data interpretation when it comes to how schema impacts LLM’s.”

People like Christopher Shin and Patrick Stox give me hope that pragmatic and sensible SEO is still fighting to get through the noise, Patrick’s LinkedIn post is proof of that.

Pragmatic SEO

The definition of pragmatic is doing things for sensible and realistic reasons and not on opinions that are based on incomplete information and conjecture.

Speaking as someone who’s been involved with SEO since virtually the birth of it, not thinking things through is why SEOs and publishers have traditionally wasted time with vaguely defined issues, spun their wheels on useless activities like superficial signals of EEAT and so on and so forth.  It’s truly dispiriting to point to documentation and official statements and get blown back with statements like, “Google lies.” That kind of attitude makes a person “want to holler.”

A little more pragmatic SEO please.

What Happens Next To The U.S. Vs. Google Antitrust Case? via @sejournal, @AlliBerry3

With the punishment for Google’s first search antitrust case expected to be delivered in August 2025, the looming question is what will happen now with a new U.S. President and a new set of Department of Justice (DOJ) appointees.

Early signs suggest the Trump administration will largely stay the course of the Biden administration when it comes to antitrust enforcement against large tech companies, including Google.

Their rationale is drastically different from that of the previous administration, but the recent nominations and appointments for the DOJ suggest that President Trump is serious about holding Google accountable, even if their preferred remedies may differ.

Before we get into it, let’s recap what has happened so far.

The U.S. Vs. Google Case

In August 2024, Federal Judge Amit Mehta ruled that Google violated the U.S. antitrust law by maintaining an illegal monopoly through exclusive agreements it had with companies like Apple to be the world’s default search engine on smartphones and web browsers.

Additionally, Google was found guilty of monopolizing general search text advertising because Google was able to raise prices on search advertising products higher than what the government claimed would have been expected or possible in a fair market.

Potential Remedies For Google

The DOJ submitted two filings with their suggestions to remediate Google’s monopolistic actions.

Proposed remedies range from restrictions on deals that feature Google’s search engine as the default on browsers and devices all the way to a breakup of the company by forcing the sale of Google’s browser Chrome.

Other intriguing remedies that have been proposed include syndicating the Google search algorithm to competitors, forced licensing of ad feeds to competitors, and divesting the Android operating system.

The DOJ under Biden made it clear in their most recent filing on November 20, 2024, that divesting Chrome is their preferred option, along with the discontinuation of exclusive agreements with browsers and phone companies.

The implications of divesting Chrome are also the most wide-reaching – not only is Chrome used by nearly two-thirds of the world’s internet users, but we learned through this trial that click data from Chrome is used to train the search algorithms using Navboost, helping Google maintain its competitive edge.

Losing Chrome’s data would almost certainly guarantee a drastically different Google search engine.

Google filed its response to the DOJ, arguing that the proposed remedies are much wider reaching than what the case was about and that America’s global leadership position in tech could be hindered by this.

Instead, they proposed allowing exclusive agreements to be made with companies like Apple and Mozilla, but with the ability to set a different default search engine on different platforms and browsing modes.

It also proposed that Android device manufacturers could preload multiple search engines, as well as preload Google apps without Google Search or Chrome.

Both sides will return to court for the remedies litigation in May 2025, with a ruling expected to be delivered in August 2025.

What Happens Now

Back to the question at hand: What happens once Trump takes office?

The initial signals, including Trump’s nominations for key roles at the FTC and the Department of Justice Antitrust Division, suggest the administration will continue to use a heavy hand against large tech companies facing antitrust troubles like Google. But, their solutions may differ from the current proposed remedies.

Trump’s Relevant Nominees

Trump has nominated several key individuals who will influence antitrust enforcement, particularly concerning Big Tech companies.

These appointments indicate that the crackdown on tech giants will likely continue, in effect, a surprising bipartisan effort. Trump’s key nominees include:

  • Gail Slater: Nominated to lead the Department of Justice’s Antitrust Division, Slater has a background as a policy advisor to Vice President-elect J.D. Vance and experience in tech policy at the National Economic Council. If confirmed, she would inherit the antitrust case against Google.
  • Andrew N. Ferguson: Appointed as Chair of the Federal Trade Commission (FTC), Ferguson has expressed intentions to reassess the agency’s approach to mergers and acquisitions, which has been uncommonly strong against mergers and acquisitions, while still maintaining oversight of dominant tech platforms.
  • Mark Meador: Appointed as an FTC Commissioner, a role previously held by Ferguson, Meador is recognized for his pro-enforcement stance, especially regarding technology companies, in his previous work with the U.S. Senate Judiciary Committee. His previous work includes drafting legislation aimed at addressing competitive practices in the tech industry.

While all three of these nominees are deeply rooted in the Republican party, they are all united in their pro-enforcement stances when it comes to Big Tech.

This is a departure from the typical Republican pro-business, anti-regulation position, signifying Trump’s seriousness in curbing the power of Google and other tech giants.

The Trump Administration’s Views On Google’s Antitrust Case

Trump’s disdain for Big Tech companies, including Google, has been consistent since his first presidency.

Why does he hate Google so much? A couple of reasons seem most likely:

  1. He has claimed the search engine is “rigged” because it presents negative stories about him.
  2. He sees weakening Big Tech companies as a way to promote “free speech” because of their misinformation moderation policies and claims the search results are biased against conservatives.

Despite this seemingly constant position against Google, President Trump has also suggested that breaking Google up may destroy the company rather than help promote fairness and competition.

He has also warned that breaking up Google may make the U.S. appear weaker to foreign powers because “China is afraid of Google.”

Elsewhere in the administration, Vice President Vance has previously called for the breakup of Google and praised the Biden administration’s Federal Trade Commission Chair, Lina Khan, for her aggressive approach to antitrust enforcement.

Whether they decide to take a stance that is pro-breaking Google up remains to be seen, but it appears that they will be taking office with a desire to strengthen competition in this market.

Final Thoughts

There is a lot of time between Trump taking office and the remedies litigation starting up again for the case against Google in May 2025.

The DOJ still needs to argue why they believe Google should be forced to sell Chrome, and if this is no longer the belief of the DOJ appointees, they will need to argue why other remedies make more sense.

It seems reasonable to assume, based on the appointees, that they will be taking some big swings at Google and arguing for the remedies that they believe would be most effective at enhancing competition.

If you are someone who believes action needs to be taken against Google, Trump’s current anti-Google stance may work in your favor regardless of whether you agree with his rationale for it.

More Resources:


Featured Image: PanuShot/Shutterstock

Are People Clicking Links In ChatGPT Search? Brands Say Yes via @sejournal, @MattGSouthern

A report from Modern Retail shows that people who use ChatGPT and Google Gemini for quick summaries also click the links these tools provide.

This is important for marketers, as it suggests that AI-driven search may change product discovery and online traffic.

While these numbers are self-reported and lack broader data, they offer insight into how consumers engage with AI search results and how brands can benefit.

What Brands Are Observing

Viv, a period care brand, noticed a trend last summer when its website traffic increased by 400%. Marketing director Kelly Donohue linked this to the rise of AI tools.

This spike coincided with a study in Environment International that found harmful heavy metals in popular tampon brands. Viv’s blog posts about product safety gained visibility as people searched for safer options.

The increased traffic resulted in more sales, with Viv reporting a 436% rise from these AI-driven referrals. This indicates that users actively clicked through to learn more and make purchases.

What To Learn From This

Viv’s experience highlights the need for brands to create comprehensive content that answers people’s questions.

Donohue pointed out that platforms like ChatGPT prefer articles with context, sources, and thorough explanations over keyword-heavy material.

Donohue explained,

“These AI tools are specifically scraping through content, but looking for more than just keywords. They’re looking for a cohesive response that they can give to people that includes context, sources, and background.”

In response, Viv focused on transparency and product safety. By creating educational articles, Viv built consumer trust and improved its visibility in AI recommendations.

The effort paid off, Donohue added:

“We ended up selling out of about six months of tampon inventory in three weeks, driven by Google’s AI-powered recommendations.”

Other Brands Report Similar Trends

Joe & Bella is an adaptive apparel brand that has gained more visitors from ChatGPT recommendations.

It makes clothing for older adults and people with mobility challenges, and during the holiday season, it saw an increase in visitors and purchases.

Jimmy Zollo, Joe & Bella’s co-founder and CEO, tells Modern Retail:

“I don’t really know how or what they would have typed or asked ChatGPT to have found us over the holidays.”

Zollo speculated that the company’s ongoing investment in SEO and its blog content likely played a role.

The brand consistently uses keywords like “adaptive clothing” in its search ads and blog posts, which may have helped position it in AI-driven results.

Zollo added:

“It was pretty cool and unexpected, but we need to better understand how to optimize for these searches going forward.”

What This Means for Marketers

These reports show that people engage with links in AI-generated search results rather than just reading summaries.

Dan Buckstaff, chief product officer at Spins, compares this to the early days of SEO.

Buckstaff said:

“Similar to 15 years ago when we were questioning how SEO worked, we’re left with questioning how brands can benefit from AI environments.”

Spins’ 2025 Industry Trends Report indicates that consumers are increasingly using AI tools like ChatGPT and social media platforms like TikTok to discover products.

While advertising on these AI tools is still developing, brands with strong, organized content are benefiting.

Looking Ahead

Consumers are increasingly clicking on links in AI-driven search results, especially younger audiences like Gen Z, who use AI tools for product discovery.

For brands like Viv, this change is crucial for content creation.

Donohue said:

“These searches are top of mind for us now, and the way we’re writing our blogs and the content on our website can play a huge part in people finding us through AI tools.”

The key takeaway is to focus on straightforward, educational content to improve your chances of being recommended by AI-powered search tools.


Featured Image: Mojahid Mottakin/Shutterstock

Google Updates Site Reputation Abuse Policy Documentation via @sejournal, @MattGSouthern

Google has updated its documentation to provide clearer guidance on its site reputation abuse policy.

The changes are meant to you better understand what qualifies as a violation and how to stay compliant.

While the updates don’t change how the policy is applied, they make the rules easier to follow by incorporating more detailed explanations from a recent blog post FAQ.

What Changed?

The updated documentation now includes content directly pulled from Google’s November blog post about site reputation abuse.

That blog post introduced a Q&A section to clarify the policy. Google has now added this FAQ guidance to its official spam policies documentation.

In a statement, Google explained:

“We updated the site reputation abuse policy to include guidance from our blog post’s FAQ on site reputation abuse. These are editorial changes only, no change in behavior.”

This means the policy hasn’t changed—it’s just been rewritten to make it easier to understand.

What Is Site Reputation Abuse?

Site reputation abuse happens when third-party content is published on a well-established website to take advantage of that site’s ranking signals.

Essentially, it occurs when someone uses a reputable site as a shortcut to boost rankings for unrelated or low-quality content rather than earning those rankings independently.

For example:

  • A news site hosting coupon pages from a third-party service purely to benefit from the site’s strong rankings in Google.
  • An educational site publishing sponsored reviews about payday loans.
  • A movie review site hosting unrelated pages about essay writing services or buying social media followers.

However, not all third-party content is considered abuse. Forums, user-generated content, syndicated news articles, and editorial pieces are generally acceptable if they’re not designed to manipulate search rankings.

Why Does This Matter?

These updates make it easier to determine whether your content violates the policy.

For example, Google’s FAQ now clarifies common scenarios, such as:

  • Third-party content: Simply having third-party content isn’t a violation unless explicitly published to exploit a site’s rankings.
  • Freelance and affiliate content: Freelance content or affiliate pages are acceptable if they’re not used to manipulate rankings. Affiliate links, when tagged appropriately (e.g., with “nofollow” or “sponsored” attributes), don’t violate the rules.

The FAQ also explains how to address violations. You can fix the issue by removing or relocating problematic content, submitting reconsideration requests in Search Console, and following Google’s spam guidelines.

This is a good reminder to review your content practices to ensure they align with Google’s policies. If you host third-party content, make sure it adds value for users and doesn’t just serve to piggyback off your site’s reputation.


Featured Image: RYO Alexandre/Shutterstock

Google’s Documentation Update Contains Hidden SEO Insights via @sejournal, @martinibuster

Google quietly updated their Estimated Salary (Occupation) Structured Data page with subtle edits that make the information more relevant and easily understood. The changes show how a page can be analyzed for weaknesses and subsequently improved.

Subtle Word Shifts Make A Difference

The art of writing is something SEO should consider now more than ever. It’s been important for at least the past six years but in my opinion it’s never been more important than it is today because of the preciseness of natural language queries for AI Overviews and AI assistants.

Three Takeaways About Content

  1. The words used on a page can exert a subtle influence in how a reader and a machine understand the page.
  2. Relevance is commonly understood as whether a web page is a match for a user’s search query and the user’s intent, which is an outdated way to think about it, in my opinion.
  3. A query is just a question and the answer is never a web page. The answer is generally a passage in a web page.

Google’s update to their “Estimated Salary (Occupation) Structured Data” web page offers a view of how Google updated one of their own web pages to be more precise.

There were only two changes that were so seemingly minimal they didn’t even merit a mention on their documentation changelog, they just updated it and pushed it live without any notice.

But the changes do make a difference in how precise the page is on the topic.

First Change: Focus Of Content

Google refers to “enriched search results” as different search experiences, like the recipe search experience, event search experience and the job experience.

The original version of the “Estimated Salary (Occupation) Structured Data” documentation focused on talking about the Job Experience search results. The updated version completely removed all references to the Job Experience and is now more precisely focused on the “estimated salary rich result” which is more precise than the less precise “Job Experience” phrasing.

This is the original version:

“Estimated salaries can appear in the job experience on Google Search and as a salary estimate rich result for a given occupation.”

This is the updated version:

“Adding Occupation structured data makes your content eligible to appear in the estimated salary rich result in Google Search results:”

Second Change: Refreshed Image And Simplified

The second change refreshes an example image.

The change has three notable qualities:

  1. Precisely models a search result
  2. Aligns with removal of “job experience”
  3. Simplifies message

The original image contained a screenshot of a laptop with a search result and a closeup of the search result overlaid. The image looks more at home on a product page than an informational page. Someone spent a lot of time creating an attractive image but it’s too complex and neglects the number one rule of content which is that all content must communicate the message quickly.

All content, whether text or image, is like a glass of water: the important part is the water, not the glass.

Screenshot Of Attractive But Slightly Less Effective Image

The image that replaced it is literally an example of the actual rich result. It’s not fancy but it doesn’t have to be. It just has to do the job of communicating.

Screenshot Of Google’s More Effective Image

The other thing this change accomplishes is that it removes the phrase “job experience” and replaces it with a sentence that aligns with the apparent goal of making this page about the Occupation structured data.

This is the new text:

“Adding Occupation structured data makes your content eligible to appear in the estimated salary rich result in Google Search results:”

Third change: Replace Confusing Sentence

The third change corrected a sentence that was grammatically incorrect and confusing.

Original version:

“You must include the required properties for your content to be eligible for display the job experience on Google and rich results.”

Google corrected the grammar error, made the sentence specific to the ‘estimated salary’ rich result, and removed the reference to Job Experience, aligning it more strongly with estimated salary rich results.

This is the updated version:

“You must include the required properties for your content to be eligible for display in the estimated salary rich result.”

Three Examples For Updating Web Pages

On one level the changes were literally about removing the focus on one topic and reinforcing a slightly different one. On another level it’s an example of giving users a better experience by communicating more precisely. Writing for humans is not just a creative art, it’s also a technical one. All writers, even novelists, understand that the craft of writing is technical because one of the most important factors is communicating ideas. Other issues like being comprehensive or fancy don’t matter as much as the communication part.

I think that the revisions Google made fits into what Google means when it says to make content for humans not search engines.

Read the updated documentation here:

Estimated salary (Occupation) structured data

Compare it to the archived original version.

Featured Image by Shutterstock/Lets Design Studio

AI Search Optimization: Make Your Structured Data Accessible via @sejournal, @MattGSouthern

A recent investigation has uncovered a problem for websites relying on JavaScript for structured data.

This data, often in JSON-LD format, is difficult for AI crawlers to access if not in the initial HTML response.

Crawlers like GPTBot (used by ChatGPT), ClaudeBot, and PerplexityBot can’t execute JavaScript and miss any structured data added later.

This creates challenges for websites using tools like Google Tag Manager (GTM) to insert JSON-LD on the client side, as many AI crawlers can’t read dynamically generated content.

Key Findings About JSON-LD & AI Crawlers

Elie Berreby, the founder of SEM King, examined what happens when JSON-LD is added using Google Tag Manager (GTM) without server-side rendering (SSR).

He found out why this type of structured data is often not seen by AI crawlers:

  1. Initial HTML Load: When a crawler requests a webpage, the server returns the first HTML version. If structured data is added with JavaScript, it won’t be in this initial response.
  2. Client-Side JavaScript Execution: JavaScript runs in the browser and changes the Document Object Model (DOM) for users. At this stage, GTM can add JSON-LD to the DOM.
  3. Crawlers Without JavaScript Rendering: AI crawlers that can’t run JavaScript cannot see changes in the DOM. This means they miss any JSON-LD added after the page loads.

In summary, structured data added only through client-side JavaScript is invisible to most AI crawlers.

Why Traditional Search Engines Are Different

Traditional search crawlers like Googlebot can read JavaScript and process changes made to a webpage after it loads, including JSON-LD data injected by Google Tag Manager (GTM).

In contrast, many AI crawlers can’t read JavaScript and only see the raw HTML from the server. As a result, they miss dynamically added content, like JSON-LD.

Google’s Warning on Overusing JavaScript

This challenge ties into a broader warning from Google about the overuse of JavaScript.

In a recent podcast, Google’s Search Relations team discussed the growing reliance on JavaScript. While it enables dynamic features, it’s not always ideal for essential SEO elements like structured data.

Martin Splitt, Google’s Search Developer Advocate, explained that websites range from simple pages to complex applications. It’s important to balance JavaScript use with making key content available in the initial HTML.

John Mueller, another Google Search Advocate, agreed, noting that developers often turn to JavaScript when simpler options, like static HTML, would be more effective.

What To Do Instead

Developers and SEO professionals should ensure structured data is accessible to all crawlers to avoid issues with AI search crawlers.

Here are some key strategies:

  1. Server-Side Rendering (SSR): Render pages on the server to include structured data in the initial HTML response.
  2. Static HTML: Use schema markup directly in the HTML to limit reliance on JavaScript.
  3. Prerendering: Offer prerendered pages where JavaScript has already been executed, providing crawlers with fully rendered HTML.

These approaches align with Google’s advice to prioritize HTML-first development and include important content like structured data in the initial server response.

Why This Matters

AI crawlers will only grow in importance, and they play by different rules than traditional search engines.

If your site depends on GTM or other client-side JavaScript for structured data, you’re missing out on opportunities to rank in AI-driven search results.

By shifting to server-side or static solutions, you can future-proof your site and ensure visibility in traditional and AI searches.


Featured Image: nexusby/Shutterstock

TikTok Ban Sparks 5000% Surge In Alternative App Searches via @sejournal, @MattGSouthern

The recent TikTok ban drama in the U.S. has caused a surge in search activity as people look for answers, alternatives, and workarounds.

The app temporarily shut down over the weekend and was restored after President-elect Donald Trump announced a 90-day extension. This led to a notable rise in search interest.

An SEO consultant named Sobhi Smat compiled a collection of search data and shared it on LinkedIn.

Here’s what the data shows about people’s reactions and what it means for marketers.

The Context: TikTok’s Uncertain Future

On January 17, the U.S. Supreme Court upheld PAFACA. The original deadline for compliance was January 19.

In response, on January 18, TikTok began shutting down its services in the U.S., removing the app from app stores and displaying service discontinuation notices.

On January 19, President-elect Donald Trump announced plans for a 90-day extension via executive order, allowing TikTok to restore operations while negotiations continue temporarily.

Search Behavior: Three Key Trends Emerge

Analysis of Google search data from January 1 to January 16 reveals three dominant categories of search behavior related to the TikTok ban:

  1. Staying Informed
  2. Exploring Alternatives
  3. Circumventing the Ban

1. Staying Informed

One of the largest spikes in search activity was caused by people trying to understand the reasons behind the ban and stay informed about recent developments.

Queries like “TikTok ban update,” “Supreme Court ruling on TikTok,” and “Is the TikTok ban extended?” saw a breakout, with search interest increasing by over 5000%.

2. Exploring TikTok Alternatives: A Battle for User Attention

As fears of TikTok’s potential shutdown grew, people turned to Google to explore alternative platforms.

The search term “TikTok alternatives” saw explosive growth, alongside interest in specific apps such as RedNote, Lemon8, Clapper, and Fanbase.

RedNote: The Rising Star

Among alternatives, RedNote attracted the most attention, with breakout search terms like “What is RedNote?”, “Is RedNote safe?”, and “TikTok vs RedNote”.

However, RedNote’s surge in popularity exposed its challenges, particularly in delivering high-quality English-language content and addressing translation issues. This led to a related search spike for “Chinese to English translation.”

Other Notable Alternatives

Other apps like Lemon8, Clapper, and Fanbase also saw increased search interest:

  • Lemon8: Questions included ” What is the Lemon8 app?” and “WWill Lemon8 be banned, too? “
  • Clapper: Searches like “what is Clapper social media” and “is Clapper safe” highlighted curiosity about this lesser-known platform.
  • Fanbase: Users searched for “how to invest in Fanbase” and “Isaac Hayes Fanbase app,” showing interest in the app’s unique monetization features.

3. Circumventing the Ban

Another trend involved users searching for ways to continue accessing TikTok despite the shutdown.

Queries like “Can I use TikTok with VPN?” “How to change location on TikTok?” and “VPN for TikTok?” spiked dramatically.

The interest in VPNs shows TikTok’s user base is determined to bypass restrictions and maintain access to the platform.

Deletion Trends

While people explored TikTok replacements, search trends indicate they were quickly disappointed.

A spike in searches like “how to delete RedNote account” and “delete Lemon8 app” suggests that not all alternatives met user expectations.

Potential Buyers

Search trends also reflect public curiosity about potential U.S. buyers, with queries mentioning various high-profile figures, including Mr. Beast, Elon Musk, and even Dolly Parton.

This aligns with the legislative requirement for ByteDance to sell to a U.S. company or cease operations.

What This Means for Marketers

For digital marketers, current events show that relying on one platform is risky.

Marketers should monitor these developments closely whether TikTok is sold, banned, or granted an extension.

This situation is a reminder of how legislative actions can influence online behavior and disrupt the market.


Featured Image: RKY Photo/Shutterstock

FTC: GoDaddy Hosting Was “Blind” To Security Threats via @sejournal, @martinibuster

The United States Federal Trade Commission (FTC) charged GoDaddy with violations of the Federal Trade Commission Act for allegedly maintaining “unreasonable” security practices that led to multiple security breaches. The FTC’s proposed settlement order will require GoDaddy to take reasonable steps to tighten security and engage third-party security assessments.

FTC Charged GoDaddy With Security Failures

The FTC complaint charged GoDaddy with misrepresenting itself as a secure web host through marketing on its website, in emails and it’s “Trust Center”, alleging that GoDaddy provided customers with “lax data security” in its web hosting environment.

The FTC complaint (PDF) stated:

“Since at least 2015, GoDaddy has marketed itself as a secure choice for customers to host their websites, touting its commitment to data security and careful threat monitoring practices in multiple locations, including its main website for hosting services, its “Trust Center,” and in email and online marketing.

In fact, GoDaddy’s data security program was unreasonable for a company of its size and complexity. Despite its representations, GoDaddy was blind to vulnerabilities and threats in its hosting environment. Since 2018, GoDaddy has violated Section 5 of the FTC Act by failing to implement standard security tools and practices to protect the environment where it hosts customers’ websites and data, and to monitor it for security threats.”

Proposed Settlement

The FTC is proposing that GoDaddy implement a security program to settle charges that it failed to secure its web hosting services, endangering their customers and the people who visited their customer’s compromised websites during major security breaches between 2019 and 2022.

The settlement proposes the following to settle the charges with GoDaddy:

“Prohibit GoDaddy from making misrepresentations about its security and the extent to which it complies with any privacy or security program sponsored by a government, self-regulatory, or standard-setting organization, including the EU-U.S. and Swiss-U.S. Privacy Shield Frameworks;

Require GoDaddy to establish and implement a comprehensive information-security program that protects the security, confidentiality, and integrity of its website-hosting services; and

Mandate that GoDaddy hire an independent third-party assessor who conducts an initial and biennial review of its information-security program.”

Read the FTC statement:

FTC Takes Action Against GoDaddy for Alleged Lax Data Security for Its Website Hosting Services

Featured Image by Shutterstock/Photo For Everything

OpenAI Secretly Funded Benchmarking Dataset Linked To o3 Model via @sejournal, @martinibuster

Revelations that OpenAI secretly funded and had access to the FrontierMath benchmarking dataset are raising concerns about whether it was used to train its reasoning o3 AI reasoning model, and the validity of the model’s high scores.

In addition to accessing the benchmarking dataset, OpenAI funded its creation, a fact that was withheld from the mathematicians who contributed to developing FrontierMath. Epoch AI belatedly disclosed OpenAI’s funding only in the final paper published on Arxiv.org, which announced the benchmark. Earlier versions of the paper omitted any mention of OpenAI’s involvement.

Screenshot Of FrontierMath Paper

Closeup Of Acknowledgement

Previous Version Of Paper That Lacked Acknowledgement

OpenAI 03 Model Scored Highly On FrontierMath Benchmark

The news of OpenAI’s secret involvement are raising questions about the high scores achieved by  the o3 reasoning AI model and causing disappointment with the FrontierMath project. Epoch AI responded with transparency about what happened and what they’re doing to check if the o3 model was trained with the FrontierMath dataset.

Giving OpenAI access to the dataset was unexpected because the whole point of it is to  test AI models but that can’t be done if the models know the questions and answers beforehand.

A post in the r/singularity subreddit expressed this disappointment and cited a document that claimed that the mathematicians didn’t know about OpenAI’s involvement:

“Frontier Math, the recent cutting-edge math benchmark, is funded by OpenAI. OpenAI allegedly has access to the problems and solutions. This is disappointing because the benchmark was sold to the public as a means to evaluate frontier models, with support from renowned mathematicians. In reality, Epoch AI is building datasets for OpenAI. They never disclosed any ties with OpenAI before.”

The Reddit discussion cited a publication that revealed OpenAI’s deeper involvement:

“The mathematicians creating the problems for FrontierMath were not (actively)[2] communicated to about funding from OpenAI.

…Now Epoch AI or OpenAI don’t say publicly that OpenAI has access to the exercises or answers or solutions. I have heard second-hand that OpenAI does have access to exercises and answers and that they use them for validation.”

Tamay Besiroglu (LinkedIn Profile), associated director at Epoch AI, acknowledged that OpenAI had access to the datasets but also asserted that there was a “holdout” dataset that OpenAI didn’t have access to.

He wrote in the cited document:

“Tamay from Epoch AI here.

We made a mistake in not being more transparent about OpenAI’s involvement. We were restricted from disclosing the partnership until around the time o3 launched, and in hindsight we should have negotiated harder for the ability to be transparent to the benchmark contributors as soon as possible. Our contract specifically prevented us from disclosing information about the funding source and the fact that OpenAI has data access to much but not all of the dataset. We own this error and are committed to doing better in the future.

Regarding training usage: We acknowledge that OpenAI does have access to a large fraction of FrontierMath problems and solutions, with the exception of a unseen-by-OpenAI hold-out set that enables us to independently verify model capabilities. However, we have a verbal agreement that these materials will not be used in model training.

OpenAI has also been fully supportive of our decision to maintain a separate, unseen holdout set—an extra safeguard to prevent overfitting and ensure accurate progress measurement. From day one, FrontierMath was conceived and presented as an evaluation tool, and we believe these arrangements reflect that purpose. “

More Facts About OpenAI & FrontierMath Revealed

Elliot Glazer (LinkedIn profile/Reddit profile), the lead mathematician at Epoch AI confirmed that OpenAI has the dataset and that they were allowed to use it to evaluate OpenAI’s o3 large language model, which is their next state of the art AI that’s referred to as a reasoning AI model. He offered his opinion that the high scores obtained by the o3 model are “legit” and that Epoch AI is conducting an independent evaluation to determine whether or not o3 had access to the FrontierMath dataset for training, which could cast the model’s high scores in a different light.

He wrote:

“Epoch’s lead mathematician here. Yes, OAI funded this and has the dataset, which allowed them to evaluate o3 in-house. We haven’t yet independently verified their 25% claim. To do so, we’re currently developing a hold-out dataset and will be able to test their model without them having any prior exposure to these problems.

My personal opinion is that OAI’s score is legit (i.e., they didn’t train on the dataset), and that they have no incentive to lie about internal benchmarking performances. However, we can’t vouch for them until our independent evaluation is complete.”

Glazer had also shared that Epoch AI was going to test o3 using a “holdout” dataset that OpenAI didn’t have access to, saying:

“We’re going to evaluate o3 with OAI having zero prior exposure to the holdout problems. This will be airtight.”

Another post on Reddit by Glazer described how the “holdout set” was created:

“We’ll describe the process more clearly when the holdout set eval is actually done, but we’re choosing the holdout problems at random from a larger set which will be added to FrontierMath. The production process is otherwise identical to how it’s always been.”

Waiting For Answers

That’s where the drama stands until the Epoch AI evaluation is completed which will indicate whether or not OpenAI had trained their AI reasoning model with the dataset or only used it for benchmarking it.

Featured Image by Shutterstock/Antonello Marangi

Confirmed: Google Is Requiring JavaScript To Block SEO Tools via @sejournal, @martinibuster

Google has made a change to how it’s search results are served which will also help to secure it against bots and scrapers. Whether this will have further effect on SEO Tools or if they can use a headless Chrome that uses JavaScript remains an open question at the moment but it’s likely that Google is using rate limiting to throttle how many pages can be requested within a set period of time.

Google Search Now Requires JavaScript

Google quietly updated their search box to require all users, including bots, to have JavaScript turned on when searching.

Surfing Google Search without JavaScript turned on results in the following message:

Turn on JavaScript to keep searching
The browser you’re using has JavaScript turned off. To continue your search, turn it on.

Screenshot Of Google Search JavaScript Message

In an email to TechCrunch a Google spokesperson shared the following details:

“Enabling JavaScript allows us to better protect our services and users from bots and evolving forms of abuse and spam, …and to provide the most relevant and up-to-date information.”

JavaScript possibly enables personalization in the search experience, which is what that spokesperson may mean by providing the most relevant information. But JavaScript can also be used for blocking bots.

Using the latest version of Chrome I copied some of the JavaScript and ran it through ChatGPT to ask what it does. One part of it may relate to restricting abusive requests for documents.

Screenshot Of Chrome Dev Tools

ChatGPT gave me the following feedback:

“Core Functionalities
Randomized Value Generation (rdb)

Generates a random value based on properties (D_d, idc, and p4b) of the input object a, constrained by p7d.
This may be used for rate-limiting, exponential backoff, or similar logic.

Purpose and Context
From its components, the script:

Likely handles request retries or access control for web resources.

Implements a policy enforcement system, where:

Policies determine if requests are valid.

Errors are logged and sometimes retried based on rules.

Randomized delays or limits might control the retry mechanism.

Appears optimized for error handling and resilience in distributed or high-traffic systems, possibly within a Google service or API.”

ChatGPT said that the code may use rate-limiting which is a way to limit the number of actions a user or a system can take within a specific time period.

Rate-Limiting:

Used to enforce a limit on the number of actions (e.g., API requests) a user or system can perform within a specific time frame.
In this code, the random values generated by rdb could be used to introduce variability in when or how often requests are allowed, helping to manage traffic effectively.

Exponential Backoff:

ChatGPT explained that exponential backoff is a way to limit the amount of retries for a failed action a user or system is allowed to make. The time period between retries for a failed action increases exponentially.

Similar Logic:

ChatGPT explained that random value generation could be used to manage access to resources to prevent abusive requests.

I don’t know for certain that this is what that specific JavaScript is doing, that’s what ChatGPT explained and it definitely matches the information that Google shared that they are using JavaScript as part of their strategy for blocking bots.