Automattic’s Legal Claims About SEO… Is This Real? via @sejournal, @martinibuster

SEO plays a role in Automattic’s counterclaim against WP Engine. The legal document mentions search engine optimization six times and SEO once as part of counterclaims asserting that WP Engine excessively used words like “WordPress” to rank in search engines as part of an “infringement” campaign that uses WordPress trademarks in commerce. A close look at those claims shows that some of the evidence may be biased and that claims about SEO rely on outdated information.

Automattic’s Claims About SEO

Automattic’s counterclaim asserts that WP Engine used SEO to rank for WordPress-related keywords and that this is causing confusion.

The counterclaim explains:

“WP Engine also has sown confusion in recent years by dramatically increasing the number of times Counterclaimants’ Marks appear on its websites. Starting in or around 2021, WP Engine began to sharply increase its use of the WordPress Marks, and starting in or around 2022, began to sharply increase its use of the WooCommerce Marks.”

Automattic next argues that the repetition of keywords on a web page is WP Engine’s SEO strategy. Here’s where their claims become controversial to those who know how search engines rank websites.

The counterclaim asserts:

“The increased number of appearances of the WordPress Marks on WP Engine’s website is particularly likely to cause confusion in the internet context.

On information and belief, internet search engines factor in the number of times a term appears in a website’s text in assessing the “relevance” of a website to the terms a user enters into the search engine when looking for websites.

WP Engine’s decision to increase the number of times the WordPress Marks appear on WP Engine’s website appears to be a conscious “search engine optimization” strategy to ensure that when internet users look for companies that offer services related to WordPress, they will be exposed to confusingly written and formatted links that take them to WP Engine’s sites rather than WordPress.org or WordPress.com.”

They call WP Engine’s strategy aggressive:

“WP Engine’s strategy included aggressive utilization of search engine optimization to use the WordPress and WooCommerce Marks extremely frequently and confuse consumers searching for authorized providers of WordPress and WooCommerce software;”

Is The Number Of Keywords Used A Ranking Factor?

I have twenty-five years of experience in search engine optimization and have a concomitantly deep understanding of how search engines rank content. The fact is that Automattic’s claim that search engines “factor in the number of times” a keyword is used in a website’s content is incorrect. Modern search engines don’t factor in the number of times a keyword appears on a web page as a ranking factor. Google’s algorithms use models like BERT to gain a semantic understanding of the meaning and intent of the keyword phrases used in search queries and content, resulting in the ability to rank content that doesn’t even contain the user’s keywords.

Those aren’t just my opinions; Google’s web page about how search works explicitly says that content is ranked according to the user’s intent, regardless of keywords, which directly contradicts Automattic’s claim about WPE’s SEO:

“To return relevant results, we first need to establish what you’re looking for – the intent behind your query. To do this, we build language models to try to decipher how the relatively few words you enter into the search box match up to the most useful content available.

This involves steps as seemingly simple as recognizing and correcting spelling mistakes, and extends to our sophisticated synonym system that allows us to find relevant documents even if they don’t contain the exact words you used.”

If Google’s documentation is not convincing enough, take a look at the search results for the phrase “Managed WordPress Hosting.” WordPress.com ranks #2, despite the phrase being completely absent from its web page.

Screenshot Of WordPress.com In Search Results

What Is The Proof?

Automattic provides a graph comparing WP Engine’s average monthly mentions of the word “WordPress” with mentions published by 18 other web hosts. The comparison of 19 total web hosts dramatically illustrates that WP Engine mentions WordPress more often than any of the other hosting providers, by a large margin.

Screenshot Of Graph

Here’s a close-up of the graph (with the values inserted) showing that WP Engine’s monthly mentions of “WordPress” far exceed the number of times words containing WordPress are used on the web pages of the other hosts.

Screenshot Of Graph Closeup

People say that numbers don’t lie, and the graph presents compelling evidence that WP Engine is deploying an aggressive use of keywords with the word WordPress in them. Leaving aside the debunked idea that keyword-term spamming actually works, a closer look at the graph comparison shows that the evidence is not so strong because it is biased.

Automattic’s Comparison Is Arguably Biased

Automattic’s counterclaim compares eighteen web hosts against WP Engine. Of those eighteen hosts, only five (including WPE) are managed WordPress hosting platforms. The remaining fourteen are generalist hosting platforms that offer cloud hosting, VPS (virtual private servers), dedicated hosting, and domain name registrations.

The significance of this fact is that the comparison can be considered biased against WP Engine because the average mention of WordPress will naturally be lower across the entire website of a company that offers multiple services (like VPS, dedicated hosting, and domain name registrations) versus a site like WP Engine that offers only one service, managed WordPress hosting.

Two of the hosts listed in the comparison, Namecheap and GoDaddy, are primarily known as domain name registrars. Namecheap is the second biggest domain name registrar in the world. There’s no need to belabor the point that these two companies in Automattic’s comparison may be biased choices to compare against WP Engine.

Of the five hosts that offer WordPress hosting, two are plugin platforms: Elementor and WPMU Dev. Both are platforms built around their respective plugins, which means that the average number of mentions of WordPress is going to be lower because the average may be diluted by documentation and blog posts about the plugins. Those two companies are also arguably biased choices for this kind of comparison.

Of the eighteen hosts that Automattic chose to compare with WP Engine, only two of them are comparable in service to WP Engine: Kinsta and Rocket.net.

Comparison Of Managed WordPress Hosts

Automattic compares the monthly mentions of phrases with “WordPress” in them, and it’s clear that the choice of hosts in the comparison biases the results against WP Engine. A fairer comparison is to compare the top-ranked web page for the phrase “managed WordPress hosting.”

The following is a comparison of the top-ranked web page for each of the three managed WordPress hosts in Automattic’s comparison list, a straightforward one-to-one comparison. I used the phrase “managed WordPress hosting” plus the domain name appended to a search query in order to surface the top-ranked page from each website and then compared how many times the word “WordPress” is used on those pages.

Here are the results:

Rocket.net

The home page of Rocket.net ranks #1 for the phrase “rocket.net managed wordpress hosting.” The home page of Rocket.net contains the word “WordPress” 21 times.

Screenshot of Google’s Search Results

Kinsta

The top ranked Kinsta page is kinsta.com/wordpress-hosting/ and that page mentions the word “WordPress” 55 times.

WP Engine

The top ranked WP Engine web page is wpengine.com/managed-wordpress-hosting/ and that page mentions the word “WordPress” 27 times.

A fair one-to-one comparison of managed WordPress host providers, selected from Automattic’s own list, shows that WP Engine is not using the word “WordPress” more often than its competitors. Its use falls directly in the middle of a fair one-to-one comparison.

Number Of Times Page Mentions WordPress

  • Rocket.net: 21 times
  • WP Engine: 27 times
  • Kinsta: 55 times

What About Other Managed WordPress Hosts?

For the sake of comparison, I compared an additional five managed WordPress hosts that Automattic omitted from its comparison to see how often the word “WordPress” was mentioned on the top-ranked web pages of WP Engine’s direct competitors.

Here are the results:

  • WPX Hosting: 9
  • Flywheel: 16
  • InstaWP: 22
  • Pressable: 23
  • Pagely: 28

It’s apparent that WP Engine’s 27 mentions put it near the upper level in that comparison, but nowhere near the level at which Kinsta mentions “WordPress.” So far, we only see part of the story. As you’ll see, other web hosts use the word “WordPress” far more than Kinsta does, and it won’t look like such an outlier when compared to generalist web hosts.

A Comparison With Generalist Web Hosts

Next, we’ll compare the generalist web hosts listed in Automattic’s comparison.

I did the same kind of search for the generalist web hosts to surface their top-ranked pages for the query “managed WordPress hosting” plus the name of the website, which is a one-to-one comparison to WP Engine.

Other Web Hosts Compared To WP Engine:

  1. InMotion Hosting: 101 times
  2. Greengeeks: 97 times
  3. Jethost: 71 times
  4. Verpex: 52 times
  5. GoDaddy: 49 times
  6. Cloudways: 47 times
  7. Namecheap: 41 times
  8. Liquidweb: 40 times
  9. Pair: 40 times
  10. Hostwinds: 37 times
  11. KnownHost: 33 times
  12. Mochahost: 33 times
  13. Panthen: 31 times
  14. Siteground: 30 times
  15. WP Engine: 27 times

Crazy, right? WP Engine uses the word “WordPress” less often than any of the other generalist web hosts. This one-to-one comparison contradicts Automattic’s graph.

And just for the record, WordPress.com’s top-ranked page wordpress.com/hosting/ uses the word “WordPress” 62 times, over twice as often as WP Engine’s web page.

Will Automattic’s SEO Claims Be Debunked?

Automattic’s claims about WP Engine’s use of SEO may be based on shaky foundations. The claims about how keywords work for SEO contradict Google’s own documentation, and the fact that WordPress.com’s own website ranks for the phrase “Managed WordPress Hosting” despite not using that exact phrase appears to debunk their assertion that search engines factor the number of times a user’s keywords are used on a web page.

The graph that Automattic presents in their counterclaim does not represent a comparison of direct competitors, which may contribute to a biased impression that WP Engine is aggressively using the “WordPress” keywords more often than competitors. However, a one-to-one comparison of the actual web pages that compete against each other for the phrase “Managed WordPress Hosting” shows that many of the web hosts in Automattic’s own list use the word “WordPress” far more often than WP Engine, which directly contradicts Automattic’s narrative.

I ran WP Engine’s Managed WordPress Hosting URL in a Keyword Density Tool, and it shows that WP Engine’s web page uses the word “WordPress” a mere 1.92% of the time, which, from an SEO point of view, could be considered a modest amount and far from excessive. It will be interesting to see how the judge decides the merits of Automattic’s SEO-related claims.

Featured Image by Shutterstock/file404

Perplexity Responds To Reddit Lawsuit Over Data Access via @sejournal, @MattGSouthern

Reddit sued Perplexity and three data-scraping firms in New York federal court, alleging the companies bypassed access controls to obtain Reddit content at scale, including by scraping Google search results.

Perplexity posted a public response, saying it summarizes Reddit discussions with citations and doesn’t train AI models on Reddit content.

The position is consistent with the company’s past statements. Whether it addresses the specific allegations in Reddit’s filing remains an open question.

The complaint names Oxylabs UAB, AWMProxy, and SerpApi as intermediaries. It alleges Perplexity is a SerpApi customer and purchased and/or utilized SerpApi services to circumvent controls and copy Reddit data.

Evidence In The Complaint

Perplexity’s argument is built around a technical distinction. The company says it summarizes and cites discussions rather than training models on Reddit posts.

Perplexity wrote in its Reddit response:

“We summarize Reddit discussions, and we cite Reddit threads in answers, just like people share links to posts here all the time.”

The complaint, however, presents technical claims that call that framework into question.

According to the filing, Reddit created a test post that was only crawlable by Google’s search engine and not accessible anywhere else on the internet. Within hours, that hidden content appeared in Perplexity’s results.

The filing also says that after Reddit sent a cease-and-desist letter, Perplexity’s citations to Reddit increased roughly forty-fold.

Similar Accusations From Publishers

Forbes previously accused Perplexity of republishing an exclusive and threatened legal action.

Wired reported that Perplexity used undisclosed IPs and spoofed user-agent strings to bypass robots.txt. Wired’s

Cloudflare later said Perplexity used “stealth, undeclared crawlers” that ignored no-crawl directives, based on tests it ran in August.

How Perplexity Has Responded

In previous disputes, Perplexity said issues stemmed from rough edges on new products and promised clearer attribution.

The company has also argued that some media organizations are trying to control “publicly reported facts.”

In this latest response, Perplexity frames Reddit’s lawsuit as leverage in broader training-data negotiations and writes:

“We summarize Reddit discussions… We won’t be extorted, and we won’t help Reddit extort Google.”

Why This Matters

This issue matters because it concerns how AI assistants use forum content that your audiences read and that publishers frequently cite.

The legal questions go beyond just training.

Courts may examine if technical controls have been bypassed, whether summarization infringes on protected expressions, and if using third-party scrapers could lead to legal liability for downstream products.

If courts accept Reddit’s anti-circumvention argument, it could lead to changes in how assistants cite or link Reddit threads.

On the other hand, if courts agree with Perplexity’s viewpoint, assistants might start relying more on forum discussions that are less restricted by licensing.

What We Don’t Know Yet

The filing alleges Perplexity obtained data via at least one scraping firm, but the public complaint doesn’t specify which vendor supplied which data or include transaction details.

OpenAI Releases Shared Project Feature To All Users via @sejournal, @martinibuster

OPenAI announced that it is upgrading all ChatGPT accounts to be eligible for the project sharing feature, which enables users to share a ChatGPT project with others who can then participate and make changes. The feature was previously available only to users on OpenAI’s Business, Enterprise, and Edu plans.

The new feature is available to users globally in the Free, Plus, Pro, and budget Go plans, whether accessed on the web, iOS, or Android devices.

There are limits specific to each plan according to the announcement:

  • Free users can collaborate on up to 5 files and with 5 collaborators
  • Plus and Go users can share up to 25 files with up to 10 collaborators
  • Pro users can share up to 40 files with up to 100 collaborators

    OpenAI suggested the following use cases:

“Group work: Upload notes, proposals, and contracts so collaborators can draft deliverables faster and stay in sync.

Content creation: Apply project-specific instructions to keep tone and style consistent across contributors.

Reporting: Store datasets and reports in one project, and return each week to generate updates without starting over.

Research: Keep transcripts, survey results, and market research in one place, so anyone in the project can query and build on the findings.

Project owners can choose to share a project with “Only those invited” or “Anyone with a link,” and can change visibility settings at any time including switching back to invite-only.”

Read more at OpenAI: Shared Projects

Featured Image by Shutterstock/LedyX

Microsoft Updates Copilot With Memory, Search Connectors, & More via @sejournal, @MattGSouthern

Microsoft announced its Copilot Fall Release, introducing features to make AI more personal and collaborative.

New capabilities include group collaboration, long-term memory, health tools, and voice-enabled learning.

Mustafa Suleyman, head of Microsoft AI, wrote in the announcement that the release represents a shift in how AI supports users.

Suleyman wrote:

“… technology should work in service of people. Not the other way around. Ever.”

What’s New

Search Improvements

Copilot Search combines AI-generated answers with traditional results in one view, providing cited responses for faster discovery.

Microsoft also highlighted its in-house models, including MAI-Voice-1, MAI-1-Preview, and MAI-Vision-1, as groundwork for more immersive Copilot experiences.

Memory & Personalization

Copilot now includes long-term memory that tracks user preferences and information across conversations.

You can ask Copilot to remember specific details like training for a marathon or an anniversary, and the AI can recall this information in future interactions. Users can edit, update, or delete memories at any time.

Search Across Services

New connector features link Copilot to OneDrive, Outlook, Gmail, Google Drive, and Google Calendar so you can search for documents, emails, and calendar events across multiple accounts using natural language.

Microsoft notes this is rolling out gradually and may not yet be available in all regions or languages.

Edge & Windows Integration

Copilot Mode in Edge is evolving into what Microsoft calls an “AI browser.”

With user permission, Copilot can see open tabs, summarize information, and take actions like booking hotels or filling forms.

Voice-only navigation enables hands-free browsing. Journeys and Actions are currently available in the U.S. only.

Shared AI Sessions

The Groups feature turns Copilot into a collaborative workspace for up to 32 people.

You can invite friends, classmates, or teammates to shared sessions. Start a session by sending a link, and anyone with the link can join and see the same conversation in real time.

This feature is U.S. only at launch.

Health Features

Copilot for health grounds responses in credible sources like Harvard Health for medical questions.

Health features are available only in the U.S. at copilot.microsoft.com and in the Copilot iOS app.

Voice Tutoring

Learn Live provides voice-enabled Socratic tutoring for educational topics.

Interactive whiteboards help you work through concepts for test preparation, language practice, or exploring new subjects. U.S. only.

“Mico” Character

Microsoft introduced Mico, an optional visual character that reacts during voice conversations.

Separately, Copilot adds a “real talk” conversation style that challenges assumptions and adapts to user preferences.

Why This Matters

These features change how Copilot fits into your workflow.

The move from individual to collaborative sessions means teams can use AI together rather than separately synthesizing results.

Long-term memory reduces the need to repeat context, which matters for ongoing projects where Copilot needs to understand your specific situation.

Looking Ahead

Features are live in the U.S. now. Microsoft says updates are rolling out across the UK, Canada, and beyond in the next few weeks.

Some features require a Microsoft 365 Personal, Family, or Premium subscription; usage limits apply. Specific availability varies by market, device, and platform.

Search Engine Journal Is Hiring! via @sejournal, @hethr_campbell

We’re looking for a powerhouse project manager to keep our marketing team inspired and on track.

This is a Philippines-based, fully-remote position working on U.S.-adjacent hours (8 p.m. – 4 a.m. PHT) to be my partner in crime execution on some exciting projects.

We do things a little differently here, and we’ve learned that culture fit is everything. When it’s a match, people tend to stay; nearly half our team has been with us for more than five years.

If you’re the kind of team member who has loads of experience leaning into complex projects, honest conversations, and big ideas, we’d love to meet you.

About SEJ

We help our advertisers communicate with precision and creativity in an AI-driven world. Our campaigns are built on data, empathy, and continuous experimentation. We manage multi-channel strategies across content, email, and social media, and we’re looking for someone who can keep the moving parts aligned without losing sight of the humans behind them.

We’re hiring a Senior Digital Marketing Project Manager to lead strategy execution, client relationships, and team coordination. You’ll help us build marketing systems that are smart, efficient, and grounded in trust.

Why This Role Is Different

This is an AI-first position. You already use tools like ChatGPT, Claude, or Gemini to work smarter, automate workflows, and uncover insights that move the needle. Your success here depends on seeing where AI enhances human creativity … and where it doesn’t.

We’re a team that values autonomy, initiative, and straight talk. We’d rather have one clear, respectful conversation than weeks of confusion. We care deeply about doing great work and making each other better through feedback and shared accountability.

What You’ll Do

  • Manage and optimize complex digital marketing campaigns from strategy to execution.
  • Translate business goals into clear, actionable plans for clients and internal teams.
  • Keep communication flowing: up, down, and across.
  • Identify opportunities to integrate AI tools into analytics and operations.
  • Support a culture of feedback, growth, and curiosity.

Who You Are

  • You’re organized and strategic, but not rigid. You like structure, but you also know when to improvise.
  • You’re skilled at managing both clients and creatives. You can lead with empathy and keep projects on schedule.
  • You don’t shy away from a tough conversation if it means getting to a better outcome.

You’re the kind of team member who says things like:

  • “Let’s make sure we’re solving the right problem.”
  • “I appreciate the feedback! Here’s what I’m hearing.”
  • “How can AI help us work smarter here?”

Why Work With Search Engine Journal?

We’re a remote-first, global team that values:

  • Clarity over chaos.
  • Progress over perfection.
  • Honest collaboration over hierarchy.
  • We’re remote, flexible, and results-focused. You’ll have real ownership, real support, and the chance to do your best work with people who actually care about doing theirs.

If this sounds like your kind of place, see the full job listing and apply here.


Featured Image: PeopleImages/Shutterstock

Google’s New BlockRank Democratizes Advanced Semantic Search via @sejournal, @martinibuster

A new research paper from Google DeepMind  proposes a new AI search ranking algorithm called BlockRank that works so well it puts advanced semantic search ranking within reach of individuals and organizations. The researchers conclude that it “can democratize access to powerful information discovery tools.”

In-Context Ranking (ICR)

The research paper describes the breakthrough of using In-Context Ranking (ICR), a way to rank web pages using a large language model’s contextual understanding abilities.

It prompts the model with:

  1. Instructions for the task (for example, “rank these web pages”)
  2. Candidate documents (the pages to rank)
  3. And the search query.

ICR is a relatively new approach first explored by researchers from Google DeepMind and Google Research in 2024 (Can Long-Context Language Models Subsume Retrieval, RAG, SQL, and More? PDF). That earlier study showed that ICR could match the performance of retrieval systems built specifically for search.

But that improvement came with a downside in that it requires escalating computing power as the number of pages to be ranked are increased.

When a large language model (LLM) compares multiple documents to decide which are most relevant to a query, it has to “pay attention” to every word in every document and how each word relates to all others. This attention process gets much slower as more documents are added because the work grows exponentially.

The new research solves that efficiency problem, which is why the research paper is called, Scalable In-context Ranking with Generative Models, because it shows how to scale In-context Ranking (ICR) with what they call BlockRank.

How BlockRank Was Developed

The researchers examined how the model actually uses attention during In-Context Retrieval and found two patterns:

  • Inter-document block sparsity:
    The researchers found that when the model reads a group of documents, it tends to focus mainly on each document separately instead of comparing them all to each other. They call this “block sparsity,” meaning there’s little direct comparison between different documents. Building on that insight, they changed how the model reads the input so that it reviews each document on its own but still compares all of them against the question being asked. This keeps the part that matters, matching the documents to the query, while skipping the unnecessary document-to-document comparisons. The result is a system that runs much faster without losing accuracy.
  • Query-document block relevance:
    When the LLM reads the query, it doesn’t treat every word in that question as equally important. Some parts of the question, like specific keywords or punctuation that signal intent, help the model decide which document deserves more attention. The researchers found that the model’s internal attention patterns, particularly how certain words in the query focus on specific documents, often align with which documents are relevant. This behavior, which they call “query-document block relevance,” became something the researchers could train the model to use more effectively.

The researchers identified these two attention patterns and then designed a new approach informed by what they learned. The first pattern, inter-document block sparsity, revealed that the model was wasting computation by comparing documents to each other when that information wasn’t useful. The second pattern, query-document block relevance, showed that certain parts of a question already point toward the right document.

Based on these insights, they redesigned how the model handles attention and how it is trained. The result is BlockRank, a more efficient form of In-Context Retrieval that cuts unnecessary comparisons and teaches the model to focus on what truly signals relevance.

Benchmarking Accuracy Of BlockRank

The researchers tested BlockRank for how well it ranks documents on three major benchmarks:

  • BEIR
    A collection of many different search and question-answering tasks used to test how well a system can find and rank relevant information across a wide range of topics.
  • MS MARCO
    A large dataset of real Bing search queries and passages, used to measure how accurately a system can rank passages that best answer a user’s question.
  • Natural Questions (NQ)
    A benchmark built from real Google search questions, designed to test whether a system can identify and rank the passages from Wikipedia that directly answer those questions.

They used a 7-billion-parameter Mistral LLM and compared BlockRank to other strong ranking models, including FIRST, RankZephyr, RankVicuna, and a fully fine-tuned Mistral baseline.

BlockRank performed as well as or better than those systems on all three benchmarks, matching the results on MS MARCO and Natural Questions and doing slightly better on BEIR.

The researchers explained the results:

“Experiments on MSMarco and NQ show BlockRank (Mistral-7B) matches or surpasses standard fine-tuning effectiveness while being significantly more efficient at inference and training. This offers a scalable and effective approach for LLM-based ICR.”

They also acknowledged that they didn’t test multiple LLMs and that these results are specific to Mistral 7B.

Is BlockRank Used By Google?

The research paper says nothing about it being used in a live environment. So it’s purely conjecture to say that it might be used. Also, it’s natural to try to identify where BlockRank fits into AI Mode or AI Overviews but the descriptions of how AI Mode’s FastSearch and RankEmbed work are vastly different from what BlockRank does. So it’s unlikely that BlockRank is related to FastSearch or RankEmbed.

Why BlockRank Is A Breakthrough

What the research paper does say is that this is a breakthrough technology that puts an advanced ranking system within reach of individuals and organizations that wouldn’t normally be able to have this kind of high quality ranking technology.

The researchers explain:

“The BlockRank methodology, by enhancing the efficiency and scalability of In-context Retrieval (ICR) in Large Language Models (LLMs), makes advanced semantic retrieval more computationally tractable and can democratize access to powerful information discovery tools. This could accelerate research, improve educational outcomes by providing more relevant information quickly, and empower individuals and organizations with better decision-making capabilities.

Furthermore, the increased efficiency directly translates to reduced energy consumption for retrieval-intensive LLM applications, contributing to more environmentally sustainable AI development and deployment.

By enabling effective ICR on potentially smaller or more optimized models, BlockRank could also broaden the reach of these technologies in resource-constrained environments.”

SEOs and publishers are free to their opinions of whether or not this could be used by Google. I don’t think there’s evidence of that but it would be interesting to ask a Googler about it.

Google appears to be in the process of making BlockRank available on GitHub, but it doesn’t appear to have any code available there yet.

Read about BlockRank here:
Scalable In-context Ranking with Generative Models

Featured Image by Shutterstock/Nithid

AI Assistants Show Significant Issues In 45% Of News Answers via @sejournal, @MattGSouthern

Leading AI assistants misrepresented or mishandled news content in nearly half of evaluated answers, according to a European Broadcasting Union (EBU) and BBC study.

The research assessed free/consumer versions of ChatGPT, Copilot, Gemini, and Perplexity, answering news questions in 14 languages across 22 public-service media organizations in 18 countries.

The EBU said in announcing the findings:

“AI’s systemic distortion of news is consistent across languages and territories.”

What The Study Found

In total, 2,709 core responses were evaluated, with qualitative examples also drawn from custom questions.

Overall, 45% of responses contained at least one significant issue, and 81% had some issue. Sourcing was the most common problem area, affecting 31% of responses at a significant level.

How Each Assistant Performed

Performance varied by platform. Google Gemini showed the most issues: 76% of its responses contained significant problems, driven by 72% with sourcing issues.

The other assistants were at or below 37% for major issues overall and below 25% for sourcing issues.

Examples Of Errors

Accuracy problems included outdated or incorrect information.

For instance, several assistants identified Pope Francis as the current Pope in late May, despite his death in April, and Gemini incorrectly characterized changes to laws on disposable vapes.

Methodology Notes

Participants generated responses between May 24 and June 10, using a shared set of 30 core questions plus optional local questions.

The study focused on the free/consumer versions of each assistant to reflect typical usage.

Many organizations had technical blocks that normally restrict assistant access to their content. Those blocks were removed for the response-generation period and reinstated afterward.

Why This Matters

When using AI assistants for research or content planning, these findings reinforce the need to verify claims against original sources.

As a publication, this could impact how your content is represented in AI answers. The high rate of errors increases the risk of misattributed or unsupported statements appearing in summaries that cite your content.

Looking Ahead

The EBU and BBC published a News Integrity in AI Assistants Toolkit alongside the report, offering guidance for technology companies, media organizations, and researchers.

Reuters reports the EBU’s view that growing reliance on assistants for news could undermine public trust.

As EBU Media Director Jean Philip De Tender put it:

“When people don’t know what to trust, they end up trusting nothing at all, and that can deter democratic participation.”


Featured Image: Naumova Marina/Shutterstock

YouTube Expands Likeness Detection To All Monetized Channels via @sejournal, @MattGSouthern

YouTube is beginning to expand access to its likeness detection tool to all channels in the YouTube Partner Program over the next few months.

The technology helps you identify unauthorized videos where your facial likeness has been altered or generated with AI.

YouTube announced the expansion after testing the tool with a small group of creators.

The tool addresses a growing concern as AI-generated content becomes more sophisticated and accessible.

How Likeness Detection Works

Channels can access the tool through YouTube Studio’s content detection tab under a new likeness section.

The onboarding process requires identity verification. You scan a QR code with your phone’s camera, then submit a photo ID and record a brief selfie video performing specific motions.

YouTube processes this information on Google servers, typically granting access within a few days.

Once verified, creators see a dashboard displaying videos that match their facial likeness. The interface shows video titles, upload dates, upload channels, view counts, and subscriber numbers. YouTube’s systems flag some matches as higher priority for review.

Taking Action On Detected Content

You have three options when reviewing matches.

You can request removal under YouTube’s privacy guidelines, submit a copyright claim, or archive the video without action. The tool automatically fills legal name and email information when starting a removal request.

Privacy removal requests apply to altered or synthetic content that violates specific criteria. YouTube’s announcement highlighted two examples: AI-generated videos showing creators endorsing political candidates, and infomercials with creators’ faces added through AI.

Copyright claims follow different rules and must consider fair use exceptions. Videos using short clips from a creator’s channel may not qualify for privacy removal but could warrant copyright action.

See a demonstration in the video below:

Policy Differences

YouTube stressed the distinction between privacy and copyright policies.

Privacy policy violations involve altered or synthetic content judged against criteria including whether the content is parody, satire, or includes AI disclosure. Copyright infringement covers unauthorized use of original content, including cropped videos to avoid detection or videos with changed audio.

The tool surfaces some short clips from creators’ own channels. These don’t qualify for privacy removal but may be eligible for copyright claims if fair use doesn’t apply.

Why This Matters

This gives YouTube Partner Program creators direct control over how AI-generated content uses their likeness.

Monetized channels can now monitor unauthorized deepfakes and request removal when videos mislead the audience about endorsements or statements that were never made.

Looking Ahead

The tool will roll out to eligible creators over the next few months. Those who see no matches shouldn’t be concerned. YouTube says this indicates no detected unauthorized use of their likeness on the platform.

Channels can withdraw consent and stop using the tool at any time through the manage likeness detection settings.

Surfer SEO Acquired By Positive Group via @sejournal, @martinibuster

The French technology group Positive acquired Surfer, the popular content optimization tool. The acquisition helps Positive create a “full-funnel” brand visibility solution together with its marketing and CRM tools.

The acquisition of Surfer extends Positive’s reach from marketing software to AI-based brand visibility. Positive described the deal as part of a European AI strategy that supports jobs and protects data. Positive’s revenue has grown fivefold in the past five years, rising from €50 million to an expected €70 million in 2025.

Surfer SEO

Founded in 2017, Surfer developed SEO tools based on language models that help marketers improve visibility on both search engines and AI assistants, which have become a growing source of website traffic and customers.

Sign Of Broader Industry Trends

The acquisition shows that search optimization continues to be an important part of business marketing as AI search and chat play a larger role in how consumers learn about products, services, and brands. This deal enables Positive to offer AI-based visibility solutions alongside its CRM and automation products, expanding its technology portfolio.

What Acquisition Means For Customers

Positive Group, based in France, is a technology solutions company that develops digital tools for marketing, CRM, automation, and data management. It operates through several divisions: User (marketing and CRM), Signitic (email signatures), and now Surfer (AI search optimization). The company is majority-owned by its executives, employs about 400 people, and keeps its servers in France and Germany. Surfer, based in Poland, brings experience in AI content optimization and a strong presence in North America. Together, they combine infrastructure, market knowledge, and product development within one technology-focused group.

Lucjan Suski, CEO and co-founder of Surfer, commented:

“SEO is evolving fast, and it matters more than ever before. We help marketers win the AI SEO era. Positive helps them grow across every other part of their digital strategy. Together, we’ll give marketers the complete toolkit to lead across AI search, email marketing automation, and beyond.”

According to Mathieu Tarnus, Positive’s founding president, and Paul de Fombelle, its CEO:

“Artificial intelligence is at the heart of our value proposition. With the acquisition of Surfer, our customers are moving from optimizing their traditional SEO positioning to optimizing their brand presence in the responses provided by conversational AI assistants. Surfer stands out from established market players by directly integrating AI into content creation and optimization.”

The acquisition adds Surfer’s AI optimization capabilities to Positive’s product ecosystem, helping customers improve visibility in AI-generated answers. For both companies, the deal is an opportunity to expand their capabilities in AI-based brand visibility.

Featured Image by Shutterstock/GhoST RideR 98

Brave Reveals Systemic Security Issues In AI Browsers via @sejournal, @MattGSouthern

Brave disclosed security vulnerabilities in AI browsers that could allow malicious websites to hijack AI assistants and access sensitive user accounts.

The issues affect Perplexity Comet, Fellou, and potentially other AI browsers that can take actions on behalf of users.

The vulnerabilities stem from indirect prompt injection attacks where websites embed hidden instructions that AI browsers process as legitimate user commands. Brave published the findings after reporting the issues to affected companies.

What Brave Found

Perplexity Comet Vulnerability

Comet’s screenshot feature can be exploited by embedding nearly invisible text in webpages.

When users take screenshots to ask questions, the AI extracts hidden text using what appears to be OCR and processes it as commands rather than untrusted content.

Brave notes Comet isn’t open-source, so this behavior is inferred and can’t be verified from source code.

The hidden instructions use faint colors that humans can barely see but AI systems extract and execute. This lets attackers issue commands to the AI assistant without the user’s knowledge.

Fellou Navigation Vulnerability

Fellou browser sends webpage content to its AI system when users navigate to a site.

Asking the AI assistant to visit a webpage causes the browser to pass the page’s visible content to the AI in a way that lets the webpage text override user intent.

This means visiting a malicious site could trigger unintended AI actions without requiring explicit user interaction with the AI assistant.

Access To Sensitive Accounts

The vulnerabilities become dangerous because AI assistants operate with user authentication privileges.

A hijacked AI browser can access banking sites, email providers, work systems, and cloud storage where users remain logged in.

Brave notes that even summarizing a Reddit post could result in attackers stealing money or private data if the post contains hidden malicious instructions.

Industry Context

Brave describes indirect prompt injection as a systemic challenge facing AI browsers rather than an isolated issue.

The problem revolves around AI systems failing to distinguish between trusted user input and untrusted webpage content when constructing prompts.

Brave is withholding details of one additional vulnerability found in another browser until next week.

Why This Matters

Brave argues that traditional web security models break when AI agents act on behalf of users.

Natural language instructions on any webpage can trigger cross-domain actions reaching banks, healthcare providers, corporate systems, and email hosts.

Same-origin policy protections become irrelevant because AI assistants execute with full user privileges across all authenticated sites.

The disclosure arrives the same day OpenAI launched ChatGPT Atlas with agent mode capabilities, highlighting the tension between AI browser functionality and security.

People using AI browsers with agent features face a tradeoff between automation capabilities and exposure to these systemic vulnerabilities.

Looking Ahead

Brave’s research continues with additional findings scheduled for disclosure next week.

The company indicated it’s exploring longer-term solutions to address the trust boundary problems in agentic browsing.


Featured Image: Who is Danny/Shutterstock