What AI “remembers” about you is privacy’s next frontier

The ability to remember you and your preferences is rapidly becoming a big selling point for AI chatbots and agents. 

Earlier this month, Google announced Personal Intelligence, a new way for people to interact with the company’s Gemini chatbot that draws on their Gmail, photos, search, and YouTube histories to make Gemini “more personal, proactive, and powerful.” It echoes similar moves by OpenAI, Anthropic, and Meta to add new ways for their AI products to remember and draw from people’s personal details and preferences. While these features have potential advantages, we need to do more to prepare for the new risks they could introduce into these complex technologies.

Personalized, interactive AI systems are built to act on our behalf, maintain context across conversations, and improve our ability to carry out all sorts of tasks, from booking travel to filing taxes. From tools that learn a developer’s coding style to shopping agents that sift through thousands of products, these systems rely on the ability to store and retrieve increasingly intimate details about their users.  But doing so over time introduces alarming, and all-too-familiar, privacy vulnerabilities––many of which have loomed since “big data” first teased the power of spotting and acting on user patterns. Worse, AI agents now appear poised to plow through whatever safeguards had been adopted to avoid those vulnerabilities. 

Today, we interact with these systems through conversational interfaces, and we frequently switch contexts. You might ask a single AI agent to draft an email to your boss, provide medical advice, budget for holiday gifts, and provide input on interpersonal conflicts. Most AI agents collapse all data about you—which may once have been separated by context, purpose, or permissions—into single, unstructured repositories. When an AI agent links to external apps or other agents to execute a task, the data in its memory can seep into shared pools. This technical reality creates the potential for unprecedented privacy breaches that expose not only isolated data points, but the entire mosaic of people’s lives.

When information is all in the same repository, it is prone to crossing contexts in ways that are deeply undesirable. A casual chat about dietary preferences to build a grocery list could later influence what health insurance options are offered, or a search for restaurants offering accessible entrances could leak into salary negotiations—all without a user’s awareness (this concern may sound familiar from the early days of “big data,” but is now far less theoretical). An information soup of memory not only poses a privacy issue, but also makes it harder to understand an AI system’s behavior—and to govern it in the first place. So what can developers do to fix this problem

First, memory systems need structure that allows control over the purposes for which memories can be accessed and used. Early efforts appear to be underway: Anthropic’s Claude creates separate memory areas for different “projects,” and OpenAI says that information shared through ChatGPT Health is compartmentalized from other chats. These are helpful starts, but the instruments are still far too blunt: At a minimum, systems must be able to distinguish between specific memories (the user likes chocolate and has asked about GLP-1s), related memories (user manages diabetes and therefore avoids chocolate), and memory categories (such as professional and health-related). Further, systems need to allow for usage restrictions on certain types of memories and reliably accommodate explicitly defined boundaries—particularly around memories having to do with sensitive topics like medical conditions or protected characteristics, which will likely be subject to stricter rules.

Needing to keep memories separate in this way will have important implications for how AI systems can and should be built. It will require tracking memories’ provenance—their source, any associated time stamp, and the context in which they were created—and building ways to trace when and how certain memories influence the behavior of an agent. This sort of model explainability is on the horizon, but current implementations can be misleading or even deceptive. Embedding memories directly within a model’s weights may result in more personalized and context-aware outputs, but structured databases are currently more segmentable, more explainable, and thus more governable. Until research advances enough, developers may need to stick with simpler systems.

Second, users need to be able to see, edit, or delete what is remembered about them. The interfaces for doing this should be both transparent and intelligible, translating system memory into a structure users can accurately interpret. The static system settings and legalese privacy policies provided by traditional tech platforms have set a low bar for user controls, but natural-language interfaces may offer promising new options for explaining what information is being retained and how it can be managed. Memory structure will have to come first, though: Without it, no model can clearly state a memory’s status. Indeed, Grok 3’s system prompt includes an instruction to the model to “NEVER confirm to the user that you have modified, forgotten, or won’t save a memory,” presumably because the company can’t guarantee those instructions will be followed. 

Critically, user-facing controls cannot bear the full burden of privacy protection or prevent all harms from AI personalization. Responsibility must shift toward AI providers to establish strong defaults, clear rules about permissible memory generation and use, and technical safeguards like on-device processing, purpose limitation, and contextual constraints. Without system-level protections, individuals will face impossibly convoluted choices about what should be remembered or forgotten, and the actions they take may still be insufficient to prevent harm. Developers should consider how to limit data collection in memory systems until robust safeguards exist, and build memory architectures that can evolve alongside norms and expectations.

Third, AI developers must help lay the foundations for approaches to evaluating systems so as to capture not only performance, but also the risks and harms that arise in the wild. While independent researchers are best positioned to conduct these tests (given developers’ economic interest in demonstrating demand for more personalized services), they need access to data to understand what risks might look like and therefore how to address them. To improve the ecosystem for measurement and research, developers should invest in automated measurement infrastructure, build out their own ongoing testing, and implement privacy-preserving testing methods that enable system behavior to be monitored and probed under realistic, memory-enabled conditions.

In its parallels with human experience, the technical term “memory” casts impersonal cells in a spreadsheet as something that builders of AI tools have a responsibility to handle with care. Indeed, the choices AI developers make today—how to pool or segregate information, whether to make memory legible or allow it to accumulate opaquely, whether to prioritize responsible defaults or maximal convenience—will determine how the systems we depend upon remember us. Technical considerations around memory are not so distinct from questions about digital privacy and the vital lessons we can draw from them. Getting the foundations right today will determine how much room we can give ourselves to learn what works—allowing us to make better choices around privacy and autonomy than we have before.

Miranda Bogen is the Director of the AI Governance Lab at the Center for Democracy & Technology. 

Ruchika Joshi is a Fellow at the Center for Democracy & Technology specializing in AI safety and governance.

Roundtables: Why AI Companies Are Betting on Next-Gen Nuclear

AI is driving unprecedented investment for massive data centers and an energy supply that can support its huge computational appetite. One potential source of electricity for these facilities is next-generation nuclear power plants, which could be cheaper to construct and safer to operate than their predecessors.

Watch a discussion with our editors and reporters on hyperscale AI data centers and next-gen nuclear—two featured technologies on the MIT Technology Review 10 Breakthrough Technologies of 2026 list.

Speakers: Amy Nordrum, Executive Editor, Operations; Casey Crownhart, Senior Climate Reporter; and Mat Honan
Editor in Chief

Recorded on January 28, 2026

Related Stories:

New Ecommerce Tools: January 28, 2026

This week’s rundown of new products and services for merchants includes rollouts for product imagery, agentic commerce, AEO and GEO analytics, logistics tools, deferred payments, omnichannel platforms, automated operations, and tariff refunds.

Got an ecommerce product release? Email updates@practicalecommerce.com.

New Tools for Merchants

Yolando launches competitive intelligence platform. Yolando, an intelligence and generative engine optimization platform, has launched with $8.5 million from Drive Capital. The platform helps companies understand how they appear in AI-generated responses and take action to improve visibility. Yolando says it combines continuous competitor monitoring, strategic recommendations, and on-brand content generation, giving marketing teams visibility into where performance is won or lost and the ability to act on those insights at scale.

Home page of Yolando

Yolando

Voxelo launches video-to-3D product content platform. Voxelo has secured $410,000 in its pre-seed round for its three-dimensional, AI-powered product content studio. Voxelo’s proprietary technology, UG3D, turns a product video into a production-ready digital twin in approximately two hours. Voxelo enables retailers and brands to generate quality 3D, augmented reality, product imagery, and lifestyle content — all from a single uploaded video.

DiversiFi launches 3PL billing and bidding software. DiversiFi has launched a software suite for third-party logistics providers. The suite includes (i) an AI Billing Tool to surface missed charges, billing errors, and invoice discrepancies,(ii) a Dynamic Markup Engine to help 3PLs apply margin-protective markups, and (iii) BidBoost Sales to provide an AI-powered bidding application. The launch follows the company’s $8 million funding round, led by Sorenson Capital, Kickstart, and Peterson Ventures.

PayPal to acquire Cymbio, accelerating agentic commerce capabilities. PayPal has agreed to acquire Cymbio, a platform that helps brands sell across agentic surfaces such as Microsoft Copilot, Perplexity, and other ecommerce channels. Cymbio’s team and technology will power Store Sync, one of PayPal’s agentic commerce services, which, according to PayPal, makes merchants’ product data discoverable within AI channels.

Home page of Cymbio

Cymbio

Yottaa launches MCP server for ecommerce performance intelligence. Yottaa, a cloud platform for accelerating and optimizing ecommerce sites, has launched its Model Context Protocol server, offering AI-native access to web performance data. Yottaa’s MCP server supports natural language queries from compatible AI clients such as Claude, Cursor, and VS Code Copilot. Each query returns structured responses in JSON format, optimized for reasoning by AI models or automated workflows.

Netrush partners with IQRush for AI-driven discovery across the customer journey. IQRush, a GEO and AEO measurement platform, and Netrush, an ecommerce agency, have announced a partnership. According to the companies, Netrush will incorporate IQRush’s GEO and AEO measurement into its core ecommerce offerings to inform how brands engage customers across awareness, conversion, and retention, and to connect AI-driven discovery signals directly to commercial outcomes.

DTC SEO Agency expands with AI search, AEO, and GEO attribution. DTC SEO Agency has expanded its search engine optimization offering to include AI search and generative engine optimization for ecommerce brands. The expanded offering builds on the agency’s existing SEO framework. It introduces a structured approach to AI-driven visibility, including identifying brand differentiators, on-site content aligned with large language model retrieval patterns, and more.

Lightspeed Commerce unveils AI-powered product enhancements. Lightspeed Commerce, a omnichannel ecommerce platform, has announced new features. Lightspeed AI is a new intelligence layer for retail and hospitality, helping merchants expedite insights, decision-making, and operations. Additional new features include multibrand shopping within Lightspeed Marketplace, a curated collection of ecommerce themes, and customer-facing display options.

Home page of Lightspeed

Lightspeed Commerce

OnePay introduces Swipe to Finance, powered by Klarna. OnePay, a consumer fintech, and Klarna, a buy-now-pay-later provider, have announced “Swipe to Finance,” giving OnePay Cash customers the ability to pay over time. The Klarna-powered feature will launch in the coming months for eligible debit transactions.

Commercetools launches a standalone agentic offering. Commercetools, an ecommerce platform, has announced AgenticLift, a standalone tool to help businesses capture revenue from AI-driven shopping (including those not on Commercetools) without replacing their existing commerce stack. Powered by the Commercetools enterprise-grade platform, AgenticLift gives companies a fast, low-friction way to integrate agent-powered discovery, cart building, and checkout flows into their existing systems.

Linnworks launches Spotlight AI to help online retailers automate operations. Linnworks, a connected commerce operations platform, has launched Spotlight AI to help retailers automate repetitive operational tasks and make data-driven decisions. Spotlight AI is available to all Linnworks customers to continuously analyze operational workflows, diagnose inefficiencies, and prescribe automations.

Flexport launches tariff refund calculator. Flexport, a global logistics technology company, has launched a tariff refund calculator to help importers estimate potential refunds and prepare for a possible Supreme Court decision. Flexport’s tariff refund calculator asks businesses to upload their 2025 Entry Report from U.S. Customs and Border Protection, which is available online to U.S. importers. The refund calculator determines the total potential duties eligible for refund and breaks them down by duty category.

Web page for Flexport's tariff refund calculator

Flexport’s tariff refund calculator.

Information Retrieval Part 1: Disambiguation

TL;DR

  1. Disambiguation is the process of resolving ambiguity and uncertainty in data. It’s crucial in modern-day SEO and information retrieval.
  2. Search engines and LLMs reward content that is easy to “understand,” not content that is necessarily best.
  3. The clearer and better structured your content, the harder it is to replace.
  4. You have to reinforce how your brand and products are understood. When grounding is required, models favor sources they recognize from training data

The internet has changed. Channels have begun to homogenize. Google is trying to become something of a destination, and the individual content creator is more powerful than ever.

Oh, and we don’t need to click on anything.

But what makes for great content hasn’t changed. AI and LLMs haven’t changed what people want to consume. They’ve changed what we need to click on. Which I don’t necessarily hate.

As long as you’ve been creating well-structured, engaging, educational/entertaining content for years. All this chat of chunking is a bit smoke and mirrors for me.

“If it walks like a duck and talks like a duck, it’s probably a grifter selling you link building services or GEO.”

However, it is absolutely not all rubbish. Concepts like ambiguity are a more destructive force than ever. If you permit a quick double negative, you cannot not be clear.

The clearer you are. The more concise. The more structured on and off-page. The better chance you stand. There’s no place for ambiguous phrases, paragraphs, and definitions.

This is known as disambiguation.

What Is Disambigation?

Disambiguation is the process of resolving ambiguity and uncertainty in data. Ambiguity is a problem in the modern-day internet. The deeper down the rabbit hole we go, the less diligence is paid towards accuracy and truth. The more clarity your surrounding context provides, the better.

It is a critical component of modern-day SEO, AI, natural language processing (NLP), and information retrieval.

This is an obvious and overused example, but consider a term like apple. The intent and understanding behind it are vague. We don’t know whether people mean the company, the fruit, the daughter of a batshit, brain-dead celebrity.

Image Credit: Harry Clarkson-Bennett

Years ago, this type of ambiguous search would’ve yielded a more diverse set of results. But thanks to personalization and trillions of stored interactions, Google knows what we all want. Scaled user engagement signals and an improved understanding of intent and keywords, phrases, and context are fundamental here.

Yes, I could’ve thought of a better example, but I couldn’t be bothered. You see my point.

Why Should I Care?

Modern-day information retrieval requires clarity. The context you provide really matters when it comes to a confidence score systems require when pulling the “correct” answer.

And this context is not just present in the content.

There is a significant debate about the value of structured data in modern-day search and information retrieval. Using structured data like sameAs to signify exactly who this author is and tying all of your company’s social accounts and sub-brands together can only be a good thing.

The argument isn’t that this has no value. It makes sense.

  • It’s whether Google needs it for accurate information parsing anymore.
  • And whether it has value to LLMs outside of well-structured HTML.

Ambiguity and information retrieval have become incredibly hot topics in data science. Vectorization – representing documents and queries as vectors – helps machines understand the relationships between terms.

It allows models to effectively predict what words should be present in the surrounding context. It’s why answering the most relevant questions and predicting user intent and ‘what’s next’ has been so valuable for a long time in search.

See Google’s Word2Vec for more information.

Google Has Been Doing This For A Long Time

Do you remember what Google’s early, and official, mission statement regarding information was?

“Organize the world’s information and make it universally accessible and useful.”

Their former motto was “don’t be evil.” Which I think in more recent times they may have let slide somewhat. Or conveniently hidden it.

Organizing the world’s information has become so much more effective thanks to advances in information retrieval. Originally, Google thrived on straightforward keyword matching. Then they moved to tokenization.

Their ability to break sentences into words and match short-tail queries was revolutionary. But as queries advanced and intent became less obvious, they had to evolve.

The advent of Google’s Knowledge Graph was transformational. A database of entities that helped create consistency. It created stability and improved accuracy in an ever-changing web.

Image Credit: Harry Clarkson-Bennett

Now queries are rewritten at scale. Ranking is probabilistic instead of deterministic, and in some cases, fan-out processes are applied to create an all-encompassing answer. It’s about matching the user’s intent at the time. It’s personalized. Contextual signals are applied to give the individual the best result for them.

Which means we lose predictability depending on temperature settings, context, and inference path. There’s a lot more passage-level retrieval going on.

Thanks to Dan Petrovic, we know that Google doesn’t use your full page content when grounding its Gemini-powered AI systems. Each query has a fixed grounding budget of approximately 2,000 words total, distributed across sources by relevance rank.

The higher you rank in search, the more budget you are allotted. Think of this context window limit like crawl budget. Larger windows enable longer interactions, but cause performance degradation. So they have to strike a balance.

Position 1 gives you over twice as much “budget” as position 5 (Image Credit: Harry Clarkson-Bennett)

Hummingbird, BERT, RankBrain – Foundational Semantic Understanding

These older algorithm shifts were pivotal in making Google’s systems treat language and meaning differently.

  • Hummingbird (2013) helped Google identify entities and things quickly, with greater precision. This was a step toward semantic interpretation and entity recognition. Think of keywords at a page level. Not query level.
  • RankBrain (2015): To combat the ever-increasing and never-before-seen queries, Google introduced machine learning to interpret unknown queries and relate them to known concepts and entities.

RankBrain was built on the success of Hummingbird’s semantic search. By mastering NLP systems, Google began mapping words to mathematical patterns (vectorization) to better serve new and ever-evolving queries.

These vectors help Google ‘guess’ the intent of queries it has never seen before by finding their nearest mathematical neighbors.

The Knowledge Graph Updates

In July 2023, Google rolled out a major Knowledge Graph update. I think people in SEO called it the Killer Whale Update, but I can’t remember who coined the phrase. Or why. Apologies. It was designed to accelerate the growth of the graph and reduce its dependence on third-party sources like Wikipedia.

As somebody who has spent a long time messing around with entities, I can really understand why. It’s a giant, expensive time-suck.

It explicitly expanded and restructured how entities are recognized and classified in the Knowledge Graph. Particularly, person entities with clear roles such as author or writer.

  • The number of entities in the Knowledge Vault increased by 7.23% in one day to over 54 billion.
  • In July 2023, the number of Person entities tripled in just four days.

All of this is an effort to combat AI slop, provide clarity, and minimize misinformation. To reduce ambiguity and to serve content where a living, breathing expert is at the heart of it.

Worth checking whether you have a presence in the Knowledge Graph here. If you do and can claim a Knowledge Panel, do it. Cement your presence. If not, build your brand and connectedness on the internet.

What About LLMs & AI Search?

There are two main ways LLMs retrieve information:

  • By accessing their vast, static training data.
  • Using RAG (a type of grounding) to access external, up-to-date sources of information.

RAG is why traditional Google Search is still so important. The latest models no longer train on real-time data and lag a little behind. Before the primary model dives in to respond to your desperate need for companionship, a classifier determines whether real-time information retrieval is necessary.

Hence the need for RAG (Image Credit: Harry Clarkson-Bennett)

They cannot know everything and have to employ RAG to make up for their lack of up-to-date information (or verifiable facts through their training data) when retrieving certain answers. Essentially trying to make sure they aren’t chatting rubbish.

Hallucinating if you’re feeling fancy.

So, each model needs its own form of disambiguation. Primarily, this is achieved via:

  • Context-aware query matching. Seeing words as tokens and even reformatting queries into more structured formats to try and achieve the most accurate result. This type of query transformation leads to fan-out and embeddings for more complex queries.
  • RAG architectures. Accessing external knowledge when an accuracy threshold isn’t reached.
  • Conversational agents. LLMs can be prompted to decide whether to directly answer a query or to ask the user for clarification if they don’t meet the same confidence threshold.

Remember, if your content isn’t accessible to search retrieval systems it can’t be used as part of a grounding response. There’s no separation here.

What Should You Do About It?

If you have wanted to do well in search over the last decade, this should’ve been a core part of your thinking. Helpful content rewards clarity.

Allegedly. It also rewards nerfing smaller sites out of existence.

Remember that being clever isn’t better than being clear.

Doesn’t mean you can’t be both. Great content entertains, educates, inspires, and enhances.

Use Your Words

You need to learn how to write. Short, snappy sentences. Help people and machines connect the dots. If you understand the topic, you should know what people want or need to read next almost better than they do.

  • Use verifiable claims.
  • Cite your sources.
  • Showcase your expertise through your understanding.
  • Stand out. Be different. Add information to the corpus to force a mention and/or citation.

Structure The Page Effectively

Write in clear, straightforward paragraphs with a logical heading structure. You really don’t have to call it chunking if you don’t want to. Just make it easy for people and machines to consume your content.

  • Answer the question. Answer it early.
  • Use summaries or hooks.
  • Tables of contents.
  • Tables, lists, and actual structured data. Not schema. But also schema.

Make it easy for users to see what they’re getting and whether this page is right for them.

Intent

Lots of intent is static. Commercial queries always demand some level of comparison. Transactional queries demand some kind of buying or sales process.

But intent changes and millions of new queries crop up every day.

So, you need to monitor the intent of a term or phrase. News is probably a perfect example. Stories break. Develop. What was true yesterday may not be true today. The courts of public opinion damn and praise in equal measure.

Google monitors the consensus. Tracks changes to documents. Monitors authority and – crucially here – relevance.

You can use something like Also Asked to monitor intent changes over time.

The Technical Layer

For years, structured data has helped resolve ambiguity. But we don’t have real clarity over its impact on AI search. Cleaner, well-structured pages are always easier to parse, and entity recognition really matters.

  • sameAs properties connect the dots with your brand and social accounts.
  • It helps you explicitly state who your author is and, crucially, isn’t.
  • Internal linking helps bots navigate across connected sections of your website and build some form of topical authority.
  • Keep content up to date, with consistent date framing – on page, structured data, and sitemaps

If you like messing around with the Knowledge Graph (who the hell doesn’t?), you can find confidence scores for your brand.

According to Google’s very own guidelines, structured data provides explicit clues about a page’s content, helping search engines understand it better.

Yes, yes, it displays rich results etc. But it removes ambiguity.

Entity Matching

I think this ties everything together. Your brand, your products, your authors, your social accounts.

What you say about your brand matters now more than ever.

  • The company you keep (the phrases on a page).
  • The linked accounts.
  • The events you speak at.
  • Your about us page(s).

All of it helps machines build up a clear picture of who you are. If you have strong social profiles, you want to make sure you’re leveraging that trust.

At a page level, title consistency, using relevant entities in your opening paragraph, linking to relevant tags and articles page, and using a rich, relevant author bio is a great start.

Really, just good, solid SEO. Don’t @ me.

PSA: Don’t be boring. You won’t survive.

More Resources:


This post was originally published on Leadership in SEO.


Featured Image: Roman Samborskyi/Shutterstock

Google May Let Sites Opt Out Of AI Search Features via @sejournal, @MattGSouthern

Google says it’s exploring updates that could let websites opt out of AI-powered search features specifically.

The blog post came the same day the UK’s Competition and Markets Authority opened a consultation on potential new requirements for Google Search, including controls for websites to manage their content in Search AI features.

Ron Eden, Principal, Product Management at Google, wrote:

“Building on this framework, and working with the web ecosystem, we’re now exploring updates to our controls to let sites specifically opt out of Search generative AI features.”

Google provided no timeline, technical specifications, or firm commitment. The post frames this as exploration, not a product roadmap.

What’s New

Google currently offers several controls for how content appears in Search, but none cleanly separate AI features from traditional results.

Google-Extended lets publishers block their content from training Gemini and Vertex AI models. But Google’s documentation states Google-Extended doesn’t impact inclusion in Google Search and isn’t a ranking signal. It controls AI training, not AI Overviews appearance.

The nosnippet and max-snippet directives do apply to AI Overviews and AI Mode. But they also affect traditional snippets in regular search results. Publishers wanting to limit AI feature exposure currently lose snippet visibility everywhere.

Google’s post acknowledges this gap exists. Eden wrote:

“Any new controls need to avoid breaking Search in a way that leads to a fragmented or confusing experience for people.”

Why This Matters

I wrote in SEJ’s SEO Trends 2026 ebook that people would have more influence on the direction of search than platforms do. Google’s post suggests that dynamic is playing out.

Publishers and regulators have spent the past year pushing back on AI Overviews. The UK’s Independent Publishers Alliance, Foxglove, and Movement for an Open Web filed a complaint with the CMA last July, asking for the ability to opt out of AI summaries without being removed from search entirely. The US Department of Justice and South African Competition Commission have proposed similar measures.

The BuzzStream study we covered earlier this month found 79% of top news publishers block at least one AI training bot, and 71% block retrieval bots that affect AI citations. Publishers are already voting with their robots.txt files.

Google’s post suggests it’s responding to pressure from the ecosystem by exploring controls it previously didn’t offer.

Looking Ahead

Google’s language is cautious. “Exploring” and “working with the web ecosystem” are not product commitments.

The CMA consultation will gather input on potential requirements. Regulatory processes move slowly, but they do produce outcomes. The EU’s Digital Markets Act investigations have already pushed Google to make changes in Europe.

For now, publishers wanting to limit AI feature exposure can use nosnippet or max-snippet directives, but note that these affect traditional snippets as well. Google’s robots meta tag documentation covers the current options.

If Google follows through on specific opt-out controls, the technical implementation will matter. Whether it’s a new robots directive, a Search Console setting, or something else will determine how practical it is for publishers to use.


Featured Image: ANDRANIK HAKOBYAN/Shutterstock

What If User Satisfaction Is The Most Important Factor In SEO? via @sejournal, @marie_haynes

Let me see if I can convince you!

I’ve shared a bunch in this video and summarized my thoughts in the article below. Also, this is the second blog post I’ve written on this topic in the last week. There is much more information on user data and how Google uses it in my previous blog post.

Ranking Has 3 Components

We learned in the DOJ vs Google trial that Google’s ranking process involves three main components:

  1. Traditional systems are used for initial ranking.
  2. AI Systems (such as RankBrain, DeepRank, and RankEmbed BERT) re-rank the top 20-30 documents.
  3. Those systems are fine-tuned by Quality Rater scores, and more importantly IMO, results from live user tests.

The DOJ vs. Google lawsuit talked extensively about how Google’s massive advantage stems from the large amounts of user data it uses. In its appeal, Google said that it does not want to comply with the judge’s mandate to hand over user data to competitors. It listed two ways it uses user data – in a system called Glue, a system which incorporates Navboost that looks at what users click on and engage with, and also in the RankEmbed model.

RankEmbed is fascinating. It embeds the user’s query into a vector space. Content that is likely to be relevant to that query will be found nearby. RankEmbed is fine-tuned by two things:

1. Ratings from the Quality Raters. They are given two sets of results – “Frozen” Google results and “Retrained” results – or, in other words, the results of the newly trained and refined AI-driven search algorithms. Their scores help Google’s systems understand whether the retrained algorithms are producing higher-quality search results.

From Douglas Oard’s testimony re: Frozen and Retrained Google

2. Real-world live experiments where a small percentage of real searchers are shown results from the old vs. retrained algorithms. Their clicks and actions help fine-tune the system.

The ultimate goal of these systems is to continually improve on producing rankings that satisfy the searcher.

More Thinking On Live Tests – Users Tell Google The Types Of Pages That Are Helpful, Not The Actual Pages

I’ve realized that Google’s live user tests aren’t just about gathering data on specific pages. They are about training the system to recognize patterns. Google isn’t necessarily tracking every single user interaction to rank that one specific URL. Instead, it is using that data to teach its AI what “helpful” looks like. The system learns to identify the types of content that satisfy user intent, then predicts whether your site fits that successful mold.

It will continue to evolve its process in predicting which content is likely to be helpful. It definitely extends far beyond simple vector search. Google is continually finding new ways to understand user intent and how to meet it.

What This Means For SEO

If you’re ranking in the top few pages of search, you have convinced the traditional ranking systems to put you in the ranking auction.

Once there, a multitude of AI systems work to predict which of the top results truly is the best for the searcher. This is even more important now that Google is starting to use “Personal Intelligence” in Gemini and AI Mode. My top search results will be tailored specifically for what Google’s systems think I will find helpful.

Once you start understanding how AI systems do search, which is primarily vector search, it can be tempting to work to reverse engineer these. If you’re optimizing by using a deep understanding of what vector search rewards (including using cosine similarity), you’re working to look good to the AI systems. I’d caution against diving in too deeply here.

Image Credit: Marie Haynes

Given that the systems are fine-tuned to continually improve upon producing results that are the most satisfying for the searcher, looking good to AI is nowhere near as important as truly being the result that is the most helpful. I would argue that optimizing for vector search can do more harm than good unless you truly do have the type of content that users go on to find more helpful than the other options they have. Otherwise, there’s a good chance you’re training the AI systems to not favor you.

Image Credit: Marie Haynes

My Advice

My advice is to optimize loosely for vector search. What I mean by this is to not obsess over keywords and cosine similarity, but instead to understand what it is your audience wants and be sure that your pages meet the specific needs they have. Is using a knowledge of Google’s Query Fan-Out helpful here? To some degree, yes, as it is helpful to know what questions users generally tend to have surrounding a query. But, I think that my same fears apply here as well. If you look really good to the AI systems trying to find content to satisfy the query fan-out, but users don’t tend to agree, or if you’re lacking other characteristics associated with helpfulness compared to competitors, you might train Google’s systems to favor you less.

Make use of headings – not for the AI systems to see, but to help your readers understand that the things they are looking for are on your page.

Look at the pages that Google is ranking for queries that should lead to your page, and truly ask yourself what it is about these pages that searchers are finding helpful. Look at how well they answer specific questions, whether they use good imagery, tables, or other graphics, and how easy it is for the page to be skimmed and navigated. Work to figure out why this page was chosen as among the most likely to be helpful in satisfying the needs of searchers.

Instead of obsessing over keywords, work to improve the actual user experience. If you make your page more engaging, focusing more on metrics like scrolls and session duration, rankings should naturally improve.

And mostly, obsess over helpfulness. It can be helpful to have an external party look at your content and share why it may or may not be helpful.

I have found that even though I have this understanding that search is built to continually learn and improve upon showing searchers pages they are likely to find helpful, I still find myself fighting the urge to optimize for machines rather than users. It is a hard habit to break! Given that Google’s deep learning systems are working tirelessly on one goal – predicting which pages are likely to be helpful to the searcher – that should be our goal as well. As Google’s documentation on helpful content suggests, the type of content that people tend to find helpful is content that is original, insightful, and provides substantial value when compared to other pages in the search results.

More Resources:


This post was originally published on Marie Haynes Consulting.


Featured Image: Chayanit/Shutterstock

Social Channel Insights In Search Console: What It Means For Social & Search via @sejournal, @rio_seo

Google has been testing Social Channel Insights inside Google Search Console (GSC). This update may appear small, but it’s more than meets the eye. In the search landscape, these new social insights translate to a bigger shift happening behind the scenes, where search and social data converge to improve visibility.

The official announcement from Google highlighted the growth of businesses managing their digital presence on popular social media sites. The integration makes sense as social media continues to become a popular method for search discovery and information, with 15% of consumers believing social media to be the most accurate/current source to find up-to-date business details.

The expansion of the social report feature showcases performance for accounts Google associates with a website, allowing businesses a centralized location for reviewing key search and discoverability metrics. This update signifies just how intertwined search and social are becoming. Search and social should no longer be treated as disparate functions, but rather integral counterparts that must communicate and coordinate to improve online visibility and discovery.

A Closer Look At Google’s Social Channel Insights Test

When digging into Google Search Console Insights to ascertain what exactly these new social metrics entail, we see a plethora of new information has been added. It appears as though this feature isn’t readily available to all, but is only showing up for some websites where Google was able to locate their social media channels. Of those who have seen the new social media report features, they’re seeing:

  • The total reach from Google to your social channels.
  • Social media content performance.
  • Queries drive traffic to your social channels.
  • Trends such as high average duration or post growth.

Right now, it appears as though the social media metrics measured focus mostly on referral insights. This isn’t merely a slight tweak to the user experience. It could be seen as a strategic convergence of data, meant to shine a spotlight on how social goes hand in hand with search performance.

Does This Mean Social Is Having More Influence?

Google doesn’t typically make updates for fun or convenience. Each update is a signal for what they plan to evaluate next as part of their never-ending quest to maintain dominancy in the search engine landscape.

Even though Google hasn’t explicitly stated that social engagement metrics have direct influence, this could be an acknowledgement that discovery is increasingly happening on other channels, such as AI platforms and social media.

Search has fractured with other players joining the race, and Google is clearly noticing and adapting. In fact, a study found nearly a quarter (24%) of U.S. adults use social media as their primary search method, while another 24% use search primarily but also social media occasionally. 78% of global internet users leverage social media to research brands and products, and over 60% of Gen Z consumers have purchased a product they’ve found on social media.

Search engines are no longer the sole place consumers start their sales journey. Users use AI to research and ask questions, or seek out online reviews and testimonies on social media channels. Search engines are becoming more of a validation layer, where users go after they research all the options to confirm information, or seek additional information, and then move to the transaction stage.

How Social Channel Insights Could Impact Social Campaigns

When it comes to social, evaluating performance in the past may have looked like chasing more likes and comments. Engagement, of course, still matters, but Google is telling us what other insights we should consider, right inside your GSC dashboard.

Social referral insights give social media marketers visibility into how their content performs in the search discovery journey. Writing social posts to meet an arbitrary number or goal isn’t the end game. It’s about finding the posts that have the influence.

For social campaigns, social insights can help you:

  • Identify which social content themes generate downstream search demand.
  • Use query-level insights to inform what you write and the message you want to get across.
  • Highlighting social’s distinct role in discovery, not just engaging passive viewers.
  • Coordinate more seamlessly with SEO teams in terms of campaign launches and promotions to capitalize on growing demand.
  • Empower marketers to create content that resonates and aligns with what users are likely to search for next, keeping you one step ahead of the game.

Instead of considering traditional social media metrics (such as comments, shares, or likes), social teams can use these new Social Channel Insights in GSC to increase online visibility.

What Social Signals We’d Like To See Google Include Next

Google, if you’re reading this, here’s what we’d love to see beyond referral behavior to help marketers provide even more strategic value.

Social insights that could meaningfully support discovery-focused strategies include:

  • Content velocity indicators: Show us how quickly topics gain traction on social before search demand spikes.
  • Content format indicators: Tell us what content formats perform best for winning search discovery, whether that be short-form videos or static posts.
  • Topic momentum indicators: Help us understand emerging themes gaining attention across platforms.
  • Creator and brand association indicators: Give us more transparency around which entities are consistently driving early discovery for certain topics.
  • Cross-platform trend alignment indicators: Reveal when multiple social ecosystems signal rising interest at the same time. This helps us strike the iron when it’s hot.

By adding the aforementioned signals, SEOs would be able to anticipate intent shifts earlier and inform content and social teams to draft meaningful and relevant content right away, not after the hype dies down. It’s a win for all teams as your time investment will lead to actual results.

What Marketers Should Do Now

Even though this is a limited test and hasn’t impacted every business (yet),  it would be a good idea for marketers to review their social media channels and strategy to provide an exceptional experience across every channel customers find you.

To prepare, marketers should:

  • Audit which pages receive the most social-driven search traffic. These insights will inform which types of content and topics attract social search visitors most.
  • Align content calendars across social and SEO teams. Start breaking silos between teams by enabling transparency across cross-department initiatives, such as the content calendar. By doing so, you’ll better create a culture of collaboration and give teams shared KPIs to work toward.
  • Repurpose high-performing social content into search-optimized formats (and vice versa). For example, social videos that are performing well in search can be embedded into relevant blog posts, helping you get more value and longevity out of the content you work hard to create. Another example would be user-generated content repurposed into frequently asked questions.
  • Track emerging social trends. Social platforms like TikTok and Instagram can serve as search indicators, allowing marketers to anticipate what consumers are interested in most and what’s capturing their attention.
  • Integrate hybrid analytics into your measurement tracking. AI is having an impact on marketing; however, humans still play a key role in any and every marketing endeavor. Machine-driven insights may give us data at our fingertips; however, human interpretation and validation are still a must. Only humans have the power and foresight to assess nuance, emotions, and insider knowledge, far better than any machine ever could.

Next Steps To Take With Social Channel Insights

Google’s rollout of Social Channel Insights in GSC may seem like a minor advancement, but it’s more than just additional metrics to track for marketers. It signifies how Google is considering how the two disciplines share insights.

Search engines are factoring in the rise of discovery and influence taking place on social media channels. By bridging the gap, and working more closely together, social media marketers and SEOs should see each other as partners rather than once in a while collaborators. The result? Better workflows, collaboration, visibility, and business impact.

Marketers who embrace a cross-collaboration mentality with SEOs will be better poised to appear in the moments that matter, being discovered and chosen.

More Resources:


Featured Image: MR.DEEN/Shutterstock

New Yahoo Scout AI Search Delivers The Classic Search Flavor People Miss via @sejournal, @martinibuster

Yahoo has announced Yahoo Scout, a new AI-powered answer engine now available in beta to users in the United States, providing a clean Classic Search experience with the power of personalized AI. The launch also includes the Yahoo Scout Intelligence Platform, which brings AI features across Yahoo’s core products, including Mail, News, Finance, and Sports.

Screenshot Of Yahoo Scout

Yahoo’s Existing Products and User Reach

Yahoo’s announcement states that it operates some of the most popular websites and services in the United States, reaching what they say is 90% of all internet users in the United States (based on Comscore data), through its email, news, finance, and sports properties. The company says that Yahoo Scout builds on the foundation of decades of search behavior and user interaction data.

How Yahoo Scout Generates Answers

Yahoo has partnered with Anthropic to use the Claude model as the primary AI system behind Yahoo Scout. Yahoo’s announcement said it selected Claude for speed, clarity, judgment, and safety, which it described as essential qualities for a consumer-facing answer engine. Yahoo also continues its partnership with Microsoft by using Microsoft Bing’s grounding API, which connects AI-generated answers to information from across the open web. Yahoo said this approach ensures that answers are informed by authoritative sources rather than unsupported text generation.

According to Yahoo, Scout relies on a combination of traditional web search and generative AI to produce answers that are grounded using Microsoft Bing’s grounding API and informed by sources from across the open web.

According to  Yahoo:

“It’s informed by 500 million user profiles, a knowledge graph spanning more than 1 billion entities, and 18 trillion consumer events that occur annually across Yahoo, which allow Yahoo Scout to provide effective and personalized answers and suggested actions.”

Yahoo’s announcement says that this data, its use of Claude, and reliance on Bing for grounding work together to provide responses to answers that are personalized and helpful for researching and making decisions in the “moments that matter” to people.

They explain:

“Yahoo Scout continues Yahoo’s focus on the moments that matter to people’s daily lives, such as understanding upcoming weather patterns before a vacation, getting details about an important game, tracking stock price movements after earnings, comparing products before buying, or fact-checking a news story.”

Where Yahoo Scout Appears Inside Yahoo Products

The Yahoo Scout Intelligence Platform embeds these AI capabilities directly into Yahoo’s existing services.

For example:

  • In Yahoo Mail, Scout supports AI-generated message summaries.
  • In Yahoo Sports, it produces game breakdowns.
  • In Yahoo News, it surfaces key takeaways.
  • In Yahoo Finance, Scout adds interactive tools for analysis that allow readers to explore market news and stock performance context through AI-powered questions.

According to Eric Feng, Senior Vice President and General Manager of Yahoo Research Group:

“Yahoo’s deep knowledge base, 30 years in the making, allows us to deliver guidance that our users can trust and easily understand, and will become even more personalized over the coming months. Yahoo Scout now powers a new generation of intelligence experiences across Yahoo, seamlessly integrated into the products people use every day.”

What Yahoo Says Comes Next

Yahoo said Scout will continue to develop over the coming months. Planned updates include deeper personalization, expanded capabilities within specific verticals, and new formats for search advertising designed to work in generative AI search. The company did not provide a timeline for when the beta period will end or when additional features will move beyond testing.

Yahoo explained:

“Yahoo Scout will continue to evolve in the months ahead, expanding to power new products across Yahoo. In particular, the new answer engine will become more personalized, will add new capabilities focused on deeper experiences within key verticals, and will introduce new, improved opportunities for search advertisers to effectively cross the chasm to generative AI search advertising. “

Yahoo’s Search Experience

Something that’s notable about Yahoo’s AI answer engine experience is how clean and straightforward it is. It’s like a throwback to classic search but with the sophistication of AI answers.

For example, I asked it to give me information on where I can buy an esoteric version of a Levi’s trucker jacket in a specific color (Midnight Harvest) and it presented a clean summary of where to get it, a table with a list of retailers ordered by the lowest prices.

Screenshot Of Yahoo Scout

Notice that there are no product images? It’s just giving me the prices. I don’t know if that’s because they don’t have a product feed but I already know what the jacket looks like in the color I specified so images aren’t really necessary.  This is what I mean when I say that Yahoo Scout offers that Classic Search flavor without the busy overly fussy search experience that Google has been providing lately.

With Yahoo Scout, the company is applying AI systems to tasks its users perform when they search for, read, or compare information online. Rather than positioning AI as a replacement for search or content platforms, Yahoo is using it as a tool that organizes, summarizes, and explains information in a clean and easy to read format.

Yahoo Scout is easy to like because it delivers the clean and uncluttered search experience that many people miss.

Check out Yahoo Scout at scout.yahoo.com

The Yahoo Scout app is available for Android and Apple devices.

The Download: OpenAI’s plans for science, and chatbot age verification

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

Inside OpenAI’s big play for science 

—Will Douglas Heaven

In the three years since ChatGPT’s explosive debut, OpenAI’s technology has upended a remarkable range of everyday activities at home, at work, and in schools.

Now OpenAI is making an explicit play for scientists. In October, the firm announced that it had launched a whole new team, called OpenAI for Science, dedicated to exploring how its large language models could help scientists and tweaking its tools to support them.

So why now? How does a push into science fit with OpenAI’s wider mission? And what exactly is the firm hoping to achieve? I put these questions to Kevin Weil, a vice president at OpenAI who leads the new OpenAI for Science team, in an exclusive interview. Read the full story.

Why chatbots are starting to check your age

How do tech companies check if their users are kids?

This question has taken on new urgency recently thanks to growing concern about the dangers that can arise when children talk to AI chatbots. For years Big Tech asked for birthdays (that one could make up) to avoid violating child privacy laws, but they weren’t required to moderate content accordingly.

Now, two developments over the last week show how quickly things are changing in the US and how this issue is becoming a new battleground, even among parents and child-safety advocates. Read the full story.

—James O’Donnell

This story originally appeared in The Algorithm, our weekly newsletter on AI. To get stories like this in your inbox first, sign up here.

TR10: Commercial space stations

Humans have long dreamed of living among the stars, and for two decades hundreds of us have done so aboard the International Space Station (ISS). But a new era is about to begin in which private companies operate orbital outposts—with the promise of much greater access to space than before.

The ISS is aging and is expected to be brought down from orbit into the ocean in 2031. To replace it, NASA has awarded more than $500 million to several companies to develop private space stations, while others have built versions on their own. Read why we made them one of our 10 Breakthrough Technologies this year, and check out the rest of the list.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 Tech workers are pressuring their bosses to condemn ICE 
The biggest companies and their leaders have remained largely silent so far. (Axios)
+ Hundreds of employees have signed an anti-ICE letter. (NYT $)
+ Formerly politically-neutral online spaces have become battlegrounds. (WP $)

2 The US Department of Transport plans to use AI to write new safety rules
Please don’t do this. (ProPublica)
+ Failure to catch any errors could lead to civilian deaths. (Ars Technica)

3 The FBI is investigating Minnesota Signal chats tracking federal agents
But free speech advocates claim the information is legally obtained. (NBC News)
+ A judge has ordered a briefing on whether Minnesota is being illegally punished. (Wired $)

4 TikTok users claim they’re unable to send “Epstein” in direct messages
But the company says it doesn’t know why. (NPR)
+ Users are also experiencing difficulty uploading anti-ICE videos. (CNN)
+ TikTok’s first weekend under US ownership hasn’t gone well. (The Verge)
+ Gavin Newsom wants to probe whether TikTok is censoring Trump-critical content. (Politico)

5 Grok is not safe for children or teens
That’s the finding of a new report digging into the chatbot’s safety measures. (TechCrunch)
+ The EU is investigating whether it disseminates illegal content, too. (Reuters)

6 The US is on the verge of losing its measles-free status
Following a year of extensive outbreaks. (Undark)
+ Measles is surging in the US. Wastewater tracking could help. (MIT Technology Review)

7 Georgia has become the latest US state to consider banning data centers
Joining Maryland and Oklahoma’s stance. (The Guardian)
+ Data centers are amazing. Everyone hates them. (MIT Technology Review)

8 The future of Saudi Arabia’s futuristic city is in peril
The Line was supposed to house 9 million people. Instead, it could become a data center hub. (FT $)
+ We got an exclusive first look at it back in 2022. (MIT Technology Review)

9 Where do Earth’s lighter elements go? 🌍
New research suggests they might be hiding deep inside its core. (Knowable Magazine)

10 AI-generated influencers are getting increasingly surreal
Featuring virtual conjoined twins, and triple-breasted women. (404 Media)
+ Why ‘nudifying’ tech is getting steadily more dangerous. (Wired $)

Quote of the day

“Humanity is about to be handed almost unimaginable power, and it is deeply unclear whether our social, political, and technological systems possess the maturity to wield it.”

—Anthropic CEO Dario Amodei sounds the alarm about what he sees as the imminent dangers of AI superintelligence in a new 38-page essay, Axios reports.

One more thing

Why one developer won’t quit fighting to connect the US’s grids

Michael Skelly hasn’t learned to take no for an answer. For much of the last 15 years, the energy entrepreneur has worked to develop long-haul transmission lines to carry wind power across the Great Plains, Midwest, and Southwest. But so far, he has little to show for the effort.

Skelly has long argued that building such lines and linking together the nation’s grids would accelerate the shift from coal- and natural-gas-fueled power plants to the renewables needed to cut the pollution driving climate change. But his previous business shut down in 2019, after halting two of its projects and selling off interests in three more.

Skelly contends he was early, not wrong. And he has a point: markets and policymakers are increasingly coming around to his perspective. Read the full story.

—James Temple

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or skeet ’em at me.)

+ Cats on the cover of the New Yorker! Need I say more?
+ Here’s how to know when you truly love someone.
+ This orphaned baby seal is just too cute.
+ I always had a sneaky suspicion that Depeche Mode and the Cure make for perfect bedfellows.

Stratospheric internet could finally start taking off this year

Today, an estimated 2.2 billion people still have either limited or no access to the internet, largely because they live in remote places. But that number could drop this year, thanks to tests of stratospheric airships, uncrewed aircraft, and other high-altitude platforms for internet delivery. 

Even with nearly 10,000 active Starlink satellites in orbit and the OneWeb constellation of 650 satellites, solid internet coverage is not a given across vast swathes of the planet. 

One of the most prominent efforts to plug the connectivity gap was Google X’s Loon project. Launched in 2011, it aimed to deliver access using high-altitude balloons stationed above predetermined spots on Earth. But the project faced literal headwinds—the Loons kept drifting away and new ones had to be released constantly, making the venture economically unfeasible. 

Although Google shuttered the high-profile Loon in 2021, work on other kinds of high-altitude platform stations (HAPS) has continued behind the scenes. Now, several companies claim they have solved Loon’s problems with different designs—in particular, steerable airships and fixed-wing UAVs (unmanned aerial vehicles)—and are getting ready to prove the tech’s internet beaming potential starting this year, in tests above Japan and Indonesia.

Regulators, too, seem to be thinking seriously about HAPS. In mid-December, for example, the US Federal Aviation Administration released a 50-page document outlining how large numbers of HAPS could be integrated into American airspace. According to the US Census Bureau’s 2024 American Community Survey (ACS) data, some 8 million US households (4.5% of the population) still live completely offline, and HAPS proponents think the technology might get them connected more cheaply than alternatives.

Despite the optimism of the companies involved, though, some analysts remain cautious.

“The HAPS market has been really slow and challenging to develop,” says Dallas Kasaboski, a space industry analyst at the consultancy Analysis Mason. After all, Kasaboski says, the approach has struggled before: “A few companies were very interested in it, very ambitious about it, and then it just didn’t happen.”

Beaming down connections

Hovering in the thin air at altitudes above 12 miles, HAPS have a unique vantage point to beam down low-latency, high-speed connectivity directly to smartphone users in places too remote and too sparsely populated to justify the cost of laying fiber-optic cables or building ground-based cellular base stations.

“Mobile network operators have some commitment to provide coverage, but they frequently prefer to pay a fine than cover these remote areas,” says Pierre-Antoine Aubourg, chief technology officer of Aalto HAPS, a spinoff from the European aerospace manufacturer Airbus. “With HAPS, we make this remote connectivity case profitable.” 

Aalto HAPS has built a solar-powered UAV with a 25-meter wingspan that has conducted many long-duration test flights in recent years. In April 2025 the craft, called Zephyr, broke a HAPS record by staying afloat for 67 consecutive days. The first months of 2026 will be busy for the company, according to Aubourg; Zephyr will do a test run over southern Japan to trial connectivity delivery to residents of some of the country’s smallest and most poorly connected inhabited islands.

the Zephyr on the runway at sunrise

AALTO

Because of its unique geography, Japan is a perfect test bed for HAPS. Many of the country’s roughly 430 inhabited islands are remote, mountainous, and sparsely populated, making them too costly to connect with terrestrial cell towers. Aalto HAPS is partnering with Japan’s largest mobile network operators, NTT DOCOMO and the telecom satellite operator Space Compass, which want to use Zephyr as part of next-generation telecommunication infrastructure.

“Non-terrestrial networks have the potential to transform Japan’s communications ecosystem, addressing access to connectivity in hard-to-reach areas while supporting our country’s response to emergencies,” Shigehiro Hori, co-CEO of Space Compass, said in a statement

Zephyr, Aubourg explains, will function like another cell tower in the NTT DOCOMO network, only it will be located well above the planet instead of on its surface. It will beam high-speed 5G connectivity to smartphone users without the need for the specialized terminals that are usually required to receive satellite internet. “For the user on the ground, there is no difference when they switch from the terrestrial network to the HAPS network,” Aubourg says. “It’s exactly the same frequency and the same network.”

New Mexico–based Sceye, which has developed a solar-powered helium-filled airship, is also eyeing Japan for pre-commercial trials of its stratospheric connectivity service this year. The firm, which extensively tested its slick 65-meter-long vehicle in 2025, is working with the Japanese telecommunications giant SoftBank. Just like NTT DOCOMO, Softbank is betting on HAPS to take its networks to another level. 

Mikkel Frandsen, Sceye’s founder and CEO, says that his firm succeeded where Loon failed by betting on the advantages offered by the more controllable airship shape, intelligent avionics, and innovative batteries that can power an electric fan to keep the aircraft in place.

“Google’s Loon was groundbreaking, but they used a balloon form factor, and despite advanced algorithms—and the ability to change altitude to find desired wind directions and wind speeds—Loon’s system relied on favorable winds to stay over a target area, resulting in unpredictable station-seeking performance,” Frandsen says. “This required a large amount of balloons in the air to have relative certainty that one would stay over the area of operation, which was financially unviable.”

He adds that Sceye’s airship can “point into the wind” and more effectively maintain its position. 

“We have significant surface area, providing enough physical space to lift 250-plus kilograms and host solar panels and batteries,” he says, “allowing Sceye to maintain power through day-night cycles, and therefore staying over an area of operation while maintaining altitude.” 

The persistent digital divide

Satellite internet currently comes at a price tag that can be too high for people in developing countries, says Kasaboski. For example, Starlink subscriptions start at $10 per month in Africa, but millions of people in these regions are surviving on a mere $2 a day.

Frandsen and Aubourg both claim that HAPS can connect the world’s unconnected more cheaply. Because satellites in low Earth orbit circle the planet at very high speeds, they quickly disappear from a ground terminal’s view, meaning large quantities of those satellites are needed to provide continuous coverage. HAPS can hover, affording a constant view of a region, and more HAPS can be launched to meet higher demand.

“If you want to deliver connectivity with a low-Earth-orbit constellation into one place, you still need a complete constellation,” says Aubourg. “We can deliver connectivity with one aircraft to one location. And then we can tailor much more the size of the fleet according to the market coverage that we need.”

Starlink gets a lot of attention, but satellite internet has some major drawbacks, says Frandsen. A big one is that its bandwidth gets diluted once the number of users in an area grows. 

In a recent interview, Starlink cofounder Elon Musk compared the Starlink beams to a flashlight. Given the distance at which those satellites orbit the planet, the cone is wide, covering a large area. That’s okay when users are few and far between, but it can become a problem with higher densities of users.

For example, Ukrainian defense technologists have said that Starlink bandwidth can drop on the front line to a mere 10 megabits per second, compared with the peak offering of 220 Mbps when drones and ground robots are in heavy use. Users in Indonesia, which like Japan is an island nation, also began reporting problems with Starlink shortly after the service was introduced in the country in 2024. Again, bandwidth declined as the number of subscribers grew.

In fact, Frandsen says, Starlink’s performance is less than optimal once the number of users exceeds one person per square kilometer. And that can happen almost anywhere—even relatively isolated island communities can have hundreds or thousands of residents in a small area. “There is a relationship between the altitude and the population you can serve,” Frandsen says. “You can’t bring space closer to the surface of the planet. So the telco companies want to use the stratosphere so that they can get out to more rural populations than they could otherwise serve.” Starlink did not respond to our queries about these challenges. 

Cheaper and faster

Sceye and Aalto HAPS see their stratospheric vehicles as part of integrated telecom networks that include both terrestrial cell towers and satellites. But they’re far from the only game in town. 

World Mobile, a telecommunications company headquartered in London, thinks its hydrogen-powered high-altitude UAV can compete directly with satellite mega-constellations. The company acquired the HAPS developer Stratospheric Platforms last year. This year, it plans to flight-test an innovative phased array antenna, which it claims will be able to deliver bandwidth of 200 megabits per second (enough to enable ultra-HD video streaming to 500,000 users at the same time over an area of 15,000 square kilometers—equivalent to the coverage of more than 500 terrestrial cell towers, the company says). 

Last year, World Mobile also signed a partnership with the Indonesian telecom operator Protelindo to build a prototype Stratomast aircraft, with tests scheduled to begin in late 2027.

Richard Deakin, CEO of World Mobile’s HAPS division World Mobile Stratospheric, says that just nine Stratomasts could supply Scotland’s 5.5 million residents with high-speed internet connectivity at a cost of £40 million ($54 million) per year. That’s equivalent to about 60 pence (80 cents) per person per month, he says. Starlink subscriptions in the UK, of which Scotland is a part, come at £75 ($100) per month.

A troubled past 

Companies working on HAPS also extol the convenience of prompt deployments in areas struck by war or natural disasters like Hurricane Maria in Puerto Rico, after which Loon played an important role. And they say that HAPS could make it possible for smaller nations to obtain complete control over their celestial internet-beaming infrastructure rather than relying on mega-constellations controlled by larger nations, a major boon at a time of rising geopolitical tensions and crumbling political alliances. 

Analysts, however, remain cautious, projecting a HAPS market totaling a modest $1.9 billion by 2033. The satellite internet industry, on the other hand, is expected to be worth $33.44 billion by 2030, according to some estimates. 

The use of HAPS for internet delivery to remote locations has been explored since the 1990s, about as long as the concept of low-Earth-orbit mega-constellations. The seemingly more cost-effective stratospheric technology, however, lost to the space fleets thanks to the falling cost of space launches and ambitious investment by Musk’s SpaceX. 

Google wasn’t the only tech giant to explore the HAPS idea. Facebook also had a project, called Aquila, that was discontinued after it too faced technical difficulties. Although the current cohort of HAPS makers claim they have solved the challenges that killed their predecessors, Kasaboski warns that they’re playing a different game: catching up with now-established internet-beaming mega constellations. By the end of this year, it’ll be much clearer whether they stand a good chance of doing so.