OpenAI ChatGPT Is Testing A Memory Feature via @sejournal, @martinibuster

OpenAI announced that it is rolling out a test of ChatGPT memory, a new feature that remembers past conversations to learn a users preferences, remembers styles and tones and format preferences.

ChatGPT gradually becomes better with time as it remembers across all chats and doing away with having to repeat instructions.

The new feature is rolling out as a limited release to some ChatGPT free and Plus users for testing and feedback. There are plans to extend the feature to more users at a future time.

Memories are saved apart from the chat histories, which is how they can be summoned in later chat sessions.

Users Control Memory

The memory feature can be selectively turned on by telling it to remember something and it can also be told to forget specific details or instructions. Memory is controlled through the settings panel, where specific memories can be deleted or disabled entirely. There is also an option to delete all memories entirely.

The settings options are available at Settings > Personalization > Memory.

How To Use ChatGPT Memory

OpenAI shared examples of how to use the feature:

  • “You’ve explained that you prefer meeting notes to have headlines, bullets and action items summarized at the bottom. ChatGPT remembers this and recaps meetings this way.
  • You’ve told ChatGPT you own a neighborhood coffee shop. When brainstorming messaging for a social post celebrating a new location, ChatGPT knows where to start.
  • You mention that you have a toddler and that she loves jellyfish. When you ask ChatGPT to help create her birthday card, it suggests a jellyfish wearing a party hat.
  • As a kindergarten teacher with 25 students, you prefer 50-minute lessons with follow-up activities. ChatGPT remembers this when helping you create lesson plans.”

Memory May Be Used For Training

OpenAI advised that content provided to ChatGPT could be used for training, which includes the memories, with an option in Data Controls where this can be turned off. By default, no data is used for training. But this does not apply to ChatGPT Team and Enterprise customers.

Another way to control ChatGPT memory is through temporary chats, which will not invoke the memory feature. Temporary chats are activated in the GPT version control at the top of the chat page.

ChatGPT Memory Toggle

According to an OpenAI temporary chat explainer:

“With Temporary Chat, you can have a conversation with a blank slate. ChatGPT won’t be aware of previous conversations or access memories. It will still follow your custom instructions if they’re enabled.”

Read the OpenAI announcement:

Memory and new controls for ChatGPT

Featured Image by Shutterstock/rafapress

Google AdSense Shifts To eCPM Payment Model via @sejournal, @MattGSouthern

Google is transitioning AdSense to an eCPM payment model for publishers.

Last November, Google stated it would be updating the revenue share structure for AdSense to modernize the way publishers earn money from their website content.

Ginny Marvin, the Google Ads Liaison, has confirmed this transition to eCPM payments for AdSense partners.

A Closer Look At The Updated Revenue Share

Previously, publishers received a consistent 68% of ad revenue. Now, the revenue share is split into separate rates for the buy-side (advertisers) and sell-side (publishers).

Google elaborates on the specifics of this new structure:

“For displaying ads with AdSense for content, publishers will receive 80% of the revenue after the advertiser platform takes its fee, whether that be Google’s buy-side or third-party platforms.”

According to an example from Google, when Google Ads purchases display ads on AdSense, it retains an average of 15% of advertiser spend. Google states that overall publisher revenue is expected to remain around 68% despite these changes.

The model differs when third-party platforms buy AdSense display ads. In these cases, publishers receive an 80% share after the third-party’s fees. Google says it lacks control or visibility into these third-party fees.

The example below shows that with both buy and sell-side fees, one dollar from an advertiser translates to sixty-eight cents for the publisher.

Google AdSense Shifts To eCPM Payment ModelScreenshot from: blog.google/products/adsense/evolving-how-publishers-monetize-with-adsense/, February 20204.

Transitioning To Per-Impression Payments

Google is transitioning AdSense to a per-impression payment model, bringing it in line with industry standards for display advertising. This will allow publishers to compare earnings across Google’s products and third-party platforms more easily.

According to Google, this payment model update will not impact the amount or type of ads publishers can show as long as they adhere to existing AdSense policies and Better Ads Standards. These standards prevent intrusive ads like pop-ups or advertisements that take over the screen.

Takeaways For Publishers

Publishers who rely on AdSense for a portion of their income may be considering the implications of these changes.

Here are some points of consideration.

Understand the Implications

The eCPM (effective cost per thousand impressions) payment model differs from the previous predominant cost-per-click (CPC) model.

With eCPM, publisher revenue is based on the number of impressions rather than clicks.

Publishers should understand how this new model works, as it could impact revenue, especially for those whose content aims for high engagement over high traffic volume.

Adapt Content & SEO Strategies

Google has stated that the earnings for most publishers will likely remain unchanged after the transition to eCPM bidding.

However, the impact may differ on an individual basis. Publishers may need to adjust their content and SEO tactics to maximize revenue within the new eCPM model.

Potential strategies include increasing website traffic volume, improving user engagement metrics, and extending session duration to serve more ad impressions.

Compliance with Ad Standards

The shift to an impression-based model increases the need for publishers to follow AdSense policies and Better Ads Standards.

Publishers must continue providing a positive ad experience for users by avoiding disruptive ads. This will be critical to sustain ad revenue and remain in good standing with the AdSense program.

In Summary

While Google’s AdSense updates aim to simplify and bring transparency to the monetization process, it’s up to publishers to leverage these changes for their benefit.

By staying informed, monitoring performance, and adapting strategies, publishers can continue to thrive.

FAQ

What is the significance of Google transitioning AdSense to an eCPM payment model for publishers?

How will the updated revenue share impact AdSense publishers?

  • The updated revenue share structure changes publishers’ compensation, introducing separate rates for buy-side (advertisers) and sell-side (publishers).
  • Publishers will receive 80% of the revenue after the advertiser platform takes fees, whether from Google’s buy-side or third-party platforms. This could affect income predictability as third-party fees can vary and are not controlled by Google.
  • The overall revenue for publishers is expected to stay around 68%, similar to the previous agreement. Still, there may be individual variations based on the details of each transaction and the parties involved.

What strategies can publishers employ to adapt to the eCPM payment model?

  • Increase website traffic volume to generate more ad impressions and potentially increase revenue.
  • Improve user engagement metrics such as time on page and pages per session, as higher engagement may lead to increased ad impressions.
  • Extend session duration on their sites to serve more ad impressions. This could involve enhancing content quality or providing additional resources to keep users engaged for extended periods.
  • Ensure compliance with AdSense policies and Better Ads Standards to provide a positive ad experience and sustain ad revenue growth.


Featured Image: Wirestock Creators/Shutterstock

Why Did Google Gemini “Leak” Chat Data? via @sejournal, @martinibuster

It only took twenty four hours after Google’s Gemini was publicly released for someone to notice that chats were being publicly displayed in Google’s search results. Google quickly responded to what appeared to be a leak. The reason how this happened is quite surprising and not as sinister as it first appears.

@shemiadhikarath tweeted:

“A few hours after the launch of @Google Gemini, search engines like Bing have indexed public conversations from Gemini.”

They posted a screenshot of the site search of gemini.google.com/share/

But if you look at the screenshot, you’ll see that there’s a message that says, “We would like to show you a description here but the site won’t allow us.”

By early morning on Tuesday February 13th the Google Gemini chats began dropping off of Google search results, Google was only showing three search results. By the afternoon the number of leaked Gemini chats showing in the search results had dwindled to just one search result.

Screenshot of Google's search results for pages indexed from the Google Gemini chat subdomain

How Did Gemini Chat Pages Get Created?

Gemini offers a way to create a link to a publicly viewable version of a private chat.

Google does not automatically create webpages out of private chats. Users create the chat pages through a link at the bottom of each chat.

Screenshot Of How To Create a Shared Chat Page

Screenshot of how to create a public webpage of a private Google Gemini Chat

Why Did Gemini Chat Pages Get Indexed?

The obvious reason for why the chat pages were crawled and indexed is because Google forgot to put a robots.txt in the root of the Gemini subdomain, (gemini.google.com).

A robots.txt file is a document for controlling crawler activity on websites. A publisher can block specific crawlers by using commands standardized in the Robots.txt Protocol.

I checked the robots.txt at 4:19 AM on February 13th and saw that one was in place:

Google Gemini robots.txt file

I next checked the Internet Archive to see how long the robots.txt file has been in place and discovered that it was there since at least February 8th, the day that the Gemini Apps were announced.

Screenshot of Google Gemini robots. txt from Internet Archive showing it was there on February 8, 2024.

That means that the obvious reason for why the chat pages were crawled is not the correct reason, it’s just the most obvious reason.

Although the Google Gemini subdomain had a robots.txt that blocked web crawlers from both Bing and Google, how did they end up crawling those pages and indexing them?

Two Ways Private Chat Pages Discovered And Indexed

  • There may be a public link somewhere.
  • Less likely but maybe possible is that they were discovered through browsing history linked from cookies.

It’s likelier that there’s a public links. But if there’s a public link then why did Google start dropping chat pages altogether? Did Google create an internal rule for the search crawler to exclude webpages from the /share/ folder from the search index, even if they’re publicly linked?

Insights Into How Bing and Google Search Index Content

Now here’s the really interesting part for all the search geeks interested in how Google and Bing index content.

The Microsoft Bing search index responded to the Gemini content differently from how Google search did. While Google was still showing three search results in the early morning of February 13th, Bing was only showing one result from the subdomain. There was a seemingly random quality to what was indexed and how much of it.

Why Did Gemini Chat Pages Leak?

Here are the known facts: Google had a robots.txt in place since the February 8th. Both Google and Bing indexed pages from the gemini.google.com subdomain. Google indexed the content regardless of the robots.txt and then began dumping them.

  • Does Googlebot have a different instructions for indexing content on Google subdomains?
  • Does Googlebot routinely crawl and index content that is blocked by robots.txt and then subsequently drop it?
  • Was the leaked data linked from somewhere that is crawlable by bots, causing the blocked content to be crawled and indexed?

Content that is blocked by Robots.txt can still be discovered, crawled and end up in the search index and ranked in the SERPs or at least through a site:search. I think this may be the case.

But if that’s the case, why did the search results begin to drop off?

If the reason for the crawling and indexing was because those private chats were linked from somewhere, was the source of the links removed?

The big question is, where are those links? Could it be related to annotations by quality raters that unintentionally leaked onto the Internet?

Google Uses AI To Detect Fake Online Reviews Faster via @sejournal, @MattGSouthern

Google is harnessing new AI technology to block a surge of fake online reviews that mislead customers and harm local businesses.

The company stopped over 170 million fake reviews in 2023 – a 45% increase over the previous year.

The crackdown provides welcome relief for local business owners who have struggled with fraudulent reviews that damage their reputations on Google Maps and Search.

In a blog post, Google discusses how its algorithms now analyze patterns over time to identify suspicious review activity quickly. This includes spotting identical reviews copied across business pages or sudden spikes of 1-star and 5-star ratings.

A Year Of More Honest Reviews

In 2023, Google received approximately 20 million daily updates to local business information, including details such as business hours and customer reviews.

To maintain data integrity amidst this high volume of incoming data, Google implemented a new algorithm to rapidly identify and remove misleading or deceptive information.

Google notes the new algorithm identified a scam where people were paid to write false positive reviews. By detecting it, Google could take action to shut it down. This prevented the further spread of deceptive reviews.

Additionally, Google reported removing or blocking the following:

  • 170+ million reviews that violated policies,
  • 14 million increase in detecting policy-violating videos
  • 2+ million attempts of fraudulent claims on business profiles being thwarted due to enhanced security protocols.

Key Benefits For Local Businesses

Google reports that its new fake review detection algorithm can help local businesses in the following ways:

  • Faster Detection: The new machine learning system detects suspicious review patterns more quickly than previous methods, which can help protect businesses from the harm of fraudulent reviews.
  • Increased Accuracy: The accuracy of identifying fake reviews has improved by 45% compared to 2022, which can give businesses more confidence that their online ratings primarily reflect real customer experiences.
  • Scam Protection: The algorithm identifies individual fake reviews and broader coordinated efforts to post scam reviews, providing businesses with enhanced protection from organized fake review campaigns.

Takeaways For Local Marketers & Business Owners

Google states its continued effort to tackle fake content will benefit users and protect local businesses by better securing their online reputations.

The update could lead to a more level playing field, with reputation becoming a more accurate reflection of service quality and genuine customer satisfaction.

For businesses, the message is to concentrate on delivering exceptional products and services, while Google works to ensure online reputation matches real-world performance.


Featured Image: Screenshot from blog.google/products/maps/how-machine-learning-keeps-contributed-content-helpful/, February 2024. 

FAQ

How has the new algorithm changed the efficiency of Google’s review moderation process?

The new algorithm has significantly enhanced the efficiency of Google’s moderation process by:

  • Identifying Patterns: By analyzing review patterns over time, the algorithm can swiftly pinpoint anomalous activities like duplicated content and unusual rating fluctuations.
  • Volume Handling: Google’s ability to manage roughly 20 million daily updates to local business information demonstrates the algorithm’s capacity to handle large volumes of data while maintaining accuracy.
  • Stopping Scams: Google’s proactive measures have shut down schemes where individuals were compensated to write falsified reviews, protecting the integrity of business ratings.

What practical measures can I take as a marketer in light of Google’s new review detection capabilities?

With Google’s improved fake review detection capabilities, marketers should consider the following actions:

  • Focus on Authenticity: Encourage genuine customer feedback instead of using artificial means to inflate ratings.
  • Monitor Reviews: Regularly assess your business reviews on Google for abnormal patterns and report suspicious activities.
  • Engage with Customers: Respond to positive and negative reviews to demonstrate active engagement and concern for customer experience.

WordPress SiteOrigin Widgets Bundle Plugin Vulnerability Affects +600,000 Sites via @sejournal, @martinibuster

SiteOrigin Widgets Bundle WordPress plugin with over 600,000 installations patched an authenticated stored cross-site scripting (XSS) vulnerability that could allow attackers to upload arbitrary files and expose site visitors to malicious scripts.

SiteOrigin Widgets Bundle Plugin

The SiteOrigins Widgets plugin, with +600,000 active installations, provides a way to easily add a multitude of widget functions like sliders, carousels, maps, change the way blog posts are displayed, and other useful webpage elements.

Stored Cross-Site Scripting Vulnerability

A Cross-Site Scripting (XSS) vulnerability is a flaw allows a hacker to inject (upload) malicious scripts. In WordPress plugins, these kinds of vulnerabilities arise from flaws in how data that’s input is not properly sanitized (filtered for untrusted data) and also from improperly securing output data (called escaping data).

This particular XSS vulnerability is called a Stored XSS because the attacker is able to inject the malicious code to the server.  According to the non-profit Open Worldwide Application Security Project (OWASP), the ability to launch an attack directly from the website makes it particularly concerning.

OWASP describes the stored XSS threat:

“This type of exploit, known as Stored XSS, is particularly insidious because the indirection caused by the data store makes it more difficult to identify the threat and increases the possibility that the attack will affect multiple users. “

In an XSS attack, where a script has successfully been injected, the attacker sends a harmful script to an unsuspecting site visitor. The user’s browser, because it trusts the website, executes the file. This can allow the attacker to access cookies, session tokens, and other sensitive website data.

Vulnerability Description

The vulnerability arose because of flaws in sanitizing inputs and escaping data.

The WordPress developer page for security explains sanitization:

“Sanitizing input is the process of securing/cleaning/filtering input data. Validation is preferred over sanitization because validation is more specific. But when “more specific” isn’t possible, sanitization is the next best thing.”

Escaping data in a WordPress plugin is a security function that filters out unwanted output.

Both of those functions needed improvement in the SiteOrigins Widgets Bundle plugin.

Wordfence described the vulnerability:

“The SiteOrigin Widgets Bundle plugin for WordPress is vulnerable to Stored Cross-Site Scripting via the onclick parameter in all versions up to, and including, 1.58.3 due to insufficient input sanitization and output escaping.”

This vulnerability requires authentication before it can be executed, which means the attacker needs at least a contributor level access in order to be able to launch an attack.

Recommended action:

The vulnerability was assigned a medium CVSS severity level, scoring 6.4/10. Plugin users should consider updating to the latest version, which is version 1.58.5, although the vulnerability was patched in version 1.58.4.

Read the Wordfence vulnerability advisory:

SiteOrigin Widgets Bundle <= 1.58.3 – Authenticated (Contributor+) Stored Cross-Site Scripting

B2B Content Marketing Strategies For High-Quality Lead Generation via @sejournal, @sejournal

Content is key for generating high-quality leads in B2B marketing.

You must know how to leverage content to begin and nurture meaningful interactions. Words and graphics have the power to drive conversions and forge lasting connections.

Ultimately, successful content marketing isn’t just about being noticed but being remembered.

And generating leads isn’t just about numbers; it’s about connecting with real people.

If you’re ready to elevate your content strategy in 2024, our latest ebook, B2B Lead Generation: Create Content That Converts, is your playbook to providing maximum value to your audience next year.

This concise yet powerful guide is full of exclusive insights from high-profile experts, while drawing on our own internal expertise in delivering leads across multiple media types.

Here’s a peek at what you’ll find inside:

  • Using content innovation for lead generation: The oversaturation of content is a big issue in B2B today – and standing out isn’t just a challenge; it’s an art form. In order to elevate your brand and leave a lasting impression, you need the right combination of innovation and strategy. The key is in surprising your audience with content that feels fresh, speaks directly to their needs, and keeps them eagerly awaiting your next move.
  • Building sustainable and scalable lead programs: Remember, it’s not just about acquiring leads; it’s about making sure they feel valued and guided, and cultivating long-lasting relationships. To build a strong lead program, you must properly qualify and score your leads to determine how to best engage with and nurture them. Be sure to map out your customer journey, tailor your communication, and provide valuable touchpoints at every step to guide your leads from curiosity to commitment.
  • Mastering the knowledge exchange: Your goal should be to “add value before you extract value” – as Curtis del Principe, Sr. Marketing Manager at Hubspot, puts it. It’s all about crafting a win-win scenario between you and your audience. When you manage to address their biggest pain points and answer their most pressing questions, they willingly share their information in return. Create the kind of content that can captivate, engage, and inspire your leads to action.
  • Leveraging consumer psychology to forge meaningful connections: At its core, B2B is about humans connecting with humans. Although interactions can sometimes feel transactional, forging genuine connections and building relationships that go beyond the surface is crucial for success in this space. Try to understand why your audience does what they do so you can adjust your content to their unique needs and preferences. Offer value to them upfront and present solutions in a way that resonates with them personally.

B2B Lead Generation: Create Content That Converts

This digital handbook is tailored for the B2B marketing professional, whether you work within a specific brand or with an agency.

Grab your copy of B2B Lead Generation: Create Content That Converts for expert insights and actionable tips to help you start driving qualified leads and building genuine relationships.


Featured Image: Paulo Bobita/Search Engine Journal

WordPress Backup Plugin DoS Vulnerability Affects +200,000 Sites via @sejournal, @martinibuster

A popular WordPress backup plugin installed in over 200,000 websites recently patched a high severity vulnerability that could lead to a denial of service attack. Wordfence assigned a CVSS severity level rating of High, with a score of 7.5/10, indicating that plugin users should take note and update their plugin.

Backuply Plugin

The vulnerability affects the Backuply WordPress backup plugin. Creating backups is a necessary function for every website, not just WordPress sites, because backups help publishers roll back to a previous version should the server fail and lose data in a catastrophic failure.

Website backups are invaluable for site migrations, hacking recovery and failed updates that render a website non-functional.

Backuply is an especially useful plugin because it backup data to multiple trusted third party cloud services and supports multiple ways to download local copies in order to create redundant backups so that if a cloud backup is bad the site can be recovered from another backup stored locally.

According to Backuply:

“Backuply comes with Local Backups and Secure Cloud backups with easy integrations with FTP, FTPS, SFTP, WebDAV, Google Drive, Microsoft OneDrive, Dropbox, Amazon S3 and easy One-click restoration.”

Vulnerability Affecting Backuply

The United States Government National Vulnerability Database warns that Backuply up to and including version 1.2.5 contains a flaw that can lead to denial of service attacks.

The warning explains:

“This is due to direct access of the backuply/restore_ins.php file and. This makes it possible for unauthenticated attackers to make excessive requests that result in the server running out of resources.”

Denial Of Service (DoS) Attack

A denial of service (DoS) attack is one in which a flaw in a software allows an attacker to make so many rapid requests that the server runs out of resources and can no longer process any further requests, including serving webpages to site visitors.

A feature of DoS attacks is that it is sometimes possible to upload scripts, HTML or other code that can then be executed, allowing the attacker to perform virtually any action.

Vulnerabilities that enable DoS attacks are considered critical, and steps to mitigate them should be taken as soon as possible.

Backuply Changelog Documentation

The official Backuply changelog, which announces the details of every update, notes that a fix was implemented in version of 1.2.6. Backuply’s transparency and rapid response is responsible and a sign of a trustworthy developer.

According to the Changelog:

“1.2.6 (FEBRUARY 08 2024)
[Security-Fix] In some cases it was possible to fill up the logs and has been fixed. Reported by Villu Orav (WordFence)”

Recommendations

In general it is highly recommended that all users of the Backuply plugin update their plugin as soon as possible in order to prevent an unwanted security event.

Read the National Vulnrability Database description of the vulnerability:

CVE-2024-0842

Read the Wordfence Backuply vulnerability report:

Backuply – Backup, Restore, Migrate and Clone <= 1.2.5 – Denial of Service

Featured Image by Shutterstock/Doppelganger4

Google Gemini Warning: Don’t Share Confidential Information via @sejournal, @martinibuster

Google Gemini privacy support pages warn that information shared with Gemini Apps may be read and annotated by human reviewers and also be included into AI training datasets.  This is what you need to know and what actions are available to prevent this from happening.

Google Gemini

Gemini is the name for the technology underlying the Google Gemini Android App available on Google Play, a functionality in the Apple iPhone Google App and a standalone chatbot called Gemini Advanced.

Gemini on Android, iPhone & Gemini Advanced

Gemini on mobile devices and the standalone chatbot are multimodal. Multimodal means that users can ask it questions with images, audio or text input. Gemini can answer questions about things in the real world, respond to questions, can perform actions, provide information about an object in a photo or provide instructions on how to use it.

All of that data in the form of images, audio and text are submitted to Google and some of it could be reviewed by humans or included in AI training datasets.

Gemini Uses Gemini Data To Create Training Datasets?

Gemini Apps uses past conversations and location data for generating responses, which is normal and reasonable. Gemini also collects and stores that same data to improve other Google products.

This is what the privacy explainer page says about it:

“Google collects your Gemini Apps conversations, related product usage information, info about your location, and your feedback. Google uses this data, consistent with our Privacy Policy, to provide, improve, and develop Google products and services and machine learning technologies, including Google’s enterprise products such as Google Cloud.”

Google’s privacy explainer says that the data is stored in a users Google Account for up to 18 months and they are able to limit the data storage to three months and 36 months.

There’s also a way to turn off saving data to a users Google Account:

“If you want to use Gemini Apps without saving your conversations to your Google Account, you can turn off your Gemini Apps Activity.

..Even when Gemini Apps Activity is off, your conversations will be saved with your account for up to 72 hours. This lets Google provide the service and process any feedback. This activity won’t appear in your Gemini Apps Activity.”

But there’s an exception to the above rule that lets Google hold on to the data for even longer.

Human Reviews Of User Gemini Data

Google’s Gemini privacy support page explains that user data that is reviewed and annotated by human reviewers is retained by Google for up to three years.

How long is reviewed data retained
Gemini Apps conversations that have been reviewed by human reviewers (as well as feedback and related data like your language, device type, or location info) are not deleted when you delete your Gemini Apps activity because they are kept separately and are not connected to your Google Account. Instead, they are retained for up to 3 years.”

The above cited support page informs that human reviewed and annotated data is used to create datasets for Chatbots:

“These are then used to create a better dataset for generative machine-learning models to learn from so our models can produce improved responses in the future.”

Google Gemini Warning:  Don’t Share Confidential Data

Google’s Gemini privacy explainer page warns that users should not share confidential information.

It explains:

“To help with quality and improve our products (such as generative machine-learning models that power Gemini Apps), human reviewers read, annotate, and process your Gemini Apps conversations. We take steps to protect your privacy as part of this process. This includes disconnecting your conversations with Gemini Apps from your Google Account before reviewers see or annotate them.

Please don’t enter confidential information in your conversations or any data you wouldn’t want a reviewer to see or Google to use to improve our products, services, and machine-learning technologies.

…Don’t enter anything you wouldn’t want a human reviewer to see or Google to use. For example, don’t enter info you consider confidential or data you don’t want to be used to improve Google products, services, and machine-learning technologies.”

There is a way to keep all that from happening. Turning off Gemini Apps Activity stops user data from being shown to human reviewers, so there is a way to opt-out and not have the data stored and used to create datasets.

But, Google still stores data up to 72 hours in order to have a backup in case of a failure but also for sharing with other Google services and with third party services that a user may interact with while using Gemini.

Using Gemini Can Lead To 3rd Party Data Sharing

Using Gemini can start a chain reaction of other apps using and storing user conversations, location data and other information.

The Gemini privacy support page explains:

“If you turn off this setting or delete your Gemini Apps activity, other settings, like Web & App Activity or Location History, may continue to save location and other data as part of your use of other Google services.

In addition, when you integrate and use Gemini Apps with other Google services, they will save and use your data to provide and improve their services, consistent with their policies and the Google Privacy Policy. If you use Gemini Apps to interact with third-party services, they will process your data according to their own privacy policies.”

The same Gemini privacy page links to a page for requesting removal of content, as well as to a Gemini Apps FAQ and a Gemini Apps Privacy Hub to learn more.

Gemini Use Comes With Strings

Many of the ways that Gemini uses data is for legitimate purposes, including submitting the information for human reviews.  But Google’s own Gemini support pages make it very clear that users should not share any confidential information that a human reviewer might see or because it might get included into an AI training dataset.

Featured Image by Judith Linine / Shutterstock.com

Google Updates Web Stories Availability via @sejournal, @martinibuster

Google updated the Web Stories documentation to reflect that Web Stories will no longer appear in Google Images, among other changes.

Google Web Stories

Web Stories is a visual short content form that is designed for short moments like when riding on a bus. They’re visual with short sentences and are in a carousel form where swiping reveals the next “page” of the content.

Google calls them snackable content and visual storytelling.

This is how Google describes Web Stories:

“Web Stories immerse your readers in fast-loading full-screen experiences. Easily create visual narratives, with engaging animations and tappable interactions.
The Web Story format (formerly known as AMP Stories) is free and part of the open web and are available for everyone to try on their websites. They can be shared and embedded across the web without being confined to a closed ecosystem or platform.

Web Stories provide content publishers with a mobile-focused format for delivering news and information as visually rich, tap-through stories. Web Stories offers a robust set of advertising opportunities for advertisers and publishers to reach a unique audience on the web with immersive experiences.”

Changes to Availability

Web Stories are available in Google Search, which includes Google Discover. They were formerly available in Google Images but that is no longer the case.

According to Google’s Developer Pages changelog:

“Updated the availability of Web Stories
What: Updated the feature availability of Web Stories.

Why: To make sure our documentation aligns with how the feature appears in Google Search. Web Stories don’t appear in Google Images anymore, and the grid view is now a carousel view in Search results.”

Google Search Central Help Page For Web Stories

Google’s Search Central developer page for Web Stories was updated to reflect the changes to where Web Stories are available.

The first change to the documentation is a new heading that emphasizes that Web Stories are a part of Google Search.

This is the new heading:

“A Web Story on Google Search”

The next change is the removable of content about how web stories appear across Google, replaced with a new heading that emphasizes that Web Stories is a Search Feature.

The second new heading:

“Feature availability “

The most substantial change is the addition of a new paragraph that accompanies the above heading.

This is the new section:

Feature availability

“Web Stories can appear as a single result on Google Search, which is available in all regions and languages where Google Search is available. Web Stories can also appear in a carousel on Google Search, which is available in the United States (in English), India (in English and Hindi), and Brazil (in Brazilian Portuguese).

In the Discover feed, Web Stories can appear as a single card where you can tap through the story. While this appearance is available in all regions and languages where Google Discover is available, it’s most likely to appear in the United States, India, and Brazil.”

Read Google’s changelog entry for this change:

February 8 Updated the availability of Web Stories

Featured Image by Shutterstock/Bplanet

Google Clarifies the “Google-Extended” Crawler Documentation via @sejournal, @martinibuster

Google recently updated the documentation of its Google-Extended web crawler user agent, reflecting changes in product naming and clarifying the impact on search, which may be a concern for those who choose to block the crawler. The updated documentation offers clearer guidance on controlling content access for use in AI model training.

Google-Extended User Agent

Introduced on September 28, 2023, Google-Extended offers web publishers a user agent that can be used to control how their sites are crawled. Publishers can allow or disallow the Google-Extended user agent using the Robots Exclusion Protocol, giving them a way to opt-out of having their content scraped and included in AI training datasets.

Google describes Google-Extended as a “standalone product token” but that’s non-standard terminology for how publishers understand the concept of User Agents.

The original announcement described the new user agent:

“Today we’re announcing Google-Extended, a new control that web publishers can use to manage whether their sites help improve Bard and Vertex AI generative APIs, including future generations of models that power those products.

By using Google-Extended to control access to content on a site, a website administrator can choose whether to help these AI models become more accurate and capable over time.”

Blocking Google-Extended is done with the “Google-Extended” User Agent:

User-agent: Google-Extended
Disallow: /

Google Changelog

Google keeps a changelog of important updates made to guidance and communication with web publishers and the search marketing community. The changelog of Google’s developer pages announced a change to the Google-Extended documentation.

The revision comes after the renaming of Bard to Gemini Apps, specifying that Google-Extended’s indexing now contributes to Gemini Apps and Vertex AI generative APIs. The new wording reassures publishers that this does not affect Google Search, addressing potential concerns about the possible implications from opting out of Google-Extended AI data collection.


What Changed?

Google’s changelog clarifies that Google-Extended crawling is exclusive to Gemini Apps and has no impact on Google Search.

The Changelog advises:

“Updated the description of the Google-Extended product token
What: With the name change of Bard to Gemini Apps, we clarified that Gemini Apps is affected by Google-Extended, and, based on publisher feedback, we specified that Google-Extended doesn’t affect Google Search.”

The updated guidance no longer uses the Bard brand name, switching it out to Gemini. And the following sentence was added:

“Google-Extended does not impact a site’s inclusion or ranking in Google Search.”

Read Google’s updated crawler overview:

Overview of Google crawlers and fetchers (user agents)

Featured Image by Shutterstock/Ribkhan