The Download: police AI, and mixed reality’s future

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

How the largest gathering of US police chiefs is talking about AI

—James O’Donnell

The International Association of Chiefs of Police bills itself as the largest gathering of its type in the United States. Leaders from many of the country’s 18,000 police departments and even some from abroad convene for product demos, discussions, parties, and awards. 

I went along last month to see how artificial intelligence was being discussed, and the message to police chiefs seemed crystal clear: If your department is slow to adopt AI, fix that now. From the expo hall, talks, and interviews, it seems they’re already enthusiastically heeding the call. Read the full story.

This story is from The Algorithm, our weekly AI newsletter. Sign up to receive it in your inbox every Monday.

Roundtables: What’s Next for Mixed Reality: Glasses, Goggles, and More

After years of trying, augmented-reality specs are at last a thing. 

If you want to learn more about where AR experiences are heading, join our editor-in-chief Mat Honan and AI hardware reporter James O’Donnell for a Roundtables conversation streamed online at 2pm ET/11am PT today. It’s for subscribers only but good news: this week our subscriptions are half price. Don’t miss out! 

Read more about mixed reality:

+ We interviewed Palmer Luckey, founder of Oculus, about his plans to bring mixed-reality goggles to soldiers. Here’s what he had to say.

+  The coolest thing about smart glasses is not the AR. It’s the AI.

+ Snap has launched new augmented-reality Spectacles. Here’s what we made of them

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 The FBI is investigating threats texted to Latino and LGBTQ+ people 
They claim recipients will be deported or sent to a re-education camp. (WP $)
+  ICE can already sidestep sanctuary city laws through data-sharing centers. (Wired $)
Trump has confirmed he plans to use the military for mass deportations. (NYT $)

2 Chinese tech groups are building AI teams in Silicon Valley 
Despite Washington’s best efforts to stymie their work. (FT $)
How a US ban on investing in Chinese startups could escalate under Trump. (Wired $)

3 How Apple will cope with looming tariffs 
The fact CEO Tim Cook already has a relationship with Trump will surely help. (Bloomberg $)

4 Two undersea cables in the Baltic Sea have been disrupted 
It looks like Russia is trying to interfere with global undersea infrastructure. (CNN)
A Russian spy ship had to be escorted out of the Irish Sea last weekend too. (The Guardian)

5 An AI tool could help solve math problems humans are stuck on
It’s a good example of how blending human and machine intelligence can produce positive results. (New Scientist $)
+ This AI system makes human tutors better at teaching children math. (MIT Technology Review)

6 Robots still struggle to match warehouse workers on some tasks
For all the advances robots have made, picking things up and moving them around remains a big challenge. (NYT $)
+ AI is poised to automate today’s most mundane manual warehouse task. (MIT Technology Review)

7 Perplexity’s AI search engine can now buy stuff for you
How long until Google follows? (The Verge)

8 Dozens of states are begging Congress to pass the kids online safety act
It has currently stalled in the House of Representatives due to censorship concerns. (The Verge)
Roblox is adding more controls to let parents set daily usage limits, block access to certain game genres, and more. (WSJ $)
+ Why child safety bills are popping up all over the US

9 The US Patent and Trademark Office banned staff from using generative AI
It cited security concerns plus the fact some tools exhibit “bias, unpredictability, and malicious behavior.” (Wired $)

10 NASA might have killed life on Mars 😬
A new paper suggests that adding water to Martian soil might have been a bad move. (Quartz $)
+  The ISS has been leaking air for 5 years, and engineers still can’t agree why. (Ars Technica)

Quote of the day

“We are bleeding cash as an industry.” 

—Thomas Laffont, co-founder of investment firm Coatue Management, says venture capital firms are struggling to make money amid a boom in AI investments, the Wall Street Journal reports.

 The big story

How mobile money supercharged Kenya’s sports betting addiction

BRIAN OTIENO

April 2022

Mobile money has mostly been hugely beneficial for Kenyans. But it has also turbo-charged the country’s sports betting sector.

Experts and public figures across the African continent are sounding the alarm over the growth of the sector increasingly loudly. It’s produced tales of riches, but it has also broken families, consumed college tuitions, and even driven some to suicide. Read the full story.

—Jonathan W. Rosen

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ I just learned a pertinent word for this season: abscission
+ Only some people will get this… but if you’re one of them, you’ll enjoy it. 
+ Why Late of the Pier were one of the most exciting UK bands of the 2000s.
+ Whether you call them crisps or chips, they’re goddamn delicious.

The rise of Bluesky, and the splintering of social

You may have read that it was a big week for Bluesky

If you’re not familiar, Bluesky is, essentially, a Twitter clone that publishes short-form status updates. It gained more than 2 million users this week. On Wednesday, The Verge reported it had crossed 15 million users. By Thursday, it was at 16 million. By Friday? 17 million and counting. It was the number one app in Apple’s app store. 

Meanwhile, Threads, Meta’s answer to Twitter, put up even bigger numbers. The company’s Adam Mosseri reported that 15 million people had signed up in November alone. Both apps are surging in usage. 

Many of these new users were seemingly fleeing X, the platform formerly known as Twitter. On the day after the election, more than 115,000 people deactivated their X accounts, according to Similarweb data. That’s a step far past not logging on. It means giving up your username and social graph. It’s nuking your account versus just ignoring it. 

Much of that migration is likely a reaction to Elon Musk’s support of Donald Trump, and his moves to elevate right-leaning content on the platform. Since Musk took over, X has reinstated a lot of previously banned accounts, very many of which are on the far right. It also tweaked its algorithm to make sure Musk’s own posts, which are often pro-Trump, get an extra level of promotion and prominence, according to Kate Conger and Ryan Mac’s new book Character Limit

There are two points I want to make here. The first is that tech and politics are just entirely enmeshed at this point. That’s due to the extreme extent to which tech has captured culture and the economy. Everything is a tech story now, including and especially politics. 

The second point is about what I see as a more long-term shift away from centralization. What’s more interesting to me than people fleeing a service because they don’t like its politics is the emergence of unique experiences and cultures across all three of these services, as well as other, smaller competitors.

Last year, we put “Twitter killers” on our list of 10 breakthrough technologies. But the breakthrough technology wasn’t the rise of one service or the decline of another. It was decentralization. At the time, I wrote: 

“Decentralized, or federated, social media allows for communication across independently hosted servers or platforms, using networking protocols such as ActivityPub, AT Protocol, or Nostr. It offers more granular moderation, more security against the whims of a corporate master or government censor, and the opportunity to control your social graph. It’s even possible to move from one server to another and follow the same people.”

In the long run, massive, centralized social networks will prove to be an aberration. We are going to use different networks for different things. 

For example, Bluesky is great for breaking news because it does not deprioritize links and defaults to a social graph that shows updates from the people you follow in chronological order. (It also has a Discover feed and you can set up others for algorithmic discovery—more on that in a moment—but the default is the classic Twitter-esque timeline.) 

Threads, which has a more algorithmically defined experience, is great for surfacing interesting conversations from the past few days. I routinely find interesting comments and posts from two or three days before I logged on. At the same time, this makes it pretty lousy at any kind of real time experience—seemingly intentionally—and essentially hides that standard timeline of updates from people you follow in favor of an algorithmically-generated “for you” feed. 

I’m going to go out on a limb here and say that while these are quite different, neither is inherently better. They offer distinct takes on product direction. And that ability to offer different experiences is a good thing. 

I think this is one area where Bluesky has a real advantage. Bluesky lets people bend the experience to their own will. You aren’t locked into the default following and discover experiences. You can roll your own custom feed, and follow custom feeds created by other people. (And Threads is now testing something similar.) That customization means my experience on Bluesky may look nothing like yours. 

This is possible because Bluesky is a service running on top of the AT Protocol, an open protocol that’s accessible to anyone and everyone. The entire idea is that social networking is too important for any one company or person to control it. So it is set up to allow anyone to run their own network using that protocol. And that’s going to lead to a wide range of outcomes. 

Take moderation, as an example. The moderation philosophy of the AT Protocol is essentially that everyone is entitled to speech but not to reach. That means it isn’t banning content at the protocol level, but that individual services can set up their own rules. 

Bluesky has its own community guidelines. But those guidelines would not necessarily apply to other services running on the protocol. Furthermore, individuals can also moderate what types of posts they want to see. It lets people set up and choose different levels of what they want to allow. That, combined with the ability to roll your own feeds, combined with the ability of different services to run on top of the same protocol, sets up a very fragmented future. 

And that’s just Bluesky. There’s also Nostr, which leans toward the crypto and tech crowds, at least for now. And Mastodon, which tends to have clusters of communities on various servers. All of them are growing. 

The era of the centralized, canonical feed is coming to an end. What’s coming next is going to be more dispersed, more fractured, more specialized. It will take place across these decentralized services, and also WhatsApp channels, Discord servers, and other smaller slices of Big Social. That’s going to be challenging. It will cause entirely new problems. But it’s also an incredible opportunity for individuals to take more control of their own experiences.

If someone forwarded you this edition of The Debrief, you can subscribe here. I appreciate your feedback on this newsletter. Drop me a line at mat.honan@technologyreview.com with any and all thoughts. And of course, I love tips.

Now read the rest of The Debrief

The News

TSMC halts advanced chip shipments for Chinese clients. It comes after some of its chips were found inside a Huawei AI processor.

Google DeepMind has come up with a new way to peer inside AI’s thought process.

An AI lab out of Chicago is building tools to help creators prevent their work from being used in training data.

Lina Khan may be on the way out, but she’s going out with a bang: The FTC is preparing to investigate Microsoft’s cloud business.

The Chat

Every week I’ll talk to one of MIT Technology Review’s reporters or editors to find out more about what they’ve been working on. For today, I spoke with Casey Crownhart, senior climate reporter, about her coverage of the COP29 UN climate conference.

Mat: COP29 is happening right now in Azerbaijan, do you have a sense of the mood?

Casey: The vibes are weird in Baku this week, in part because of the US election. The US has been a strong leader in international climate talks in recent years, and an incoming Trump administration will certainly mean a big change.

And the main goal of these talks—reaching a climate finance agreement—is a little daunting. Developing countries need something like $1 trillion dollars annually to cope with climate change. That’s a huge jump from the current target, so there are questions about how this agreement will shake out.

Mat: Azerbaijan seems like a weird choice to host. I read one account from the conference saying you could smell the oil in the air. Why there?

Casey: Azerbaijan’s economy is super reliant on fossil fuels, which definitely makes it an ironic spot for international climate negotiations.

There’s a whole complicated process of picking the COP host each year—five regions rotate hosting, and the countries in that region have to all agree on a pick when it’s their turn. Russia apparently vetoed most of the other choices in the Eastern European group this year, and the region settled on Azerbaijan as one of the only viable options.

Mat: You write that if Trump pulls out of the UN Framework Convention on Climate Change, it would be like riding away on a rocket. Why would that be so much worse than dropping out of Paris?

Casey: Trump withdrew from the Paris Agreement once already, and it was relatively easy for Biden to rejoin when he came into office. If, during his second term, Trump were to go a step further and pull out of the UNFCCC, it’s not just an agreement he’s walking away from, it’s the whole negotiating framework. So the statement would be much bigger.

There’s also the question of reversibility. It’s not clear if Trump can actually withdraw from the UNFCCC on his own, and it’s also not clear what it would take to rejoin it. When the US joined in the ’90s, the Senate had to agree, so getting back in might not be as simple as a future president signing something.

Mat: What from COP29 are you optimistic about?

Casey: Tough to find a glimmer of hope in all this, but if there is one, I’d say I’m optimistic that we’ll see some countries step up, including the UK and China. The UK announced a new emissions target at the talks already, and it’ll be really interesting to see what role China plays at COP29 and moving forward.

The Recommendation

Once upon a time I was a gadget blogger. It’s fun writing about gadgets! I miss it! Especially because at some point your phone became the only device you need. But! My beloved wife bought me a Whoop fitness tracker for my birthday. It’s an always-on device that you wear around your wrist. I’ve been Oura-curious for some time, but frankly I am a little bit terrified of rings. I spent a number of months going to a hand rehab clinic after a bike accident, and while I was there first learned about degloving and how commonly it happens to people because a ring gets caught on something. Just thought I’d put that in your head too. Anyway! The whoop is a fabric bracelet with a little monitor on it. It tracks your movement, your heart rate, your sleep, and a lot more. There’s no screen, so it’s very low profile and unobtrusive. It is, however, pretty spendy: The device is free but the plan costs $239 annually.

The Download: Bluesky’s rapid rise, and harmful fertility stereotypes

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

The rise of Bluesky, and the splintering of social

You may have read that it was a big week for Bluesky. If you’re not familiar, Bluesky is, essentially, a Twitter clone that publishes short-form status updates. Last Wednesday, The Verge reported it had crossed 15 million users. It’s just ticked over 19 million now, and is the number one app in Apple’s app store.

Meanwhile, Threads, Meta’s answer to Twitter, reportedly signed up 15 million people in November alone. Both apps are surging in usage.

Many of these new users were seemingly fleeing X, the platform formerly known as Twitter, in reaction to Elon Musk’s support of Donald Trump, and his moves to elevate right-leaning content on the platform. But there’s a deeper trend at play here. We’re seeing a long-term shift away from massive centralized social networks. Read the full story

—Mat Honan

This story is from The Debrief, our newly-launched newsletter written by our editor-in-chief Mat Honan. It’s his weekly take on the real stories behind the biggest news in tech—with some links to stories we love and the occasional recommendation thrown in for good measure. Sign up to get it every Friday!

Why the term “women of childbearing age” is problematic

—Jessica Hamzelou

Every journalist has favorite topics. Mine include the quest to delay or reverse human aging, and new technologies for reproductive health and fertility. So when I saw trailers for The Substance, a film centered on one middle-aged woman’s attempt to reexperience youth, I had to watch it.

I won’t spoil the movie for anyone who hasn’t seen it yet (although I should warn that it is not for the squeamish). But a key premise of the film involves harmful attitudes toward female aging. 

“Hey, did you know that a woman’s fertility starts to decrease by the age of 25?” a powerful male character asks early in the film. “At 50, it just stops,” he later adds. He never explains what stops, exactly, but to the viewer the message is pretty clear: If you’re a woman, your worth is tied to your fertility. Once your fertile window is over, so are you. 

The insidious idea that women’s bodies are, above all else, vessels for growing children has plenty of negative consequences for us all. But it also sets back scientific research and health policy. Read Jess’s story to learn how

This story is from The Checkup, MIT Technology Review’s weekly biotech newsletter.  Sign up to receive it in your inbox every Thursday.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 Trump plans to loosen US rules for self-driving cars 
No prizes for guessing who might be behind that idea. (Bloomberg $)
Elon Musk is ramping up his legal fight against OpenAI and Microsoft. (WSJ $)
Trump has appointed the FCC’s Brendan Carr to lead the agency. (NPR)
Robotaxis are here. It’s time to decide what to do about them. (MIT Technology Review)

2 How Bluesky is handling its explosive growth
It has just 20 employees, and they’re working round the clock to deal with bugs, outages and moderation issues. (NYT $)
+ Just joined Bluesky? Here’s how to use it. (The Verge)
How to fix the internet. (MIT Technology Review)

 3 Biden agreed to some small but significant AI limits with Xi Jinping 
I think we can all get behind the idea that nuclear weapons should be exclusively controlled by humans. (Politico)
Biden has lifted a ban on Ukraine using long-raise missiles to strike inside Russia. (BBC)

4 Big Tech is trying to sink the US online child safety bill 
And, as it stands, its lobbying efforts look very likely to succeed. (WSJ $)

5 Amazon has launched a rival to Temu and Shein 
Nothing on ‘Haul’ costs more than $20. (BBC)
+ Welcome to the slop era of online shopping. (The Atlantic $)

6 The Mike Tyson-Jake Paul fight on Netflix was plagued by glitches
Despite that, 60 million households still tuned in. (Deadline)

7 AI models can work together faster in their own language 
Linking different models together could help tackle thorny problems individual ones can’t solve. (New Scientist $)

8 Tech companies are training their AI on movie subtitles 
A database called OpenSubtitles provides a rare glimpse into what goes into these systems. (The Atlantic $)

9 McDonald’s is trying to bring back NFTs
Remember those? (Gizmodo)

10 A lot of people are confusing Starlink satellites with UFOs
Guess it’ll take us a while for us to get used to seeing them. (Ars Technica)

Quote of the day

“F*** you, Elon Musk.”

—Brazil’s first lady, Janja Lula da Silva, makes her views clear during a speech calling for tougher social media regulation ahead of the G20 summit in Rio de Janeiro, Reuters reports.

 The big story

Alina Chan tweeted life into the idea that the virus came from a lab

Alina Chan

COURTESY PHOTO

June 2021

Alina Chan started asking questions in March 2020. She was chatting with friends on Facebook about the virus then spreading out of China. She thought it was strange that no one had found any infected animal. She wondered why no one was admitting another possibility, which to her seemed very obvious: the outbreak might have been due to a lab accident.

Chan is a postdoc in a gene therapy lab at the Broad Institute, a prestigious research institute affiliated with both Harvard and MIT. Throughout 2020, Chan relentlessly stoked scientific argument, and wasn’t afraid to pit her brain against the best virologists in the world. Her persistence even helped change some researchers’ minds. Read the full story.

—Antonio Regalado

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ Why Quincy Jones was the best of the best.
+ These handy apps are a great way to save articles to read later on (Pocket is my own personal favorite.)
+ How to resurrect a ghost river in the Bronx.
+ Look after your stainless steel pans, and your stainless steel pans will look after you.

The AI lab waging a guerrilla war over exploitative AI

Ben Zhao remembers well the moment he officially jumped into the fight between artists and generative AI: when one artist asked for AI bananas. 

A computer security researcher at the University of Chicago, Zhao had made a name for himself by building tools to protect images from facial recognition technology. It was this work that caught the attention of Kim Van Deun, a fantasy illustrator who invited him to a Zoom call in November 2022 hosted by the Concept Art Association, an advocacy organization for artists working in commercial media. 

On the call, artists shared details of how they had been hurt by the generative AI boom, which was then brand new. At that moment, AI was suddenly everywhere. The tech community was buzzing over image-generating AI models, such as Midjourney, Stable Diffusion, and OpenAI’s DALL-E 2, which could follow simple word prompts to depict fantasylands or whimsical chairs made of avocados. 

But these artists saw this technological wonder as a new kind of theft. They felt the models were effectively stealing and replacing their work. Some had found that their art had been scraped off the internet and used to train the models, while others had discovered that their own names had become prompts, causing their work to be drowned out online by AI knockoffs.

Zhao remembers being shocked by what he heard. “People are literally telling you they’re losing their livelihoods,” he told me one afternoon this spring, sitting in his Chicago living room. “That’s something that you just can’t ignore.” 

So on the Zoom, he made a proposal: What if, hypothetically, it was possible to build a mechanism that would help mask their art to interfere with AI scraping?

“I would love a tool that if someone wrote my name and made a prompt, like, garbage came out,” responded Karla Ortiz, a prominent digital artist. “Just, like, bananas or some weird stuff.” 

That was all the convincing Zhao needed—the moment he joined the cause.

Fast-forward to today, and millions of artists have deployed two tools born from that Zoom: Glaze and Nightshade, which were developed by Zhao and the University of Chicago’s SAND Lab (an acronym for “security, algorithms, networking, and data”).

Arguably the most prominent weapons in an artist’s arsenal against nonconsensual AI scraping, Glaze and Nightshade work in similar ways: by adding what the researchers call “barely perceptible” perturbations to an image’s pixels so that machine-learning models cannot read them properly. Glaze, which has been downloaded more than 6 million times since it launched in March 2023, adds what’s effectively a secret cloak to images that prevents AI algorithms from picking up on and copying an artist’s style. Nightshade, which I wrote about when it was released almost exactly a year ago this fall, cranks up the offensive against AI companies by adding an invisible layer of poison to images, which can break AI models; it has been downloaded more than 1.6 million times. 

Thanks to the tools, “I’m able to post my work online,” Ortiz says, “and that’s pretty huge.” For artists like her, being seen online is crucial to getting more work. If they are uncomfortable about ending up in a massive for-profit AI model without compensation, the only option is to delete their work from the internet. That would mean career suicide. “It’s really dire for us,” adds Ortiz, who has become one of the most vocal advocates for fellow artists and is part of a class action lawsuit against AI companies, including Stability AI, over copyright infringement. 

But Zhao hopes that the tools will do more than empower individual artists. Glaze and Nightshade are part of what he sees as a battle to slowly tilt the balance of power from large corporations back to individual creators. 

“It is just incredibly frustrating to see human life be valued so little,” he says with a disdain that I’ve come to see as pretty typical for him, particularly when he’s talking about Big Tech. “And to see that repeated over and over, this prioritization of profit over humanity … it is just incredibly frustrating and maddening.” 

As the tools are adopted more widely, his lofty goal is being put to the test. Can Glaze and Nightshade make genuine security accessible for creators—or will they inadvertently lull artists into believing their work is safe, even as the tools themselves become targets for haters and hackers? While experts largely agree that the approach is effective and Nightshade could prove to be powerful poison, other researchers claim they’ve already poked holes in the protections offered by Glaze and that trusting these tools is risky. 

But Neil Turkewitz, a copyright lawyer who used to work at the Recording Industry Association of America, offers a more sweeping view of the fight the SAND Lab has joined. It’s not about a single AI company or a single individual, he says: “It’s about defining the rules of the world we want to inhabit.” 

Poking the bear

The SAND Lab is tight knit, encompassing a dozen or so researchers crammed into a corner of the University of Chicago’s computer science building. That space has accumulated somewhat typical workplace detritus—a Meta Quest headset here, silly photos of dress-up from Halloween parties there. But the walls are also covered in original art pieces, including a framed painting by Ortiz.  

Years before fighting alongside artists like Ortiz against “AI bros” (to use Zhao’s words), Zhao and the lab’s co-leader, Heather Zheng, who is also his wife, had built a record of combating harms posed by new tech. 

group of students and teachers posing in Halloween costumes
When I visited the SAND Lab in Chicago, I saw how tight knit the group was. Alongside the typical workplace stuff were funny Halloween photos like this one. (Front row: Ronik Bhaskar, Josephine Passananti, Anna YJ Ha, Zhuolin Yang, Ben Zhao, Heather Zheng. Back row: Cathy Yuanchen Li, Wenxin Ding, Stanley Wu, and Shawn Shan.)
COURTESY OF SAND LAB

Though both earned spots on MIT Technology Review’s 35 Innovators Under 35 list for other work nearly two decades ago, when they were at the University of California, Santa Barbara (Zheng in 2005 for “cognitive radios” and Zhao a year later for peer-to-peer networks), their primary research focus has become security and privacy. 

The pair left Santa Barbara in 2017, after they were poached by the new co-director of the University of Chicago’s Data Science Institute, Michael Franklin. All eight PhD students from their UC Santa Barbara lab decided to follow them to Chicago too. Since then, the group has developed a “bracelet of silence” that jams the microphones in AI voice assistants like the Amazon Echo. It has also created a tool called Fawkes—“privacy armor,” as Zhao put it in a 2020 interview with the New York Times—that people can apply to their photos to protect them from facial recognition software. They’ve also studied how hackers might steal sensitive information through stealth attacks on virtual-reality headsets, and how to distinguish human art from AI-generated images. 

“Ben and Heather and their group are kind of unique because they’re actually trying to build technology that hits right at some key questions about AI and how it is used,” Franklin tells me. “They’re doing it not just by asking those questions, but by actually building technology that forces those questions to the forefront.”

It was Fawkes that intrigued Van Deun, the fantasy illustrator, two years ago; she hoped something similar might work as protection against generative AI, which is why she extended that fateful invite to the Concept Art Association’s Zoom call. 

That call started something of a mad rush in the weeks that followed. Though Zhao and Zheng collaborate on all the lab’s projects, they each lead individual initiatives; Zhao took on what would become Glaze, with PhD student Shawn Shan (who was on this year’s Innovators Under 35 list) spearheading the development of the program’s algorithm. 

In parallel to Shan’s coding, PhD students Jenna Cryan and Emily Wenger sought to learn more about the views and needs of the artists themselves. They created a user survey that the team distributed to artists with the help of Ortiz. In replies from more than 1,200 artists—far more than the average number of responses to user studies in computer science—the team found that the vast majority of creators had read about art being used to train models, and 97% expected AI to decrease some artists’ job security. A quarter said AI art had already affected their jobs. 

Almost all artists also said they posted their work online, and more than half said they anticipated reducing or removing that online work, if they hadn’t already—no matter the professional and financial consequences.

The first scrappy version of Glaze was developed in just a month, at which point Ortiz gave the team her entire catalogue of work to test the model on. At the most basic level, Glaze acts as a defensive shield. Its algorithm identifies features from the image that make up an artist’s individual style and adds subtle changes to them. When an AI model is trained on images protected with Glaze, the model will not be able to reproduce styles similar to the original image. 

A painting from Ortiz later became the first image publicly released with Glaze on it: a young woman, surrounded by flying eagles, holding up a wreath. Its title is Musa Victoriosa, “victorious muse.” 

It’s the one currently hanging on the SAND Lab’s walls. 

Despite many artists’ initial enthusiasm, Zhao says, Glaze’s launch caused significant backlash. Some artists were skeptical because they were worried this was a scam or yet another data-harvesting campaign. 

The lab had to take several steps to build trust, such as offering the option to download the Glaze app so that it adds the protective layer offline, which meant no data was being transferred anywhere. (The images are then shielded when artists upload them.)  

Soon after Glaze’s launch, Shan also led the development of the second tool, Nightshade. Where Glaze is a defensive mechanism, Nightshade was designed to act as an offensive deterrent to nonconsensual training. It works by changing the pixels of images in ways that are not noticeable to the human eye but manipulate machine-learning models so they interpret the image as something different from what it actually shows. If poisoned samples are scraped into AI training sets, these samples trick the AI models: Dogs become cats, handbags become toasters. The researchers say only a relatively few examples are enough to permanently damage the way a generative AI model produces images.

Currently, both tools are available as free apps or can be applied through the project’s website. The lab has also recently expanded its reach by offering integration with the new artist-supported social network Cara, which was born out of a backlash to exploitative AI training and forbids AI-produced content.

In dozens of conversations with Zhao and the lab’s researchers, as well as a handful of their artist-collaborators, it’s become clear that both groups now feel they are aligned in one mission. “I never expected to become friends with scientists in Chicago,” says Eva Toorenent, a Dutch artist who worked closely with the team on Nightshade. “I’m just so happy to have met these people during this collective battle.” 

Belladonna artwork shows a central character with a skull head in a dark forest illuminated around them by the belladonna flower slung over their shoulder
Images online of Toorenent’s Belladonna have been treated with the SAND Lab’s Nightshade tool.
EVA TOORENENT

Her painting Belladonna, which is also another name for the nightshade plant, was the first image with Nightshade’s poison on it. 

“It’s so symbolic,” she says. “People taking our work without our consent, and then taking our work without consent can ruin their models. It’s just poetic justice.” 

No perfect solution

The reception of the SAND Lab’s work has been less harmonious across the AI community.

After Glaze was made available to the public, Zhao tells me, someone reported it to sites like VirusTotal, which tracks malware, so that it was flagged by antivirus programs. Several people also started claiming on social media that the tool had quickly been broken. Nightshade similarly got a fair share of criticism when it launched; as TechCrunch reported in January, some called it a “virus” and, as the story explains, “another Reddit user who inadvertently went viral on X questioned Nightshade’s legality, comparing it to ‘hacking a vulnerable computer system to disrupt its operation.’” 

“We had no idea what we were up against,” Zhao tells me. “Not knowing who or what the other side could be meant that every single new buzzing of the phone meant that maybe someone did break Glaze.” 

Both tools, though, have gone through rigorous academic peer review and have won recognition from the computer security community. Nightshade was accepted at the IEEE Symposium on Security and Privacy, and Glaze received a distinguished paper award and the 2023 Internet Defense Prize at the Usenix Security Symposium, a top conference in the field. 

“In my experience working with poison, I think [Nightshade is] pretty effective,” says Nathalie Baracaldo, who leads the AI security and privacy solutions team at IBM and has studied data poisoning. “I have not seen anything yet—and the word yet is important here—that breaks that type of defense that Ben is proposing.” And the fact that the team has released the source code for Nightshade for others to probe, and it hasn’t been broken, also suggests it’s quite secure, she adds. 

At the same time, at least one team of researchers does claim to have penetrated the protections of Glaze, or at least an old version of it. 

As researchers from Google DeepMind and ETH Zurich detailed in a paper published in June, they found various ways Glaze (as well as similar but less popular protection tools, such as Mist and Anti-DreamBooth) could be circumvented using off-the-shelf techniques that anyone could access—such as image upscaling, meaning filling in pixels to increase the resolution of an image as it’s enlarged. The researchers write that their work shows the “brittleness of existing protections” and warn that “artists may believe they are effective. But our experiments show they are not.”

Florian Tramèr, an associate professor at ETH Zurich who was part of the study, acknowledges that it is “very hard to come up with a strong technical solution that ends up really making a difference here.” Rather than any individual tool, he ultimately advocates for an almost certainly unrealistic ideal: stronger policies and laws to help create an environment in which people commit to buying only human-created art. 

What happened here is common in security research, notes Baracaldo: A defense is proposed, an adversary breaks it, and—ideally—the defender learns from the adversary and makes the defense better. “It’s important to have both ethical attackers and defenders working together to make our AI systems safer,” she says, adding that “ideally, all defenses should be publicly available for scrutiny,” which would both “allow for transparency” and help avoid creating a false sense of security. (Zhao, though, tells me the researchers have no intention to release Glaze’s source code.)

Still, even as all these researchers claim to support artists and their art, such tests hit a nerve for Zhao. In Discord chats that were later leaked, he claimed that one of the researchers from the ETH Zurich–Google DeepMind team “doesn’t give a shit” about people. (That researcher did not respond to a request for comment, but in a blog post he said it was important to break defenses in order to know how to fix them. Zhao says his words were taken out of context.) 

Zhao also emphasizes to me that the paper’s authors mainly evaluated an earlier version of Glaze; he says its new update is more resistant to tampering. Messing with images that have current Glaze protections would harm the very style that is being copied, he says, making such an attack useless. 

This back-and-forth reflects a significant tension in the computer security community and, more broadly, the often adversarial relationship between different groups in AI. Is it wrong to give people the feeling of security when the protections you’ve offered might break? Or is it better to have some level of protection—one that raises the threshold for an attacker to inflict harm—than nothing at all? 

Yves-Alexandre de Montjoye, an associate professor of applied mathematics and computer science at Imperial College London, says there are plenty of examples where similar technical protections have failed to be bulletproof. For example, in 2023, de Montjoye and his team probed a digital mask for facial recognition algorithms, which was meant to protect the privacy of medical patients’ facial images; they were able to break the protections by tweaking just one thing in the program’s algorithm (which was open source). 

Using such defenses is still sending a message, he says, and adding some friction to data profiling. “Tools such as TrackMeNot”—which protects users from data profiling—“have been presented as a way to protest; as a way to say I do not consent.”  

“But at the same time,” he argues, “we need to be very clear with artists that it is removable and might not protect against future algorithms.”

While Zhao will admit that the researchers pointed out some of Glaze’s weak spots, he unsurprisingly remains confident that Glaze and Nightshade are worth deploying, given that “security tools are never perfect.” Indeed, as Baracaldo points out, the Google DeepMind and ETH Zurich researchers showed how a highly motivated and sophisticated adversary will almost certainly always find a way in.

Yet it is “simplistic to think that if you have a real security problem in the wild and you’re trying to design a protection tool, the answer should be it either works perfectly or don’t deploy it,” Zhao says, citing spam filters and firewalls as examples. Defense is a constant cat-and-mouse game. And he believes most artists are savvy enough to understand the risk. 

Offering hope

The fight between creators and AI companies is fierce. The current paradigm in AI is to build bigger and bigger models, and there is, at least currently, no getting around the fact that they require vast data sets hoovered from the internet to train on. Tech companies argue that anything on the public internet is fair game, and that it is “impossible” to build advanced AI tools without copyrighted material; many artists argue that tech companies have stolen their intellectual property and violated copyright law, and that they need ways to keep their individual works out of the models—or at least receive proper credit and compensation for their use. 

So far, the creatives aren’t exactly winning. A number of companies have already replaced designers, copywriters, and illustrators with AI systems. In one high-profile case, Marvel Studios used AI-generated imagery instead of human-created art in the title sequence of its 2023 TV series Secret Invasion. In another, a radio station fired its human presenters and replaced them with AI. The technology has become a major bone of contention between unions and film, TV, and creative studios, most recently leading to a strike by video-game performers. There are numerous ongoing lawsuits by artists, writers, publishers, and record labels against AI companies. It will likely take years until there is a clear-cut legal resolution. But even a court ruling won’t necessarily untangle the difficult ethical questions created by generative AI. Any future government regulation is not likely to either, if it ever materializes. 

That’s why Zhao and Zheng see Glaze and Nightshade as necessary interventions—tools to defend original work, attack those who would help themselves to it, and, at the very least, buy artists some time. Having a perfect solution is not really the point. The researchers need to offer something now because the AI sector moves at breakneck speed, Zheng says, means that companies are ignoring very real harms to humans. “This is probably the first time in our entire technology careers that we actually see this much conflict,” she adds.

On a much grander scale, she and Zhao tell me they hope that Glaze and Nightshade will eventually have the power to overhaul how AI companies use art and how their products produce it. It is eye-wateringly expensive to train AI models, and it’s extremely laborious for engineers to find and purge poisoned samples in a data set of billions of images. Theoretically, if there are enough Nightshaded images on the internet and tech companies see their models breaking as a result, it could push developers to the negotiating table to bargain over licensing and fair compensation. 

That’s, of course, still a big “if.” MIT Technology Review reached out to several AI companies, such as Midjourney and Stability AI, which did not reply to requests for comment. A spokesperson for OpenAI, meanwhile, did not confirm any details about encountering data poison but said the company takes the safety of its products seriously and is continually improving its safety measures: “We are always working on how we can make our systems more robust against this type of abuse.”

In the meantime, the SAND Lab is moving ahead and looking into funding from foundations and nonprofits to keep the project going. They also say there has also been interest from major companies looking to protect their intellectual property (though they decline to say which), and Zhao and Zheng are exploring how the tools could be applied in other industries, such as gaming, videos, or music. In the meantime, they plan to keep updating Glaze and Nightshade to be as robust as possible, working closely with the students in the Chicago lab—where, on another wall, hangs Toorenent’s Belladonna. The painting has a heart-shaped note stuck to the bottom right corner: “Thank you! You have given hope to us artists.”

This story has been updated with the latest download figures for Glaze and Nightshade.

Google DeepMind has a new way to look inside an AI’s “mind”

AI has led to breakthroughs in drug discovery and robotics and is in the process of entirely revolutionizing how we interact with machines and the web. The only problem is we don’t know exactly how it works, or why it works so well. We have a fair idea, but the details are too complex to unpick. That’s a problem: It could lead us to deploy an AI system in a highly sensitive field like medicine without understanding that it could have critical flaws embedded in its workings.

A team at Google DeepMind that studies something called mechanistic interpretability has been working on new ways to let us peer under the hood. At the end of July, it released Gemma Scope, a tool to help researchers understand what is happening when AI is generating an output. The hope is that if we have a better understanding of what is happening inside an AI model, we’ll be able to control its outputs more effectively, leading to better AI systems in the future.

“I want to be able to look inside a model and see if it’s being deceptive,” says Neel Nanda, who runs the mechanistic interpretability team at Google DeepMind. “It seems like being able to read a model’s mind should help.”

Mechanistic interpretability, also known as “mech interp,” is a new research field that aims to understand how neural networks actually work. At the moment, very basically, we put inputs into a model in the form of a lot of data, and then we get a bunch of model weights at the end of training. These are the parameters that determine how a model makes decisions. We have some idea of what’s happening between the inputs and the model weights: Essentially, the AI is finding patterns in the data and making conclusions from those patterns, but these patterns can be incredibly complex and often very hard for humans to interpret.

It’s like a teacher reviewing the answers to a complex math problem on a test. The student—the AI, in this case—wrote down the correct answer, but the work looks like a bunch of squiggly lines. This example assumes the AI is always getting the correct answer, but that’s not always true; the AI student may have found an irrelevant pattern that it’s assuming is valid. For example, some current AI systems will give you the result that 9.11 is bigger than 9.8. Different methods developed in the field of mechanistic interpretability are beginning to shed a little bit of light on what may be happening, essentially making sense of the squiggly lines.

“A key goal of mechanistic interpretability is trying to reverse-engineer the algorithms inside these systems,” says Nanda. “We give the model a prompt, like ‘Write a poem,’ and then it writes some rhyming lines. What is the algorithm by which it did this? We’d love to understand it.”

To find features—or categories of data that represent a larger concept—in its AI model, Gemma, DeepMind ran a tool known as a “sparse autoencoder” on each of its layers. You can think of a sparse autoencoder as a microscope that zooms in on those layers and lets you look at their details. For example, if you prompt Gemma about a chihuahua, it will trigger the “dogs” feature, lighting up what the model knows about “dogs.” The reason it is considered “sparse” is that it’s limiting the number of neurons used, basically pushing for a more efficient and generalized representation of the data.

The tricky part of sparse autoencoders is deciding how granular you want to get. Think again about the microscope. You can magnify something to an extreme degree, but it may make what you’re looking at impossible for a human to interpret. But if you zoom too far out, you may be limiting what interesting things you can see and discover. 

DeepMind’s solution was to run sparse autoencoders of different sizes, varying the number of features they want the autoencoder to find. The goal was not for DeepMind’s researchers to thoroughly analyze the results on their own. Gemma and the autoencoders are open-source, so this project was aimed more at spurring interested researchers to look at what the sparse autoencoders found and hopefully make new insights into the model’s internal logic. Since DeepMind ran autoencoders on each layer of their model, a researcher could map the progression from input to output to a degree we haven’t seen before.

“This is really exciting for interpretability researchers,” says Josh Batson, a researcher at Anthropic. “If you have this model that you’ve open-sourced for people to study, it means that a bunch of interpretability research can now be done on the back of those sparse autoencoders. It lowers the barrier to entry to people learning from these methods.”

Neuronpedia, a platform for mechanistic interpretability, partnered with DeepMind in July to build a demo of Gemma Scope that you can play around with right now. In the demo, you can test out different prompts and see how the model breaks up your prompt and what activations your prompt lights up. You can also mess around with the model. For example, if you turn the feature about dogs way up and then ask the model a question about US presidents, Gemma will find some way to weave in random babble about dogs, or the model may just start barking at you.

One interesting thing about sparse autoencoders is that they are unsupervised, meaning they find features on their own. That leads to surprising discoveries about how the models break down human concepts. “My personal favorite feature is the cringe feature,” says Joseph Bloom, science lead at Neuronpedia. “It seems to appear in negative criticism of text and movies. It’s just a great example of tracking things that are so human on some level.” 

You can search for concepts on Neuronpedia and it will highlight what features are being activated on specific tokens, or words, and how strongly each one is activated. “If you read the text and you see what’s highlighted in green, that’s when the model thinks the cringe concept is most relevant. The most active example for cringe is somebody preaching at someone else,” says Bloom.

Some features are proving easier to track than others. “One of the most important features that you would want to find for a model is deception,” says Johnny Lin, founder of Neuronpedia. “It’s not super easy to find: ‘Oh, there’s the feature that fires when it’s lying to us.’ From what I’ve seen, it hasn’t been the case that we can find deception and ban it.”

DeepMind’s research is similar to what another AI company, Anthropic, did back in May with Golden Gate Claude. It used sparse autoencoders to find the parts of Claude, their model, that lit up when discussing the Golden Gate Bridge in San Francisco. It then amplified the activations related to the bridge to the point where Claude literally identified not as Claude, an AI model, but as the physical Golden Gate Bridge and would respond to prompts as the bridge.

Although it may just seem quirky, mechanistic interpretability research may prove incredibly useful. “As a tool for understanding how the model generalizes and what level of abstraction it’s working at, these features are really helpful,” says Batson.

For example, a team lead by Samuel Marks, now at Anthropic, used sparse autoencoders to find features that showed a particular model was associating certain professions with a specific gender. They then turned off these gender features to reduce bias in the model. This experiment was done on a very small model, so it’s unclear if the work will apply to a much larger model.

Mechanistic interpretability research can also give us insights into why AI makes errors. In the case of the assertion that 9.11 is larger than 9.8, researchers from Transluce saw that the question was triggering the parts of an AI model related to Bible verses and September 11. The researchers concluded the AI could be interpreting the numbers as dates, asserting the later date, 9/11, as greater than 9/8. And in a lot of books like religious texts, section 9.11 comes after section 9.8, which may be why the AI thinks of it as greater. Once they knew why the AI made this error, the researchers tuned down the AI’s activations on Bible verses and September 11, which led to the model giving the correct answer when prompted again on whether 9.11 is larger than 9.8.

There are also other potential applications. Currently, a system-level prompt is built into LLMs to deal with situations like users who ask how to build a bomb. When you ask ChatGPT a question, the model is first secretly prompted by OpenAI to refrain from telling you how to make bombs or do other nefarious things. But it’s easy for users to jailbreak AI models with clever prompts, bypassing any restrictions. 

If the creators of the models are able to see where in an AI the bomb-building knowledge is, they can theoretically turn off those nodes permanently. Then even the most cleverly written prompt wouldn’t elicit an answer about how to build a bomb, because the AI would literally have no information about how to build a bomb in its system.

This type of granularity and precise control are easy to imagine but extremely hard to achieve with the current state of mechanistic interpretability. 

“A limitation is the steering [influencing a model by adjusting its parameters] is just not working that well, and so when you steer to reduce violence in a model, it ends up completely lobotomizing its knowledge in martial arts. There’s a lot of refinement to be done in steering,” says Lin. The knowledge of “bomb making,” for example, isn’t just a simple on-and-off switch in an AI model. It most likely is woven into multiple parts of the model, and turning it off would probably involve hampering the AI’s knowledge of chemistry. Any tinkering may have benefits but also significant trade-offs.

That said, if we are able to dig deeper and peer more clearly into the “mind” of AI, DeepMind and others are hopeful that mechanistic interpretability could represent a plausible path to alignment—the process of making sure AI is actually doing what we want it to do.

What’s on the table at this year’s UN climate conference

This article is from The Spark, MIT Technology Review’s weekly climate newsletter. To receive it in your inbox every Wednesday, sign up here.

It’s time for a party—the Conference of the Parties, that is. Talks kicked off this week at COP29 in Baku, Azerbaijan. Running for a couple of weeks each year, the global summit is the largest annual meeting on climate change.

The issue on the table this time around: Countries need to agree to set a new goal on how much money should go to developing countries to help them finance the fight against climate change. Complicating things? A US president-elect whose approach to climate is very different from that of the current administration (understatement of the century).

This is a big moment that could set the tone for what the next few years of the international climate world looks like. Here’s what you need to know about COP29 and how Donald Trump’s election is coloring things.

The UN COP meetings are an annual chance for nearly 200 nations to get together to discuss (and hopefully act on) climate change. Greatest hits from the talks include the Paris Agreement, a 2015 global accord that set a goal to limit global warming to 1.5 °C (2.7 °F) above preindustrial levels.

This year, the talks are in Azerbaijan, a petrostate if there ever was one. Oil and gas production makes up over 90% of the country’s export revenue and nearly half its GDP as of 2022. A perfectly ironic spot for a global climate summit!

The biggest discussion this year centers on global climate finance—specifically, how much of it is needed to help developing countries address climate change and adapt to changing conditions. The current goal, set in 2009, is for industrialized countries to provide $100 billion each year to developing nations. The deadline was 2020, and that target was actually met for the first time in 2022, according to the Organization for Economic Cooperation and Development, which keeps track of total finance via reports from contributing countries. Currently, most of that funding is in the form of public loans and grants.

The thing is, that $100 billion number was somewhat arbitrary—in Paris in 2015, countries agreed that a new, larger target should be set in 2025 to take into account how much countries actually need.

It’s looking as if the magic number is somewhere around $1 trillion each year. However, it remains to be seen how this goal will end up shaking out, because there are disagreements about basically every part of this. What should the final number be? What kind of money should count—just public funds, or private investments as well? Which nations should pay? How long will this target stand? What, exactly, would this money be going toward?

Working out all those details is why nations are gathering right now. But one shadow looming over these negotiations is the impending return of Donald Trump.

As I covered last week, Trump’s election will almost certainly result in less progress on cutting emissions than we might have seen under a more climate-focused administration. But arguably an even bigger deal than domestic progress (or lack thereof) will be how Trump shifts the country’s climate position on the international stage.

The US has emitted more carbon pollution into the atmosphere than any other country, it currently leads the world in per capita emissions, and it’s the world’s richest economy. If anybody should be a leader at the table in talks about climate finance, it’s the US. And yet, Trump is coming into power soon, and we’ve all seen this film before. 

Last time Trump was in office, he pulled the US out of the Paris Agreement. He’s made promises to do it again—and could go one step further by backing out of the UN Framework Convention on Climate Change (UNFCCC) altogether. If leaving the Paris Agreement is walking away from the table, withdrawing from the UNFCCC is like hopping on a rocket and blasting in a different direction. It’s a more drastic action and could be tougher to reverse in the future, though experts also aren’t sure if Trump could technically do this on his own.

The uncertainty of what happens next in the US is a cloud hanging over these negotiations. “This is going to be harder because we don’t have a dynamic and pushy and confident US helping us on climate action,” said Camilla Born, an independent climate advisor and former UK senior official at COP26, during an online event last week hosted by Carbon Brief.

Some experts are confident that others will step up to fill the gap. “There are many drivers of climate action beyond the White House,” said Mohamed Adow, founding director of Power Shift Africa, at the CarbonBrief event.

If I could characterize the current vibe in the climate world, it’s uncertainty. But the negotiations over the next couple of weeks could provide clues to what we can expect for the next few years. Just how much will a Trump presidency slow global climate action? Will the European Union step up? Could this cement the rise of China as a climate leader? We’ll be watching it all.


Now read the rest of The Spark

Related reading

In case you want some additional context from the last few years of these meetings, here’s my coverage of last year’s fight at COP28 over a transition away from fossil fuels, and a newsletter about negotiations over the “loss and damages” fund at COP27.

For the nitty-gritty details about what’s on the table at COP29, check out this very thorough explainer from Carbon Brief.

The White House in Washington DC under dark stormy clouds

DAN THORNBERG/ADOBE STOCK

Another thing

Trump’s election will have significant ripple effects across the economy and our lives. His victory is a tragic loss for climate progress, as my colleague James Temple wrote in an op-ed last week. Give it a read, if you haven’t already, to dig into some of the potential impacts we might see over the next four years and beyond. 

Keeping up with climate  

The US Environmental Protection Agency finalized a rule to fine oil and gas companies for methane emissions. The fee was part of the Inflation Reduction Act of 2022. (Associated Press)
→ This rule faces a cloudy future under the Trump administration; industry groups are already talking about repealing it. (NPR)

Speaking of the EPA, Donald Trump chose Lee Zeldin, a former Republican congressman from New York, to lead the agency. Zeldin isn’t particularly known for climate or economic policy. (New York Times)

Oil giant BP is scaling back its early-stage hydrogen projects. The company revealed in an earnings report that it’s canceling 18 such projects and currently plans to greenlight between five and 10. (TechCrunch)

Investors betting against renewable energy scored big last week, earning nearly $1.2 billion as stocks in that sector tumbled. (Financial Times)

Lithium iron phosphate batteries are taking over the world, or at least electric vehicles. These lithium-ion batteries are cheaper and longer-lasting than their nickel-containing cousins, though they also tend to be heavier. (Canary Media
→ I wrote about this trend last year in a newsletter about batteries and their ingredients. (MIT Technology Review)

The US unveiled plans to triple its nuclear energy capacity by 2050. That’s an additional 200 gigawatts’ worth of consistently available power. (Bloomberg)

Five subsea cables that can help power millions of homes just got the green light in Great Britain. The projects will help connect the island to other power grids, as well as to offshore wind farms in Dutch and Belgian waters. (The Guardian)

How this grassroots effort could make AI voices more diverse

We are on the cusp of a voice AI boom, with tech companies such as Apple and OpenAI rolling out the next generation of artificial-intelligence-powered assistants. But the default voices for these assistants are often white American—British, if you’re lucky—and most definitely speak English. They represent only a tiny proportion of the many dialects and accents in the English language, which spans many regions and cultures. And if you’re one of the billions of people who don’t speak English, bad luck: These tools don’t sound nearly as good in other languages.

This is because the data that has gone into training these models is limited. In AI research, most data used to train models is extracted from the English-language internet, which reflects Anglo-American culture. But there is a massive grassroots effort underway to change this status quo and bring more transparency and diversity to what AI sounds like: Mozilla’s Common Voice initiative. 

The data set Common Voice has created over the past seven years is one of the most useful resources for people wanting to build voice AI. It has seen a massive spike in downloads, partly thanks to the current AI boom; it recently hit the 5 million mark, up from 38,500 in 2020. Creating this data set has not been easy, mainly because the data collection relies on an army of volunteers. Their numbers have also jumped, from just under 500,000 in 2020 to over 900,000 in 2024. But by giving its data away, some members of this community argue, Mozilla is encouraging volunteers to effectively do free labor for Big Tech. 

Since 2017, volunteers for the Common Voice project have collected a total of 31,000 hours of voice data in around 180 languages as diverse as Russian, Catalan, and Marathi. If you’ve used a service that uses audio AI, it’s likely been trained at least partly on Common Voice. 

Mozilla’s cause is a noble one. As AI is integrated increasingly into our lives and the ways we communicate, it becomes more important that the tools we interact with sound like us. The technology could break down communication barriers and help convey information in a compelling way to, for example, people who can’t read. But instead, an intense focus on English risks entrenching a new colonial world order and wiping out languages entirely.

“It would be such an own goal if, rather than finally creating truly multimodal, multilingual, high-performance translation models and making a more multilingual world, we actually ended up forcing everybody to operate in, like, English or French,” says EM Lewis-Jong, a director for Common Voice. 

Common Voice is open source, which means anyone can see what has gone into the data set, and users can do whatever they want with it for free. This kind of transparency is unusual in AI data governance. Most large audio data sets simply aren’t publicly available, and many consist of data that has been scraped from sites like YouTube, according to research conducted by a team from the University of Washington, and Carnegie Mellon andNorthwestern universities. 

The vast majority of language data is collected by volunteers such as Bülent Özden, a researcher from Turkey. Since 2020, he has been not only donating his voice but also raising awareness around the project to get more people to donate. He recently spent two full-time months correcting data and checking for typos in Turkish. For him, improving AI models is not the only motivation to do this work. 

“I’m doing it to preserve cultures, especially low-resource [languages],” Özden says. He tells me he has recently started collecting samples of Turkey’s smaller languages, such as Circassian and Zaza.

However, as I dug into the data set, I noticed that the coverage of languages and accents is very uneven. There are only 22 hours of Finnish voices from 231 people. In comparison, the data set contains 3,554 hours of English from 94,665 speakers. Some languages, such as Korean and Punjabi, are even less well represented. Even though they have tens of millions of speakers, they account for only a couple of hours of recorded data. 

This imbalance has emerged because data collection efforts are started from the bottom up by language communities themselves, says Lewis-Jong. 

“We’re trying to give communities what they need to create their own AI training data sets. We have a particular focus on doing this for language communities where there isn’t any data, or where maybe larger tech organizations might not be that interested in creating those data sets,” Lewis-Jong says. They hope that with the help of volunteers and various bits of grant funding, the Common Voice data set will have close to 200 languages by the end of the year.

Common Voice’s permissive license means that many companies rely on it—for example, the Swedish startup Mabel AI, which builds translation tools for health-care providers. One of the first languages the company used was Ukrainian; it built a translation tool to help Ukrainian refugees interact with Swedish social services, says Karolina Sjöberg, Mabel AI’s founder and CEO. The team has since expanded to other languages, such as Arabic and Russian. 

The problem with a lot of other audio data is that it consists of people reading from books or texts. The result is very different from how people really speak, especially when they are distressed or in pain, Sjöberg says. Because anyone can submit sentences to Common Voice for others to read aloud, Mozilla’s data set also includes sentences that are more colloquial and feel more natural, she says.

Not that it is perfectly representative. The Mabel AI team soon found out that most voice data in the languages it needed was donated by younger men, which is fairly typical for the data set. 

“The refugees that we intended to use the app with were really anything but younger men,” Sjöberg says. “So that meant that the voice data that we needed did not quite match the voice data that we had.” The team started collecting its own voice data from Ukrainian women, as well as from elderly people. 

Unlike other data sets, Common Voice asks participants to share their gender and details about their accent. Making sure different genders are represented is important to fight bias in AI models, says Rebecca Ryakitimbo, a Common Voice fellow who created the project’s gender action plan. More diversity leads not only to better representation but also to better models. Systems that are trained on narrow and homogenous data tend to spew stereotyped and harmful results.

“We don’t want a case where we have a chatbot that is named after a woman but does not give the same response to a woman as it would a man,” she says. 

Ryakitimbo has collected voice data in Kiswahili in Tanzania, Kenya, and the Democratic Republic of Congo. She tells me she wanted to collect voices from a socioeconomically diverse set of Kiswahili speakers and has reached out to women young and old living in rural areas, who might not always be literate or even have access to devices. 

This kind of data collection is challenging. The importance of collecting AI voice data can feel abstract to many people, especially if they aren’t familiar with the technologies. Ryakitimbo and volunteers would approach women in settings where they felt safe to begin with, such as presentations on menstrual hygiene, and explain how the technology could, for example, help disseminate information about menstruation. For women who did not know how to read, the team read out sentences that they would repeat for the recording. 

The Common Voice project is bolstered by the belief that languages form a really important part of identity. “We think it’s not just about language, but about transmitting culture and heritage and treasuring people’s particular cultural context,” says Lewis-Jong. “There are all kinds of idioms and cultural catchphrases that just don’t translate,” they add. 

Common Voice is the only audio data set where English doesn’t dominate, says Willie Agnew, a researcher at Carnegie Mellon University who has studied audio data sets. “I’m very impressed with how well they’ve done that and how well they’ve made this data set that is actually pretty diverse,” Agnew says. “It feels like they’re way far ahead of almost all the other projects we looked at.” 

I spent some time verifying the recordings of other Finnish speakers on the Common Voice platform. As their voices echoed in my study, I felt surprisingly touched. We had all gathered around the same cause: making AI data more inclusive, and making sure our culture and language was properly represented in the next generation of AI tools. 

But I had some big questions about what would happen to my voice if I donated it. Once it was in the data set, I would have no control about how it might be used afterwards. The tech sector isn’t exactly known for giving people proper credit, and the data is available for anyone’s use. 

“As much as we want it to benefit the local communities, there’s a possibility that also Big Tech could make use of the same data and build something that then comes out as the commercial product,” says Ryakitimbo. Though Mozilla does not share who has downloaded Common Voice, Lewis-Jong tells me Meta and Nvidia have said that they have used it.

Open access to this hard-won and rare language data is not something all minority groups want, says Harry H. Jiang, a researcher at Carnegie Mellon University, who was part of the team doing audit research. For example, Indigenous groups have raised concerns. 

“Extractivism” is something that Mozilla has been thinking about a lot over the past 18 months, says Lewis-Jong. Later this year the project will work with communities to pilot alternative licenses including Nwulite Obodo Open Data License, which was created by researchers at the University of Pretoria for sharing African data sets more equitably. For example, people who want to download the data might be asked to write a request with details on how they plan to use it, and they might be allowed to license it only for certain products or for a limited time. Users might also be asked to contribute to community projects that support poverty reduction, says Lewis-Jong.  

Lewis-Jong says the pilot is a learning exercise to explore whether people will want data with alternative licenses, and whether they are sustainable for communities managing them. The hope is that it could lead to something resembling “open source 2.0.”

In the end, I decided to donate my voice. I received a list of phrases to say, sat in front of my computer, and hit Record. One day, I hope, my effort will help a company or researcher build voice AI that sounds less generic, and more like me. 

This story has been updated.

Why the term “women of childbearing age” is problematic

This article first appeared in The Checkup, MIT Technology Review’s weekly biotech newsletter. To receive it in your inbox every Thursday, and read articles like this first, sign up here.

Every journalist has favorite topics. Regular Checkup readers might already know some of mine, which include the quest to delay or reverse human aging, and new technologies for reproductive health and fertility. So when I saw trailers for The Substance, a film centered on one middle-aged woman’s attempt to reexperience youth, I had to watch it.

I won’t spoil the movie for anyone who hasn’t seen it yet (although I should warn that it is not for the squeamish, or anyone with an aversion to gratuitous close-ups of bums and nipples). But a key premise of the film involves harmful attitudes toward female aging.

“Hey, did you know that a woman’s fertility starts to decrease by the age of 25?” a powerful male character asks early in the film. “At 50, it just stops,” he later adds. He never explains what stops, exactly, but to the viewer the message is pretty clear: If you’re a woman, your worth is tied to your fertility. Once your fertile window is over, so are you.

The insidious idea that women’s bodies are, above all else, vessels for growing children has plenty of negative consequences for us all. But it has also set back scientific research and health policy.

Earlier this week, I chatted about this with Alana Cattapan, a political scientist at the University of Waterloo in Ontario, Canada. Cattapan has been exploring the concept of “women of reproductive age”—a descriptor that is ubiquitous in health research and policy.

The idea for the research project came to her when the Zika virus was making headlines around eight years ago. “I was planning on going to the Caribbean for a trip related to my partner’s research, and I kept getting advice that women of reproductive age shouldn’t go,” she told me. At the time, Zika was being linked to microcephaly—unusually small heads—in newborn babies. It was thought that the virus was affecting key stages of fetal development.

Cattapan wasn’t pregnant. And she wasn’t planning on becoming pregnant at the time. So why was she being advised to stay away from areas with the virus?

The experience got her thinking about the ways in which attitudes toward our bodies are governed by the idea of potential pregnancy. Take, for example, biomedical research on the causes and treatment of disease. Women’s health has lagged behind men’s as a focus of such work, for multiple reasons. Male bodies have long been considered the “default” human form, for example. And clinical trials have historically been designed in ways that make them less accessible for women.

Fears about the potential effects of drugs on fetuses have also played a significant role in keeping people who have the potential to become pregnant out of studies. “Scientific research has excluded women of ‘reproductive age,’ or women who might potentially conceive, in a blanket way,” says Cattapan. “The research that we have on many, many drugs does not include women and certainly doesn’t include women in pregnancy.”  

This lack of research goes some way to explaining why women are much more likely to experience side effects from drugs—some of them fatal. Over the last couple of decades, greater effort has been made to include people with ovaries and uteruses in clinical research. But we still have a long way to go.

Women are also often subjected to medical advice designed to protect a potential fetus, whether they are pregnant or not. Official guidelines on how much mercury-containing fish it is safe to eat can be different for “women of childbearing age,” according to the US Environmental Protection Agency, for example.  And in 2021, the World Health Organization used the same language to describe people who should be a focus of policies to reduce alcohol consumption

The takeaway message is that it’s women who should be thinking about fetal health, says Cattapan. Not the industries producing these chemicals or the agencies that regulate them. Not even the men who contribute to a pregnancy. Just women who stand a chance of getting pregnant, whether they intend to or not. “It puts the onus of the health of future generations squarely on the shoulders of women,” she says.

Another problem is the language itself. The term “women of reproductive age” typically includes women between 15 and 44. Women at one end of that spectrum will have very different bodies and a very different set of health risks from those at the other. And the term doesn’t account for people who might be able to get pregnant but don’t necessarily identify as female.

In other cases it is overly broad. In the context of the Zika virus, for example, it was not all women between the ages of 15 and 44 who should have considered taking precautions. The travel advice didn’t apply to people who’d had hysterectomies or did not have sex with men, for example, says Cattapan. “Precision here matters,” she says. 

More nuanced health advice would be helpful in cases like these. Guidelines often read as though they’re written for people assumed to be stupid, she adds. “I don’t think that needs to be the case.”

Another thing

On Thursday, president-elect Donald Trump said that he will nominate Robert F. Kennedy Jr. to lead the US Department of Health and Human Services. The news was not entirely a surprise, given that Trump had told an audience at a campaign rally that he would let Kennedy “go wild” on health, “the foods,” and “the medicines.”

The role would give Kennedy some control over multiple agencies, including the Food and Drug Administration, which regulates medicines in the US, and the Centers for Disease Control and Prevention, which coordinates public health advice and programs.

That’s extremely concerning to scientists, doctors, and health researchers, given Kennedy’s positions on evidence-based medicine, including his antivaccine stance. A few weeks ago, in a post on X, he referred to the FDA’s “aggressive suppression of psychedelics, peptides, stem cells, raw milk, hyperbaric therapies, chelating compounds, ivermectin, hydroxychloroquine, vitamins, clean foods, sunshine, exercise, nutraceuticals and anything else that advances human health and can’t be patented by Pharma.”  

“If you work for the FDA and are part of this corrupt system, I have two messages for you,” continued the post. “1. Preserve your records, and 2. Pack your bags.”

There’s a lot to unpack here. But briefly, we don’t yet have good evidence that mind-altering psychedelic drugs are the mental-health cure-alls some claim they are. There’s not enough evidence to support the many unapproved stem-cell treatments sold by clinics throughout the US and beyond, either. These “treatments” can be dangerous.

Health agencies are currently warning against the consumption of raw unpasteurized milk, because it might carry the bird flu virus that has been circulating in US dairy farms. And it’s far too simplistic to lump all vitamins together—some might be of benefit to some people, but not everyone needs supplements, and high doses can be harmful.

Kennedy’s 2021 book The Real Anthony Fauci has already helped spread misinformation about AIDS. Here at MIT Technology Review, we’ll continue our work reporting on whatever comes next. Watch this space.


Now read the rest of The Checkup

Read more from MIT Technology Review’s archive

The tech industry has a gender problem, as the Gamergate and various #MeToo scandals made clear. A new generation of activists is hoping to remedy it

Male and female immune systems work differently. Which is another reason why it’s vital to study both women and female animals as well as males

Both of the above articles were published in the Gender issue of MIT Technology Review magazine. You can read more from that issue online here.

Women are more likely to receive abuse online. My colleague Charlotte Jee spoke to the technologists working on an alternative way to interact online: a feminist internet.

From around the web 

The scientific community and biopharma investors are reacting to the news of Robert F. Kennedy Jr.’s nomination to lead the Department of Health and Human Services. “It’s hard to see HHS functioning,” said one biotech analyst. (STAT)

Virologist Beata Halassy successfully treated her own breast cancer with viruses she grew in the lab. She has no regrets. (Nature)

Could diet influence the growth of endometriosis lesions? Potentially, according to research in mice fed high-fat, low-fiber “Western” diets. (BMC Medicine)

Last week, 43 female rhesus macaque monkeys escaped from a lab in South Carolina. The animals may have a legal claim to freedom. (Vox)

What Africa needs to do to become a major AI player

Kessel Okinga-Koumu paced around a crowded hallway. It was her first time presenting at the Deep Learning Indaba, she told the crowd gathered to hear her, filled with researchers from Africa’s machine-learning community. The annual weeklong conference (‘Indaba’ is a Zulu word for gathering), was held most recently in September at Amadou Mahtar Mbow University in Dakar, Senegal. It attracted over 700 attendees to hear about—and debate—the potential of Africa-centric AI and how it’s being deployed in agriculture, education, health care, and other critical sectors of the continent’s economy.     

A 28-year-old computer science student at the University of the Western Cape in Cape Town, South Africa, Okinga-Koumu spoke about how she’s tackling a common problem: the lack of lab equipment at her university. Lecturers have long been forced to use chalkboards or printed 2D representations of equipment to simulate practical lessons that need microscopes, centrifuges, or other expensive tools. “In some cases, they even ask students to draw the equipment during practical lessons,” she lamented. 

Okinga-Koumu pulled a phone from the pocket of her blue jeans and opened a prototype web app she’s built. Using VR and AI features, the app allows students to simulate using the necessary lab equipment—exploring 3D models of the tools in a real-world setting, like a classroom or lab. “Students could have detailed VR of lab equipment, making their hands-on experience more effective,” she said. 

Established in 2017, the Deep Learning Indaba now has chapters in 47 of the 55 African nations and aims to boost AI development across the continent by providing training and resources to African AI researchers like Okinga-Koumu. Africa is still early in the process of adopting AI technologies, but organizers say the continent is uniquely hospitable to it for several reasons, including a relatively young and increasingly well-educated population, a rapidly growing ecosystem of AI startups, and lots of potential consumers. 

“The building and ownership of AI solutions tailored to local contexts is crucial for equitable development,” says Shakir Mohamed, a senior research scientist at Google DeepMind and cofounder of the organization sponsoring the conference. Africa, more than other continents in the world, can address specific challenges with AI and will benefit immensely from its young talent, he says: “There is amazing expertise everywhere across the continent.” 

However, researchers’ ambitious efforts to develop AI tools that answer the needs of Africans face numerous hurdles. The biggest are inadequate funding and poor infrastructure. Not only is it very expensive to build AI systems, but research to provide AI training data in original African languages has been hamstrung by poor financing of linguistics departments at many African universities and the fact that citizens increasingly don’t speak or write local languages themselves. Limited internet access and a scarcity of domestic data centers also mean that developers might not be able to deploy cutting-edge AI capabilities.

Attendees of Deep Learning Indaba 2024 in session hall on their computers

DEEP LEARNING INDABA 2024

Complicating this further is a lack of overarching policies or strategies for harnessing AI’s immense benefits—and regulating its downsides. While there are various draft policy documents, researchers are in conflict over a continent-wide strategy. And they disagree about which policies would most benefit Africa, not the wealthy Western governments and corporations that have often funded technological innovation.

Taken together, researchers worry, these issues will hold Africa’s AI sector back and hamper its efforts to pave its own pathway in the global AI race.          

On the cusp of change

Africa’s researchers are already making the most of generative AI’s impressive capabilities. In South Africa, for instance, to help address the HIV epidemic, scientists have designed an app called Your Choice, powered by an LLM-based chatbot that interacts with people to obtain their sexual history without stigma or discrimination. In Kenya, farmers are using AI apps  to diagnose diseases in crops and increase productivity. And in Nigeria, Awarri, a newly minted AI startup, is trying to build the country’s first large language model, with the endorsement of the government, so that Nigerian languages can be integrated into AI tools. 

The Deep Learning Indaba is another sign of how Africa’s AI research scene is starting to flourish. At the Dakar meeting, researchers presented 150 posters and 62 papers. Of those, 30 will be published in top-tier journals, according to Mohamed. 

Meanwhile, an analysis of 1,646 publications in AI between 2013 and 2022 found “a significant increase in publications” from Africa. And Masakhane, a cousin organization to Deep Learning Indaba that pushes for natural-language-processing research in African languages, has released over 400 open-source models and 20 African-language data sets since it was founded in 2018. 

“These metrics speak a lot to the capacity building that’s happening,” says Kathleen Siminyu, a computer scientist from Kenya, who researches NLP tools for her native Kiswahili. “We’re starting to see a critical mass of people having basic foundational skills. They then go on to specialize.”      

She adds: “It’s like a wave that cannot be stopped.”   

Khadija Ba, a Senegalese entrepreneur and investor at the pan-African VC fund P1 Ventures who was at this year’s conference, says that she sees African AI startups as particularly attractive because their local approaches have potential to be scaled for the global market. African startups often build solutions in the absence of robust infrastructure, yet “these innovations work efficiently, making them adaptable to other regions facing similar challenges,” she says. 

In recent years, funding in Africa’s tech ecosystem has picked up: VC investment totaled $4.5 billion last year, more than double what it was just five years ago, according to a report by the African Private Capital Association. And this October, Google announced a $5.8 million commitment to support AI training initiatives in Kenya, Nigeria, and South Africa. But researchers say local funding remains sluggish. Take the Google-backed fund rolled out, also in October, in Nigeria, Africa’s most populous country. It will pay out $6,000 each to 10 AI startups—not even enough to purchase the equipment needed to power their systems.

Lilian Wanzare, a lecturer and NLP researcher at Maseno University in Kisumu, Kenya, bridles at African governments’ lackadaisical support for local AI initiatives and complains as well that the government charges exorbitant fees for access to publicly generated data, hindering data sharing and collaboration. “[We] researchers are just blocked,” she says. “The government is saying they’re willing to support us, but the structures have not been put in place for us.”

Language barriers 

Researchers who want to make Africa-centric AI don’t face just insufficient local investment and inaccessible data. There are major linguistic challenges, too.  

During one discussion at the Indaba, Ife Adebara, a Nigerian computational linguist, posed a question: “How many people can write a bachelor’s thesis in their native African language?” 

Zero hands went up. 

Then the audience disintegrated into laughter.   

Africans want AI to speak their local languages, but many Africans cannot speak and write in these languages themselves, Adebara said.      

Although Africa accounts for one-third of all languages in the world, many oral languages are slowly disappearing, their population of native speakers declining. And LLMs developed by Western-based tech companies fail to serve African languages; they don’t understand locally relevant context and culture. 

For Adebara and others researching NLP tools, the lack of people who have the ability to read and write in African languages poses a major hurdle to development of bespoke AI-enabled technologies. “Without literacy in our local languages, the future of AI in Africa is not as bright as we think,” she says.      

On top of all that, there’s little machine-readable data for African languages. One reason is that linguistic departments in public universities are poorly funded, Adebara says, limiting linguists’ participation in work that could create such data and benefit AI development. 

This year, she and her colleagues established EqualyzAI, a for-profit company seeking to preserve African languages through digital technology. They have built voice tools and AI models, covering about 517 African languages.       

Lelapa AI, a software company that’s building data sets and NLP tools for African languages, is also trying to address these language-specific challenges. Its cofounders met in 2017 at the first Deep Learning Indaba and launched the company in 2022. In 2023, it released its first AI tool, Vulavula, a speech-to-text program that recognizes several languages spoken in South Africa. 

This year, Lelapa AI released InkubaLM, a first-of-its-kind small language model that currently supports a range of African languages: IsiXhosa, Yoruba, Swahili, IsiZulu, and Hausa. InkubaLM can answer questions and perform tasks like English translation and sentiment analysis. In tests, it performed as well as some larger models. But it’s still in early stages. The hope is that InkubaLM will someday power Vulavula, says Jade Abbott, cofounder and chief operating officer of Lelapa AI. 

“It’s the first iteration of us really expressing our long-term vision of what we want, and where we see African AI in the future,” Abbott says. “What we’re really building is a small language model that punches above its weight.”

InkubaLM is trained on two open-source data sets with 1.9 billion tokens, built and curated by Masakhane and other African developers who worked with real people in local communities. They paid native speakers of languages to attend writing workshops to create data for their model.

Fundamentally, this approach will always be better, says Wanzare, because it’s informed by people who represent the language and culture.

A clash over strategy

Another issue that came up again and again at the Indaba was that Africa’s AI scene lacks the sort of regulation and support from governments that you find elsewhere in the world—in Europe, the US, China, and, increasingly, the Middle East. 

Of the 55 African nations, only seven—Senegal, Egypt, Mauritius, Rwanda, Algeria, Nigeria, and Benin—have developed their own formal AI strategies. And many of those are still in the early stages.  

A major point of tension at the Indaba, though, was the regulatory framework that will govern the approach to AI across the entire continent. In March, the African Union Development Agency published a white paper, developed over a three-year period, that lays out this strategy. The 200-page document includes recommendations for industry codes and practices, standards to assess and benchmark AI systems, and a blueprint of AI regulations for African nations to adopt. The hope is that it will be endorsed by the heads of African governments in February 2025 and eventually passed by the African Union.  

But in July, the African Union Commission in Addis Ababa, Ethiopia, another African governing body that wields more power than the development agency, released a rival continental AI strategy—a 66-page document that diverges from the initial white paper. 

It’s unclear what’s behind the second strategy, but Seydina Ndiaye, a program director at the Cheikh Hamidou Kane Digital University in Dakar who helped draft the development agency’s white paper, claims it was drafted by a tech lobbyist from Switzerland. The commission’s strategy calls for African Union member states to declare AI a national priority, promote AI startups, and develop regulatory frameworks to address safety and security challenges. But Ndiaye expressed concerns that the document does not reflect the perspectives, aspirations, knowledge, and work of grassroots African AI communities. “It’s a copy-paste of what’s going on outside the continent,” he says.               

Vukosi Marivate, a computer scientist at the University of Pretoria in South Africa who helped found the Deep Learning Indaba and is known as an advocate for the African machine-learning movement, expressed fury over this turn of events at the conference. “These are things we shouldn’t accept,” he declared. The room full of data wonks, linguists, and international funders brimmed with frustration. But Marivate encouraged the group to forge ahead with building AI that benefits Africans: “We don’t have to wait for the rules to act right,” he said.  

Barbara Glover, a program manager for the African Union Development Agency, acknowledges that AI researchers are angry and frustrated. There’s been a push to harmonize the two continental AI strategies, but she says the process has been fractious: “That engagement didn’t go as envisioned.” Her agency plans to keep its own version of the continental AI strategy, Glover says, adding that it was developed by African experts rather than outsiders. “We are capable, as Africans, of driving our own AI agenda,” she says.       

crowd of attendees mingle around display booths at Deep Learning Indaba 2024. Booth signs for Mila, Meta and OpenAI can be seen in the frame.

DEEP LEARNING INDABA 2024

This all speaks to a broader tension over foreign influence in the African AI scene, one that goes beyond any single strategic document. Mirroring the skepticism toward the African Union Commission strategy, critics say the Deep Learning Indaba is tainted by its reliance on funding from big foreign tech companies; roughly 50% of its $500,000 annual budget comes from international donors and the rest from corporations like Google DeepMind, Apple, Open AI, and Meta. They argue that this cash could pollute the Indaba’s activities and influence the topics and speakers chosen for discussion. 

But Mohamed, the Indaba cofounder who is a researcher at Google DeepMind, says that “almost all that goes back to our beneficiaries across the continent,” and the organization helps connect them to training opportunities in tech companies. He says it benefits from some of its cofounders’ ties with these companies but that they do not set the agenda.

Ndiaye says that the funding is necessary to keep the conference going. “But we need to have more African governments involved,” he says.     

To Timnit Gebru, founder and executive director at the nonprofit Distributed AI Research Institute (DAIR), which supports equitable AI research in Africa, the angst about foreign funding for AI development comes down to skepticism of exploitative, profit-driven international tech companies. “Africans [need] to do something different and not replicate the same issues we’re fighting against,” Gebru says. She warns about the pressure to adopt “AI for everything in Africa,” adding that there’s “a lot of push from international development organizations” to use AI as an “antidote” for all Africa’s challenges.       

Siminyu, who is also a researcher at DAIR, agrees with that view. She hopes that African governments will fund and work with people in Africa to build AI tools that reach underrepresented communities—tools that can be used in positive ways and in a context that works for Africans. “We should be afforded the dignity of having AI tools in a way that others do,” she says.     

Life-seeking, ice-melting robots could punch through Europa’s icy shell

At long last, NASA’s Europa Clipper mission is on its way. After overcoming financial and technological hurdles, the $5 billion mission launched on October 14 from Florida’s Kennedy Space Center. It is now en route to its target: Jupiter’s ice-covered moon Europa, whose frozen shell almost certainly conceals a warm saltwater ocean. When the spacecraft gets there, it will conduct dozens of close flybys in order to determine what that ocean is like and, crucially, where it might be hospitable to life.

Europa Clipper is still years away from its destination—it is not slated to reach the Jupiter system until 2030. But that hasn’t stopped engineers and scientists from working on what would come next if the results are promising: a mission capable of finding evidence of life itself.

This would likely have three parts: a lander, an autonomous ice-thawing robot, and some sort of self-navigating submersible. Indeed, several groups from multiple countries already have working prototypes of ice-diving robots and smart submersibles that they are set to test in Earth’s own frigid landscapes, from Alaska to Antarctica, in the next few years

But Earth’s oceans are pale simulacra of Europa’s extreme environment. To plumb the ocean of this Jovian moon, engineers must work out a way to get missions to survive a  never-ending rain of radiation that fries electronic circuits. They must also plow through an ice shell that’s at least twice as thick as Mount Everest is tall.

“There are a lot of hard problems that push up right against the limits of what’s possible,” says Richard Camilli, an expert on autonomous robotic systems at the Woods Hole Oceanographic Institution’s Deep Submergence Laboratory. But you’ve got to start somewhere, and Earth’s seas will be a vital testing ground. 

“We’re doing something nobody has done before,” says Sebastian Meckel, a researcher at the Center for Marine Environmental Sciences at the University of Bremen, Germany, who is helping to develop one such futuristic Europan submersible. If the field tests prove successful, the descendants of these aquatic explorers could very well be those that uncover the first evidence of extraterrestrial life.

Hellish descent

The hunt for signs of extraterrestrial biology has predominantly taken place on Mars, our dusty, diminutive planetary neighbor. Looking for life in an icy ocean world is a whole new kettle of (alien) fish, but exobiologists think it’s certainly worth the effort. On Mars, scientists hope to find microscopic evidence of past life on, or just under, its dry and frozen surface. But on Europa, which has a wealth of liquid water (kept warm by Jupiter, whose intense gravity generates plenty of internal friction and heat there), it is possible that microbial critters, and perhaps even more advanced small aquatic animals, may be present in the here and now.

The bad news is that Europa is one of the most hostile environments in the solar system—at least, for anything above its concealed ocean. 

When NASA’s Clipper mission arrives in 2030, it will be confronted by an endless storm of high-energy particles being whipped about by Jupiter’s immense and intense magnetic field, largely raining down onto Europa itself. “It’s enough to kill a regular person within a few seconds,” says Camilli. No human will be present on Europa, but that radiation is so extreme that it can frazzle most electronic circuits. This poses a major hazard for Europa Clipper, which is why it’s doing only quick flybys of the moon as its orbit around Jupiter periodically dips close.

Clipper has an impressive collection of remote sensing tools that will allow it to survey the ocean’s physical and chemical properties, even though it will never touch the moon itself. But almost all scientists expect that uncovering evidence of biological activity will require something to pierce through the ice shell and swim about in the ocean.

A cross-section view of an ice-melting probe called PRIME on the surface of the moon, with small robots being deployed in the subsurface ocean, against the backdrop of Jupiter.
An illustration of two Europa exploration concepts from NASA. An ice-melting probe called PRIME sits on the surface of the moon, with small wedge-shaped SWIM robots deployed below.
NASA/JPL-CALTECH

The good news is that any Europan life-hunting mission has a great technological legacy to build upon. Over the years, scientists have developed and deployed robotic subs that have uncovered a cornucopia of strange life and bizarre geology dwelling in the deep. These include remotely operated vehicles (ROVs), which are often tethered to a surface vessel and are piloted by a person atop the waves, and autonomous underwater vehicles (AUVs), which freely traverse the seas by themselves before reporting back to the surface.

Hopeful Europa explorers usually cite an AUV as their best option—something that a lander can drop off and let loose in those alien waters that will then return and share its data so it can be beamed back to Earth. “The whole idea is very exciting and cool,” says Bill Chadwick, a research professor at Oregon State University’s Hatfield Marine Science Center in Newport, Oregon. But on a technical level, he adds, “it seems incredibly daunting.”

Presuming that a life-finding robotic mission is sufficiently radiation-proof and can land and sit safely on Europa’s surface, it would then encounter the colossal obstacle that is Europa’s ice shell, estimated to be 10 to 15 miles thick. Something is going to have to drill or melt its way through all that before reaching the ocean, a process that will likely take several years. “And there’s no guarantee that the ice is going to be static as you’re going through,” says Camilli. Thanks to gravitational tugs from Jupiter, and the internal heat they generate, Europa is a geologically tumultuous world, with ice constantly fragmenting, convulsing and even erupting on its surface. “How do you deal with that?”

Europa’s lack of an atmosphere is also an issue. Say your robot does reach the ocean below all that ice. That’s great, but if the thawed tunnel isn’t sealed shut behind the robot, then the higher pressure of the oceanic depths will come up against a vacuum high above. “If you drill through and you don’t have some kind of pressure control, you can get the equivalent of a blowout, like an oil well,” says Camilli—and your robot could get rudely blasted into space.

Even if you manage to pass through that gauntlet, you must then make sure the diver maintains a link with the surface lander, and with Earth. “What would be worse than finally finding life somewhere else and not being able to tell anyone about it?” says Morgan Cable, a research scientist at NASA’s Jet Propulsion Laboratory (JPL).

Pioneering probes

What these divers will do when they breach Europa’s ocean almost doesn’t matter at this stage. The scientific analysis is currently secondary to the primary problem: Can robots actually get through that ice shell and survive the journey? 

A simple way to start is with a cryobot—a melt probe that can gradually thaw its way through the shell, pulled down by gravity. That’s the idea behind NASA’s Probe using Radioisotopes for Icy Moons Exploration, or PRIME. As the name suggests, this cryobot would use the heat from the radioactive decay of an element like plutonium-238 to melt ice. If you know the thickness of the ice shell, you know exactly how many tablespoons of radioactive matter to bring aboard. 

Once it gets through the ice, the cryobot could unfurl a suite of scientific investigation tools, or perhaps deploy an independent submersible that could work in tandem with the cryobot—all while making sure none of that radioactive matter contaminates the ocean. NASA’s Sensing with Independent Micro-Swimmers project, for example, has sketched out plans to deploy a school of wedge-shaped robots—a fleet of sleuths that would work together to survey the depths before reporting back to base.

These concepts remain hypothetical. To get an idea of what’s technically possible, several teams are building and field-testing their own prototype ice divers. 

One of the furthest-along efforts is the Ocean Worlds Reconnaissance and Characterization of Astrobiological Analogs project, or ORCAA, led by JPL. After some preliminary fieldwork, the group is now ready for prime time; next year, a team will set up camp on Alaska’s expansive Juneau Icefield and deploy an eight-foot tall, two-inch wide cryobot. Its goal will be to get through 1,000 feet of ice, through a glasslike upper layer, down into ancient ices, and ultimately into a subglacial lake.

A shows two team members near a supraglacial lake (a body of water on top of the glacier), where biologists could take water samples and compare them to samples taken from the borehole.
ORCAA team members stand by a lake on top of a glacier during Alaska fieldwork.
NASA/JPL-CALTECH

This cryobot won’t be powered by radioactive matter. “I don’t see NASA and the Department of Energy being game for that yet,” says Samuel Howell, an ocean worlds scientist at JPL and the ORCAA principal investigator. Instead, it will be electrically heated (with power delivered via a tether to the surface), and that heat will pump warm water out in front of the cryobot, melting the ice and allowing it to migrate downward.

The cryobot will be permanently tethered to the surface, using that link to communicate its rudimentary scientific data and return samples of water back to a team of scientists at base camp atop the ice. Those scientists will act as if they are an astrobiology suite of instruments similar to what might eventually be fitted on a cryobot sent to Europa. 

The 2025 field experiment “has all the pieces of a cryobot mission,” says Howell. “We’re just duct-taping them together and trying to see what breaks.”

Space scientists and marine engineers are also teaming up at Germany’s Center for Marine Environmental Sciences (MARUM) to forge their own underwater explorer. Under the auspices of the Technologies for Rapid Ice Penetration and Subglacial Lake Exploration project, or TRIPLE, they are developing an ice-thawing cryobot, an astrobiological laboratory suite, and an AUV designed to be used in Earth’s seas and Europa’s ocean.

Their cryobot is somewhat like the one ORCAA is using; it’s an electrically heated thawing machine tethered to the surface. But onboard MARUM’s “ice shuttle” will be a remarkably small AUV, just 20 inches long and four inches wide. The team plans to deploy both on the Antarctic ice shelf, near the Neumayer III station, in the spring of 2026. 

Model of the miniature underwater vehicle being developed at MARUM with partners from industry. It will have a diameter of around ten and a length of about 50 centimeters.
Germany’s Center for Marine Environmental Sciences is developing a small AUV that it plans to deploy in Antarctica in 2026.
MARUM – CENTER FOR MARINE ENVIRONMENTAL SCIENCES, UNIVERSITY OF BREMEN.

From a surface station, the ice shuttle will thaw its way down through the ice shell, aiming to reach the bitingly cold water hundreds of feet below. Once it does so, a hatch will open and the tiny AUV will be dropped off to swim about (on a probably preprogrammed route), wirelessly communicating with the ice shuttle throughout. It will take a sample of the water, return to the ice shuttle, dock with it, and recharge its batteries. For the field test, the ice shuttle, which will have some rudimentary scientific tools, will return the water sample back to the surface for analysis; for the space mission itself, the idea is that an array of instruments onboard the shuttle will examine that water.

As with ORCAA, the scientific aspect of this is not paramount. “What we’re focusing on now is form and function,” says project member Ralf Bachmayer, a marine robotics researcher at MARUM. Can their prototype Europan explorer get down to the hidden waters, deploy a scout, and return to base intact?

Bachmayer can’t wait to find out. “For engineers, it’s a dream come true to work on this project,” he says.

Swarms and serpents

A submersible-like AUV isn’t the only way scientists are thinking of investigating icy oceanic moons. JPL’s Exobiology Extant Life Surveyor, or EELS, involves a working, wriggling, serpentine robot inspired by the desire to crawl through the vents of Saturn’s own water-laden moon, Enceladus. The robotic snake has already been field-tested; it recently navigated through the icy crevasses and moulins of the Athabasca Glacier in Alberta, Canada.

Although an AUV-like cryobot mission is likely to be the first explorer of an icy oceanic moon, “a crazy idea like a robotic snake could work,” says Cable, the science lead for EELS. She hopes the project is “opening the eyes of scientists and engineers alike to new possibilities when it comes to accessing the hard-to-reach, and often most scientifically compelling, places of planetary environments.”

It might be that we’ll need such creative, and perhaps unexpected, designs to find our way to Europa’s ocean. Space agencies exploring the solar system have achieved remarkable things, but “NASA has never flown an aqueous instrument before,” says Howell.

But one day, thanks to this work, it might—and, just maybe, one of them will find life blooming in Europa’s watery shadows.

Robin George Andrews is an award-winning science journalist and doctor of volcanoes based in London. He regularly writes about the Earth, space, and planetary sciences, and is the author of two critically acclaimed books: Super Volcanoes (2021) and How To Kill An Asteroid (October 2024).