The Download: Clear’s identity ambitions, and the climate blame game

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

Inside Clear’s ambitions to manage your identity beyond the airport 

Clear Secure is the most visible biometric identity company in the United States. Best known for its line-jumping service in airports, it’s also popping up at  sports arenas and stadiums all over the country. You can also use its identity verification platform to rent tools at Home Depot, put your profile in front of recruiters on LinkedIn, and, as of this month, verify your identity as a rider on Uber.

And soon enough, if Clear has its way, it may also be in your favorite retailer, bank, and even doctor’s office—or anywhere else that you currently have to pull out a wallet (or wait in line). 

While the company has been building toward this sweeping vision for years, it now seems its time has finally come. But as biometrics go mainstream, what—and who—bears the cost? Read the full story.  

—Eileen Guo

LinkedIn Live: Facial verification tech promises a frictionless future. But at what cost?

Do you use your face to unlock your phone, or speed through airport security? As biometrics companies move into more and more spaces, where else would you use this technology? The trade off seems simple: you scan your face, you get a frictionless future. But is it really? Join MIT Technology Review’s features and investigations team for a LinkedIn Live this Thursday, November 21, about the rise of facial verification tech and what it means to give up your face. Register for free.

Who’s to blame for climate change? It’s surprisingly complicated.

Once again, global greenhouse-gas emissions are projected to hit a new high in 2024. 

In this time of shifting political landscapes and ongoing international negotiations, many are quick to blame one country or another for an outsize role in causing climate change.

But assigning responsibility is complicated. These three visualizations help explain why.

—Casey Crownhart

Cyber Week Sale: subscriptions are half price!

Take advantage of epic savings on award-winning reporting, razor-sharp analysis, and expert insights on your favorite technology topics. Subscribe today to save 50% on an annual subscription, plus receive a free digital copy of our “Generative AI and the future of work” report. Don’t miss out.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 AI can now translate your voice in real-time during meetings
It’s part of Microsoft’s drive to push more AI into its products, but how well it works in the wild remains to be seen. (WP $)
Apple is having less success on that front, at least if its AI notification summaries are anything to go by. (The Atlantic $)

2 Anyone can buy data tracking US soldiers in Germany 
And the Pentagon is powerless to stop it.(Wired $)
It’s shockingly easy to buy sensitive data about US military personnel. (MIT Technology Review)

3 Bluesky now has over 20 million users 📈
Its user base has tripled in the last three months. (Engadget)
Truth Social, on the other hand, is not doing quite so well. (WP $)
+ The rise of Bluesky, and the splintering of social. (MIT Technology Review)

4 How Google created a culture of concealment 
It’s been preparing for antitrust action for over a decade, enforcing a policy where employees delete messages by default. (NYT $)
+ The company reacted angrily to reports it may be forced to sell Chrome. (BBC)

5 Project 2025 is already infiltrating the Trump administration 
Despite repeated denials, it’s clearly a blueprint for his next term. (Vox)
A hacker reportedly gained access to damaging testimonies about Matt Gaetz, his pick to be attorney general. (NYT $)

6 Quantum computers hit a major milestone for error-free calculation
This is a crucial part of making them useful for real-world tasks. (New Scientist $)

7 Technology is changing political speech
Slogans are becoming less effective. Now it’s more about saying different things to different audiences. (New Yorker $)

8 Lab-grown foie gras, anyone?
This could be the cultivated meat industry’s future: as a luxury product for the few. (Wired $)

9 Niantic is using Pokémon Go player data to build an AI navigation system
If it works, it could unlock some amazing capabilities in AR, robotics and beyond. (404 Media)

10 Minecraft is expanding into the real world
It has struck a $110 million deal with one of the world’s biggest theme park operators. (The Guardian)

Quote of the day

“Nobody believes that these cables were severed by accident.”

—Germany’s defense minister Boris Pistorius, tells reporters that the severing of two fiber-optic cables in the Baltic Sea was a deliberate act of sabotage, the New York Times reports. 

 The big story

Are we alone in the universe?

a road leads to a sky filled with orbiting cubes and glowing lights

ARIEL DAVIS

November 2023

The quest to determine if life is out there has gained greater scientific footing over the past 50 years. Back then, astronomers had yet to spot a single planet outside our solar system. Now we know the galaxy is teeming with a diversity of worlds.

We’re getting closer than ever before to learning how common living worlds like ours actually are. New tools, including artificial intelligence, could help scientists look past their preconceived notions of what constitutes life. Read the full story.

—Adam Mann

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ How to not only survive but thrive during the winter.
+ Fancy working from somewhere new? Here are some of the best cities for a workcation. 
+ Want to see David Bowie imitating Mick Jagger? Of course you do
+ It’s an old(ish) joke but still funny.

AI can now create a replica of your personality

Imagine sitting down with an AI model for a spoken two-hour interview. A friendly voice guides you through a conversation that ranges from your childhood, your formative memories, and your career to your thoughts on immigration policy. Not long after, a virtual replica of you is able to embody your values and preferences with stunning accuracy.

That’s now possible, according to a new paper from a team including researchers from Stanford and Google DeepMind, which has been published on arXiv and has not yet been peer-reviewed. 

Led by Joon Sung Park, a Stanford PhD student in computer science, the team recruited 1,000 people who varied by age, gender, race, region, education, and political ideology. They were paid up to $100 for their participation. From interviews with them, the team created agent replicas of those individuals. As a test of how well the agents mimicked their human counterparts, participants did a series of personality tests, social surveys, and logic games, twice each, two weeks apart; then the agents completed the same exercises. The results were 85% similar. 

“If you can have a bunch of small ‘yous’ running around and actually making the decisions that you would have made—that, I think, is ultimately the future,” Joon says. 

In the paper the replicas are called simulation agents, and the impetus for creating them is to make it easier for researchers in social sciences and other fields to conduct studies that would be expensive, impractical, or unethical to do with real human subjects. If you can create AI models that behave like real people, the thinking goes, you can use them to test everything from how well interventions on social media combat misinformation to what behaviors cause traffic jams. 

Such simulation agents are slightly different from the agents that are dominating the work of leading AI companies today. Called tool-based agents, those are models built to do things for you, not converse with you. For example, they might enter data, retrieve information you have stored somewhere, or—someday—book travel for you and schedule appointments. Salesforce announced its own tool-based agents in September, followed by Anthropic in October, and OpenAI is planning to release some in January, according to Bloomberg

The two types of agents are different but share common ground. Research on simulation agents, like the ones in this paper, is likely to lead to stronger AI agents overall, says John Horton, an associate professor of information technologies at the MIT Sloan School of Management, who founded a company to conduct research using AI-simulated participants. 

“This paper is showing how you can do a kind of hybrid: use real humans to generate personas which can then be used programmatically/in-simulation in ways you could not with real humans,” he told MIT Technology Review in an email. 

The research comes with caveats, not the least of which is the danger that it points to. Just as image generation technology has made it easy to create harmful deepfakes of people without their consent, any agent generation technology raises questions about the ease with which people can build tools to personify others online, saying or authorizing things they didn’t intend to say. 

The evaluation methods the team used to test how well the AI agents replicated their corresponding humans were also fairly basic. These included the General Social Survey—which collects information on one’s demographics, happiness, behaviors, and more—and assessments of the Big Five personality traits: openness to experience, conscientiousness, extroversion, agreeableness, and neuroticism. Such tests are commonly used in social science research but don’t pretend to capture all the unique details that make us ourselves. The AI agents were also worse at replicating the humans in behavioral tests like the “dictator game,” which is meant to illuminate how participants consider values such as fairness. 

To build an AI agent that replicates people well, the researchers needed ways to distill our uniqueness into language AI models can understand. They chose qualitative interviews to do just that, Joon says. He says he was convinced that interviews are the most efficient way to learn about someone after he appeared on countless podcasts following a 2023 paper that he wrote on generative agents, which sparked a huge amount of interest in the field. “I would go on maybe a two-hour podcast podcast interview, and after the interview, I felt like, wow, people know a lot about me now,” he says. “Two hours can be very powerful.”

These interviews can also reveal idiosyncrasies that are less likely to show up on a survey. “Imagine somebody just had cancer but was finally cured last year. That’s very unique information about you that says a lot about how you might behave and think about things,” he says. It would be difficult to craft survey questions that elicit these sorts of memories and responses. 

Interviews aren’t the only option, though. Companies that offer to make “digital twins” of users, like Tavus, can have their AI models ingest customer emails or other data. It tends to take a pretty large data set to replicate someone’s personality that way, Tavus CEO Hassaan Raza told me, but this new paper suggests a more efficient route. 

“What was really cool here is that they show you might not need that much information,” Raza says, adding that his company will experiment with the approach. “How about you just talk to an AI interviewer for 30 minutes today, 30 minutes tomorrow? And then we use that to construct this digital twin of you.”

How the largest gathering of US police chiefs is talking about AI

This story is from The Algorithm, our weekly newsletter on AI. To get it in your inbox first, sign up here.

It can be tricky for reporters to get past certain doors, and the door to the International Association of Chiefs of Police conference is one that’s almost perpetually shut to the media. Thus, I was pleasantly surprised when I was able to attend for a day in Boston last month. 

It bills itself as the largest gathering of police chiefs in the United States, where leaders from many of the country’s 18,000 police departments and even some from abroad convene for product demos, discussions, parties, and awards. 

I went along to see how artificial intelligence was being discussed, and the message to police chiefs seemed crystal clear: If your department is slow to adopt AI, fix that now. The future of policing will rely on it in all its forms.

In the event’s expo hall, the vendors (of which there were more than 600) offered a glimpse into the ballooning industry of police-tech suppliers. Some had little to do with AI—booths showcased body armor, rifles, and prototypes of police-branded Cybertrucks, and others displayed new types of gloves promising to protect officers from needles during searches. But one needed only to look to where the largest crowds gathered to understand that AI was the major draw. 

The hype focused on three uses of AI in policing. The flashiest was virtual reality, exemplified by the booth from V-Armed, which sells VR systems for officer training. On the expo floor, V-Armed built an arena complete with VR goggles, cameras, and sensors, not unlike the one the company recently installed at the headquarters of the Los Angeles Police Department. Attendees could don goggles and go through training exercises on responding to active shooter situations. Many competitors of V-Armed were also at the expo, selling systems they said were cheaper, more effective, or simpler to maintain. 

The pitch on VR training is that in the long run, it can be cheaper and more engaging to use than training with actors or in a classroom. “If you’re enjoying what you’re doing, you’re more focused and you remember more than when looking at a PDF and nodding your head,” V-Armed CEO Ezra Kraus told me. 

The effectiveness of VR training systems has yet to be fully studied, and they can’t completely replicate the nuanced interactions police have in the real world. AI is not yet great at the soft skills required for interactions with the public. At a different company’s booth, I tried out a VR system focused on deescalation training, in which officers were tasked with calming down an AI character in distress. It suffered from lag and was generally quite awkward—the character’s answers felt overly scripted and programmatic. 

The second focus was on the changing way police departments are collecting and interpreting data. Rather than buying a gunshot detection tool from one company and a license plate reader or drone from another, police departments are increasingly using expanding suites of sensors, cameras, and so on from a handful of leading companies that promise to integrate the data collected and make it useful. 

Police chiefs attended classes on how to build these systems, like one taught by Microsoft and the NYPD about the Domain Awareness System, a web of license plate readers, cameras, and other data sources used to track and monitor crime in New York City. Crowds gathered at massive, high-tech booths from Axon and Flock, both sponsors of the conference. Flock sells a suite of cameras, license plate readers, and drones, offering AI to analyze the data coming in and trigger alerts. These sorts of tools have come in for heavy criticism from civil liberties groups, which see them as an assault on privacy that does little to help the public. 

Finally, as in other industries, AI is also coming for the drudgery of administrative tasks and reporting. Many companies at the expo, including Axon, offer generative AI products to help police officers write their reports. Axon’s offering, called Draft One, ingests footage from body cameras, transcribes it, and creates a first draft of a report for officers. 

“We’ve got this thing on an officer’s body, and it’s recording all sorts of great stuff about the incident,” Bryan Wheeler, a senior vice president at Axon, told me at the expo. “Can we use it to give the officer a head start?”

On the surface, it’s a writing task well suited for AI, which can quickly summarize information and write in a formulaic way. It could also save lots of time officers currently spend on writing reports. But given that AI is prone to “hallucination,” there’s an unavoidable truth: Even if officers are the final authors of their reports, departments adopting these sorts of tools risk injecting errors into some of the most critical documents in the justice system. 

“Police reports are sometimes the only memorialized account of an incident,” wrote Andrew Ferguson, a professor of law at American University, in July in the first law review article about the serious challenges posed by police reports written with AI. “Because criminal cases can take months or years to get to trial, the accuracy of these reports are critically important.” Whether certain details were included or left out can affect the outcomes of everything from bail amounts to verdicts. 

By showing an officer a generated version of a police report, the tools also expose officers to details from their body camera recordings before they complete their report, a document intended to capture the officer’s memory of the incident. That poses a problem. 

“The police certainly would never show video to a bystander eyewitness before they ask the eyewitness about what took place, as that would just be investigatory malpractice,” says Jay Stanley, a senior policy analyst with the ACLU Speech, Privacy, and Technology Project, who will soon publish work on the subject. 

A spokesperson for Axon says this concern “isn’t reflective of how the tool is intended to work,” and that Draft One has robust features to make sure officers read the reports closely, add their own information, and edit the reports for accuracy before submitting them.

My biggest takeaway from the conference was simply that the way US police are adopting AI is inherently chaotic. There is no one agency governing how they use the technology, and the roughly 18,000 police departments in the United States—the precise figure is not even known—have remarkably high levels of autonomy to decide which AI tools they’ll buy and deploy. The police-tech companies that serve them will build the tools police departments find attractive, and it’s unclear if anyone will draw proper boundaries for ethics, privacy, and accuracy. 

That will only be made more apparent in an upcoming Trump administration. In a policing agenda released last year during his campaign, Trump encouraged more aggressive tactics like “stop and frisk,” deeper cooperation with immigration agencies, and increased liability protection for officers accused of wrongdoing. The Biden administration is now reportedly attempting to lock in some of its proposed policing reforms before January. 

Without federal regulation on how police departments can and cannot use AI, the lines will be drawn by departments and police-tech companies themselves.

“Ultimately, these are for-profit companies, and their customers are law enforcement,” says Stanley. “They do what their customers want, in the absence of some very large countervailing threat to their business model.”


Now read the rest of The Algorithm

Deeper Learning

The AI lab waging a guerrilla war over exploitative AI

When generative AI tools landed on the scene, artists were immediately concerned, seeing them as a new kind of theft. Computer security researcher Ben Zhao jumped into action in response, and his lab at the University of Chicago started building tools like Nightshade and Glaze to help artists keep their work from being scraped up by AI models. My colleague Melissa Heikkilä spent time with Zhao and his team to look at the ongoing effort to make these tools strong enough to stop AI’s relentless hunger for more images, art, and data to train on.  

Why this matters: The current paradigm in AI is to build bigger and bigger models, and these require vast data sets to train on. Tech companies argue that anything on the public internet is fair game, while artists demand compensation or the right to refuse. Settling this fight in the courts or through regulation could take years, so tools like Nightshade and Glaze are what artists have for now. If the tools disrupt AI companies’ efforts to make better models, that could push them to the negotiating table to bargain over licensing and fair compensation. But it’s a big “if.” Read more from Melissa Heikkilä.

Bits and Bytes

Tech elites are lobbying Elon Musk for jobs in Trump’s administration

Elon Musk is the tech leader who most has Trump’s ear. As such, he’s reportedly the conduit through which AI and tech insiders are pushing to have an influence in the incoming administration. (The New York Times)

OpenAI is getting closer to launching an AI agent to automate your tasks

AI agents—models that can do tasks for you on your behalf—are all the rage. OpenAI is reportedly closer to releasing one, news that comes a few weeks after Anthropic announced its own. (Bloomberg)

How this grassroots effort could make AI voices more diverse

A massive volunteer-led effort to collect training data in more languages, from people of more ages and genders, could help make the next generation of voice AI more inclusive and less exploitative. (MIT Technology Review

Google DeepMind has a new way to look inside an AI’s “mind”

Autoencoders let us peer into the black box of artificial intelligence. They could help us create AI that is better understood and more easily controlled. (MIT Technology Review)

Musk has expanded his legal assault on OpenAI to target Microsoft

Musk has expanded his federal lawsuit against OpenAI, which alleges that the company has abandoned its nonprofit roots and obligations. He’s now going after Microsoft too, accusing it of antitrust violations in its work with OpenAI. (The Washington Post)

The Download: police AI, and mixed reality’s future

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

How the largest gathering of US police chiefs is talking about AI

—James O’Donnell

The International Association of Chiefs of Police bills itself as the largest gathering of its type in the United States. Leaders from many of the country’s 18,000 police departments and even some from abroad convene for product demos, discussions, parties, and awards. 

I went along last month to see how artificial intelligence was being discussed, and the message to police chiefs seemed crystal clear: If your department is slow to adopt AI, fix that now. From the expo hall, talks, and interviews, it seems they’re already enthusiastically heeding the call. Read the full story.

This story is from The Algorithm, our weekly AI newsletter. Sign up to receive it in your inbox every Monday.

Roundtables: What’s Next for Mixed Reality: Glasses, Goggles, and More

After years of trying, augmented-reality specs are at last a thing. 

If you want to learn more about where AR experiences are heading, join our editor-in-chief Mat Honan and AI hardware reporter James O’Donnell for a Roundtables conversation streamed online at 2pm ET/11am PT today. It’s for subscribers only but good news: this week our subscriptions are half price. Don’t miss out! 

Read more about mixed reality:

+ We interviewed Palmer Luckey, founder of Oculus, about his plans to bring mixed-reality goggles to soldiers. Here’s what he had to say.

+  The coolest thing about smart glasses is not the AR. It’s the AI.

+ Snap has launched new augmented-reality Spectacles. Here’s what we made of them

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 The FBI is investigating threats texted to Latino and LGBTQ+ people 
They claim recipients will be deported or sent to a re-education camp. (WP $)
+  ICE can already sidestep sanctuary city laws through data-sharing centers. (Wired $)
Trump has confirmed he plans to use the military for mass deportations. (NYT $)

2 Chinese tech groups are building AI teams in Silicon Valley 
Despite Washington’s best efforts to stymie their work. (FT $)
How a US ban on investing in Chinese startups could escalate under Trump. (Wired $)

3 How Apple will cope with looming tariffs 
The fact CEO Tim Cook already has a relationship with Trump will surely help. (Bloomberg $)

4 Two undersea cables in the Baltic Sea have been disrupted 
It looks like Russia is trying to interfere with global undersea infrastructure. (CNN)
A Russian spy ship had to be escorted out of the Irish Sea last weekend too. (The Guardian)

5 An AI tool could help solve math problems humans are stuck on
It’s a good example of how blending human and machine intelligence can produce positive results. (New Scientist $)
+ This AI system makes human tutors better at teaching children math. (MIT Technology Review)

6 Robots still struggle to match warehouse workers on some tasks
For all the advances robots have made, picking things up and moving them around remains a big challenge. (NYT $)
+ AI is poised to automate today’s most mundane manual warehouse task. (MIT Technology Review)

7 Perplexity’s AI search engine can now buy stuff for you
How long until Google follows? (The Verge)

8 Dozens of states are begging Congress to pass the kids online safety act
It has currently stalled in the House of Representatives due to censorship concerns. (The Verge)
Roblox is adding more controls to let parents set daily usage limits, block access to certain game genres, and more. (WSJ $)
+ Why child safety bills are popping up all over the US

9 The US Patent and Trademark Office banned staff from using generative AI
It cited security concerns plus the fact some tools exhibit “bias, unpredictability, and malicious behavior.” (Wired $)

10 NASA might have killed life on Mars 😬
A new paper suggests that adding water to Martian soil might have been a bad move. (Quartz $)
+  The ISS has been leaking air for 5 years, and engineers still can’t agree why. (Ars Technica)

Quote of the day

“We are bleeding cash as an industry.” 

—Thomas Laffont, co-founder of investment firm Coatue Management, says venture capital firms are struggling to make money amid a boom in AI investments, the Wall Street Journal reports.

 The big story

How mobile money supercharged Kenya’s sports betting addiction

BRIAN OTIENO

April 2022

Mobile money has mostly been hugely beneficial for Kenyans. But it has also turbo-charged the country’s sports betting sector.

Experts and public figures across the African continent are sounding the alarm over the growth of the sector increasingly loudly. It’s produced tales of riches, but it has also broken families, consumed college tuitions, and even driven some to suicide. Read the full story.

—Jonathan W. Rosen

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ I just learned a pertinent word for this season: abscission
+ Only some people will get this… but if you’re one of them, you’ll enjoy it. 
+ Why Late of the Pier were one of the most exciting UK bands of the 2000s.
+ Whether you call them crisps or chips, they’re goddamn delicious.

The rise of Bluesky, and the splintering of social

You may have read that it was a big week for Bluesky

If you’re not familiar, Bluesky is, essentially, a Twitter clone that publishes short-form status updates. It gained more than 2 million users this week. On Wednesday, The Verge reported it had crossed 15 million users. By Thursday, it was at 16 million. By Friday? 17 million and counting. It was the number one app in Apple’s app store. 

Meanwhile, Threads, Meta’s answer to Twitter, put up even bigger numbers. The company’s Adam Mosseri reported that 15 million people had signed up in November alone. Both apps are surging in usage. 

Many of these new users were seemingly fleeing X, the platform formerly known as Twitter. On the day after the election, more than 115,000 people deactivated their X accounts, according to Similarweb data. That’s a step far past not logging on. It means giving up your username and social graph. It’s nuking your account versus just ignoring it. 

Much of that migration is likely a reaction to Elon Musk’s support of Donald Trump, and his moves to elevate right-leaning content on the platform. Since Musk took over, X has reinstated a lot of previously banned accounts, very many of which are on the far right. It also tweaked its algorithm to make sure Musk’s own posts, which are often pro-Trump, get an extra level of promotion and prominence, according to Kate Conger and Ryan Mac’s new book Character Limit

There are two points I want to make here. The first is that tech and politics are just entirely enmeshed at this point. That’s due to the extreme extent to which tech has captured culture and the economy. Everything is a tech story now, including and especially politics. 

The second point is about what I see as a more long-term shift away from centralization. What’s more interesting to me than people fleeing a service because they don’t like its politics is the emergence of unique experiences and cultures across all three of these services, as well as other, smaller competitors.

Last year, we put “Twitter killers” on our list of 10 breakthrough technologies. But the breakthrough technology wasn’t the rise of one service or the decline of another. It was decentralization. At the time, I wrote: 

“Decentralized, or federated, social media allows for communication across independently hosted servers or platforms, using networking protocols such as ActivityPub, AT Protocol, or Nostr. It offers more granular moderation, more security against the whims of a corporate master or government censor, and the opportunity to control your social graph. It’s even possible to move from one server to another and follow the same people.”

In the long run, massive, centralized social networks will prove to be an aberration. We are going to use different networks for different things. 

For example, Bluesky is great for breaking news because it does not deprioritize links and defaults to a social graph that shows updates from the people you follow in chronological order. (It also has a Discover feed and you can set up others for algorithmic discovery—more on that in a moment—but the default is the classic Twitter-esque timeline.) 

Threads, which has a more algorithmically defined experience, is great for surfacing interesting conversations from the past few days. I routinely find interesting comments and posts from two or three days before I logged on. At the same time, this makes it pretty lousy at any kind of real time experience—seemingly intentionally—and essentially hides that standard timeline of updates from people you follow in favor of an algorithmically-generated “for you” feed. 

I’m going to go out on a limb here and say that while these are quite different, neither is inherently better. They offer distinct takes on product direction. And that ability to offer different experiences is a good thing. 

I think this is one area where Bluesky has a real advantage. Bluesky lets people bend the experience to their own will. You aren’t locked into the default following and discover experiences. You can roll your own custom feed, and follow custom feeds created by other people. (And Threads is now testing something similar.) That customization means my experience on Bluesky may look nothing like yours. 

This is possible because Bluesky is a service running on top of the AT Protocol, an open protocol that’s accessible to anyone and everyone. The entire idea is that social networking is too important for any one company or person to control it. So it is set up to allow anyone to run their own network using that protocol. And that’s going to lead to a wide range of outcomes. 

Take moderation, as an example. The moderation philosophy of the AT Protocol is essentially that everyone is entitled to speech but not to reach. That means it isn’t banning content at the protocol level, but that individual services can set up their own rules. 

Bluesky has its own community guidelines. But those guidelines would not necessarily apply to other services running on the protocol. Furthermore, individuals can also moderate what types of posts they want to see. It lets people set up and choose different levels of what they want to allow. That, combined with the ability to roll your own feeds, combined with the ability of different services to run on top of the same protocol, sets up a very fragmented future. 

And that’s just Bluesky. There’s also Nostr, which leans toward the crypto and tech crowds, at least for now. And Mastodon, which tends to have clusters of communities on various servers. All of them are growing. 

The era of the centralized, canonical feed is coming to an end. What’s coming next is going to be more dispersed, more fractured, more specialized. It will take place across these decentralized services, and also WhatsApp channels, Discord servers, and other smaller slices of Big Social. That’s going to be challenging. It will cause entirely new problems. But it’s also an incredible opportunity for individuals to take more control of their own experiences.

If someone forwarded you this edition of The Debrief, you can subscribe here. I appreciate your feedback on this newsletter. Drop me a line at mat.honan@technologyreview.com with any and all thoughts. And of course, I love tips.

Now read the rest of The Debrief

The News

TSMC halts advanced chip shipments for Chinese clients. It comes after some of its chips were found inside a Huawei AI processor.

Google DeepMind has come up with a new way to peer inside AI’s thought process.

An AI lab out of Chicago is building tools to help creators prevent their work from being used in training data.

Lina Khan may be on the way out, but she’s going out with a bang: The FTC is preparing to investigate Microsoft’s cloud business.

The Chat

Every week I’ll talk to one of MIT Technology Review’s reporters or editors to find out more about what they’ve been working on. For today, I spoke with Casey Crownhart, senior climate reporter, about her coverage of the COP29 UN climate conference.

Mat: COP29 is happening right now in Azerbaijan, do you have a sense of the mood?

Casey: The vibes are weird in Baku this week, in part because of the US election. The US has been a strong leader in international climate talks in recent years, and an incoming Trump administration will certainly mean a big change.

And the main goal of these talks—reaching a climate finance agreement—is a little daunting. Developing countries need something like $1 trillion dollars annually to cope with climate change. That’s a huge jump from the current target, so there are questions about how this agreement will shake out.

Mat: Azerbaijan seems like a weird choice to host. I read one account from the conference saying you could smell the oil in the air. Why there?

Casey: Azerbaijan’s economy is super reliant on fossil fuels, which definitely makes it an ironic spot for international climate negotiations.

There’s a whole complicated process of picking the COP host each year—five regions rotate hosting, and the countries in that region have to all agree on a pick when it’s their turn. Russia apparently vetoed most of the other choices in the Eastern European group this year, and the region settled on Azerbaijan as one of the only viable options.

Mat: You write that if Trump pulls out of the UN Framework Convention on Climate Change, it would be like riding away on a rocket. Why would that be so much worse than dropping out of Paris?

Casey: Trump withdrew from the Paris Agreement once already, and it was relatively easy for Biden to rejoin when he came into office. If, during his second term, Trump were to go a step further and pull out of the UNFCCC, it’s not just an agreement he’s walking away from, it’s the whole negotiating framework. So the statement would be much bigger.

There’s also the question of reversibility. It’s not clear if Trump can actually withdraw from the UNFCCC on his own, and it’s also not clear what it would take to rejoin it. When the US joined in the ’90s, the Senate had to agree, so getting back in might not be as simple as a future president signing something.

Mat: What from COP29 are you optimistic about?

Casey: Tough to find a glimmer of hope in all this, but if there is one, I’d say I’m optimistic that we’ll see some countries step up, including the UK and China. The UK announced a new emissions target at the talks already, and it’ll be really interesting to see what role China plays at COP29 and moving forward.

The Recommendation

Once upon a time I was a gadget blogger. It’s fun writing about gadgets! I miss it! Especially because at some point your phone became the only device you need. But! My beloved wife bought me a Whoop fitness tracker for my birthday. It’s an always-on device that you wear around your wrist. I’ve been Oura-curious for some time, but frankly I am a little bit terrified of rings. I spent a number of months going to a hand rehab clinic after a bike accident, and while I was there first learned about degloving and how commonly it happens to people because a ring gets caught on something. Just thought I’d put that in your head too. Anyway! The whoop is a fabric bracelet with a little monitor on it. It tracks your movement, your heart rate, your sleep, and a lot more. There’s no screen, so it’s very low profile and unobtrusive. It is, however, pretty spendy: The device is free but the plan costs $239 annually.

The Download: Bluesky’s rapid rise, and harmful fertility stereotypes

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology.

The rise of Bluesky, and the splintering of social

You may have read that it was a big week for Bluesky. If you’re not familiar, Bluesky is, essentially, a Twitter clone that publishes short-form status updates. Last Wednesday, The Verge reported it had crossed 15 million users. It’s just ticked over 19 million now, and is the number one app in Apple’s app store.

Meanwhile, Threads, Meta’s answer to Twitter, reportedly signed up 15 million people in November alone. Both apps are surging in usage.

Many of these new users were seemingly fleeing X, the platform formerly known as Twitter, in reaction to Elon Musk’s support of Donald Trump, and his moves to elevate right-leaning content on the platform. But there’s a deeper trend at play here. We’re seeing a long-term shift away from massive centralized social networks. Read the full story

—Mat Honan

This story is from The Debrief, our newly-launched newsletter written by our editor-in-chief Mat Honan. It’s his weekly take on the real stories behind the biggest news in tech—with some links to stories we love and the occasional recommendation thrown in for good measure. Sign up to get it every Friday!

Why the term “women of childbearing age” is problematic

—Jessica Hamzelou

Every journalist has favorite topics. Mine include the quest to delay or reverse human aging, and new technologies for reproductive health and fertility. So when I saw trailers for The Substance, a film centered on one middle-aged woman’s attempt to reexperience youth, I had to watch it.

I won’t spoil the movie for anyone who hasn’t seen it yet (although I should warn that it is not for the squeamish). But a key premise of the film involves harmful attitudes toward female aging. 

“Hey, did you know that a woman’s fertility starts to decrease by the age of 25?” a powerful male character asks early in the film. “At 50, it just stops,” he later adds. He never explains what stops, exactly, but to the viewer the message is pretty clear: If you’re a woman, your worth is tied to your fertility. Once your fertile window is over, so are you. 

The insidious idea that women’s bodies are, above all else, vessels for growing children has plenty of negative consequences for us all. But it also sets back scientific research and health policy. Read Jess’s story to learn how

This story is from The Checkup, MIT Technology Review’s weekly biotech newsletter.  Sign up to receive it in your inbox every Thursday.

The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology.

1 Trump plans to loosen US rules for self-driving cars 
No prizes for guessing who might be behind that idea. (Bloomberg $)
Elon Musk is ramping up his legal fight against OpenAI and Microsoft. (WSJ $)
Trump has appointed the FCC’s Brendan Carr to lead the agency. (NPR)
Robotaxis are here. It’s time to decide what to do about them. (MIT Technology Review)

2 How Bluesky is handling its explosive growth
It has just 20 employees, and they’re working round the clock to deal with bugs, outages and moderation issues. (NYT $)
+ Just joined Bluesky? Here’s how to use it. (The Verge)
How to fix the internet. (MIT Technology Review)

 3 Biden agreed to some small but significant AI limits with Xi Jinping 
I think we can all get behind the idea that nuclear weapons should be exclusively controlled by humans. (Politico)
Biden has lifted a ban on Ukraine using long-raise missiles to strike inside Russia. (BBC)

4 Big Tech is trying to sink the US online child safety bill 
And, as it stands, its lobbying efforts look very likely to succeed. (WSJ $)

5 Amazon has launched a rival to Temu and Shein 
Nothing on ‘Haul’ costs more than $20. (BBC)
+ Welcome to the slop era of online shopping. (The Atlantic $)

6 The Mike Tyson-Jake Paul fight on Netflix was plagued by glitches
Despite that, 60 million households still tuned in. (Deadline)

7 AI models can work together faster in their own language 
Linking different models together could help tackle thorny problems individual ones can’t solve. (New Scientist $)

8 Tech companies are training their AI on movie subtitles 
A database called OpenSubtitles provides a rare glimpse into what goes into these systems. (The Atlantic $)

9 McDonald’s is trying to bring back NFTs
Remember those? (Gizmodo)

10 A lot of people are confusing Starlink satellites with UFOs
Guess it’ll take us a while for us to get used to seeing them. (Ars Technica)

Quote of the day

“F*** you, Elon Musk.”

—Brazil’s first lady, Janja Lula da Silva, makes her views clear during a speech calling for tougher social media regulation ahead of the G20 summit in Rio de Janeiro, Reuters reports.

 The big story

Alina Chan tweeted life into the idea that the virus came from a lab

Alina Chan

COURTESY PHOTO

June 2021

Alina Chan started asking questions in March 2020. She was chatting with friends on Facebook about the virus then spreading out of China. She thought it was strange that no one had found any infected animal. She wondered why no one was admitting another possibility, which to her seemed very obvious: the outbreak might have been due to a lab accident.

Chan is a postdoc in a gene therapy lab at the Broad Institute, a prestigious research institute affiliated with both Harvard and MIT. Throughout 2020, Chan relentlessly stoked scientific argument, and wasn’t afraid to pit her brain against the best virologists in the world. Her persistence even helped change some researchers’ minds. Read the full story.

—Antonio Regalado

We can still have nice things

A place for comfort, fun and distraction to brighten up your day. (Got any ideas? Drop me a line or tweet ’em at me.)

+ Why Quincy Jones was the best of the best.
+ These handy apps are a great way to save articles to read later on (Pocket is my own personal favorite.)
+ How to resurrect a ghost river in the Bronx.
+ Look after your stainless steel pans, and your stainless steel pans will look after you.

The AI lab waging a guerrilla war over exploitative AI

Ben Zhao remembers well the moment he officially jumped into the fight between artists and generative AI: when one artist asked for AI bananas. 

A computer security researcher at the University of Chicago, Zhao had made a name for himself by building tools to protect images from facial recognition technology. It was this work that caught the attention of Kim Van Deun, a fantasy illustrator who invited him to a Zoom call in November 2022 hosted by the Concept Art Association, an advocacy organization for artists working in commercial media. 

On the call, artists shared details of how they had been hurt by the generative AI boom, which was then brand new. At that moment, AI was suddenly everywhere. The tech community was buzzing over image-generating AI models, such as Midjourney, Stable Diffusion, and OpenAI’s DALL-E 2, which could follow simple word prompts to depict fantasylands or whimsical chairs made of avocados. 

But these artists saw this technological wonder as a new kind of theft. They felt the models were effectively stealing and replacing their work. Some had found that their art had been scraped off the internet and used to train the models, while others had discovered that their own names had become prompts, causing their work to be drowned out online by AI knockoffs.

Zhao remembers being shocked by what he heard. “People are literally telling you they’re losing their livelihoods,” he told me one afternoon this spring, sitting in his Chicago living room. “That’s something that you just can’t ignore.” 

So on the Zoom, he made a proposal: What if, hypothetically, it was possible to build a mechanism that would help mask their art to interfere with AI scraping?

“I would love a tool that if someone wrote my name and made a prompt, like, garbage came out,” responded Karla Ortiz, a prominent digital artist. “Just, like, bananas or some weird stuff.” 

That was all the convincing Zhao needed—the moment he joined the cause.

Fast-forward to today, and millions of artists have deployed two tools born from that Zoom: Glaze and Nightshade, which were developed by Zhao and the University of Chicago’s SAND Lab (an acronym for “security, algorithms, networking, and data”).

Arguably the most prominent weapons in an artist’s arsenal against nonconsensual AI scraping, Glaze and Nightshade work in similar ways: by adding what the researchers call “barely perceptible” perturbations to an image’s pixels so that machine-learning models cannot read them properly. Glaze, which has been downloaded more than 6 million times since it launched in March 2023, adds what’s effectively a secret cloak to images that prevents AI algorithms from picking up on and copying an artist’s style. Nightshade, which I wrote about when it was released almost exactly a year ago this fall, cranks up the offensive against AI companies by adding an invisible layer of poison to images, which can break AI models; it has been downloaded more than 1.6 million times. 

Thanks to the tools, “I’m able to post my work online,” Ortiz says, “and that’s pretty huge.” For artists like her, being seen online is crucial to getting more work. If they are uncomfortable about ending up in a massive for-profit AI model without compensation, the only option is to delete their work from the internet. That would mean career suicide. “It’s really dire for us,” adds Ortiz, who has become one of the most vocal advocates for fellow artists and is part of a class action lawsuit against AI companies, including Stability AI, over copyright infringement. 

But Zhao hopes that the tools will do more than empower individual artists. Glaze and Nightshade are part of what he sees as a battle to slowly tilt the balance of power from large corporations back to individual creators. 

“It is just incredibly frustrating to see human life be valued so little,” he says with a disdain that I’ve come to see as pretty typical for him, particularly when he’s talking about Big Tech. “And to see that repeated over and over, this prioritization of profit over humanity … it is just incredibly frustrating and maddening.” 

As the tools are adopted more widely, his lofty goal is being put to the test. Can Glaze and Nightshade make genuine security accessible for creators—or will they inadvertently lull artists into believing their work is safe, even as the tools themselves become targets for haters and hackers? While experts largely agree that the approach is effective and Nightshade could prove to be powerful poison, other researchers claim they’ve already poked holes in the protections offered by Glaze and that trusting these tools is risky. 

But Neil Turkewitz, a copyright lawyer who used to work at the Recording Industry Association of America, offers a more sweeping view of the fight the SAND Lab has joined. It’s not about a single AI company or a single individual, he says: “It’s about defining the rules of the world we want to inhabit.” 

Poking the bear

The SAND Lab is tight knit, encompassing a dozen or so researchers crammed into a corner of the University of Chicago’s computer science building. That space has accumulated somewhat typical workplace detritus—a Meta Quest headset here, silly photos of dress-up from Halloween parties there. But the walls are also covered in original art pieces, including a framed painting by Ortiz.  

Years before fighting alongside artists like Ortiz against “AI bros” (to use Zhao’s words), Zhao and the lab’s co-leader, Heather Zheng, who is also his wife, had built a record of combating harms posed by new tech. 

group of students and teachers posing in Halloween costumes
When I visited the SAND Lab in Chicago, I saw how tight knit the group was. Alongside the typical workplace stuff were funny Halloween photos like this one. (Front row: Ronik Bhaskar, Josephine Passananti, Anna YJ Ha, Zhuolin Yang, Ben Zhao, Heather Zheng. Back row: Cathy Yuanchen Li, Wenxin Ding, Stanley Wu, and Shawn Shan.)
COURTESY OF SAND LAB

Though both earned spots on MIT Technology Review’s 35 Innovators Under 35 list for other work nearly two decades ago, when they were at the University of California, Santa Barbara (Zheng in 2005 for “cognitive radios” and Zhao a year later for peer-to-peer networks), their primary research focus has become security and privacy. 

The pair left Santa Barbara in 2017, after they were poached by the new co-director of the University of Chicago’s Data Science Institute, Michael Franklin. All eight PhD students from their UC Santa Barbara lab decided to follow them to Chicago too. Since then, the group has developed a “bracelet of silence” that jams the microphones in AI voice assistants like the Amazon Echo. It has also created a tool called Fawkes—“privacy armor,” as Zhao put it in a 2020 interview with the New York Times—that people can apply to their photos to protect them from facial recognition software. They’ve also studied how hackers might steal sensitive information through stealth attacks on virtual-reality headsets, and how to distinguish human art from AI-generated images. 

“Ben and Heather and their group are kind of unique because they’re actually trying to build technology that hits right at some key questions about AI and how it is used,” Franklin tells me. “They’re doing it not just by asking those questions, but by actually building technology that forces those questions to the forefront.”

It was Fawkes that intrigued Van Deun, the fantasy illustrator, two years ago; she hoped something similar might work as protection against generative AI, which is why she extended that fateful invite to the Concept Art Association’s Zoom call. 

That call started something of a mad rush in the weeks that followed. Though Zhao and Zheng collaborate on all the lab’s projects, they each lead individual initiatives; Zhao took on what would become Glaze, with PhD student Shawn Shan (who was on this year’s Innovators Under 35 list) spearheading the development of the program’s algorithm. 

In parallel to Shan’s coding, PhD students Jenna Cryan and Emily Wenger sought to learn more about the views and needs of the artists themselves. They created a user survey that the team distributed to artists with the help of Ortiz. In replies from more than 1,200 artists—far more than the average number of responses to user studies in computer science—the team found that the vast majority of creators had read about art being used to train models, and 97% expected AI to decrease some artists’ job security. A quarter said AI art had already affected their jobs. 

Almost all artists also said they posted their work online, and more than half said they anticipated reducing or removing that online work, if they hadn’t already—no matter the professional and financial consequences.

The first scrappy version of Glaze was developed in just a month, at which point Ortiz gave the team her entire catalogue of work to test the model on. At the most basic level, Glaze acts as a defensive shield. Its algorithm identifies features from the image that make up an artist’s individual style and adds subtle changes to them. When an AI model is trained on images protected with Glaze, the model will not be able to reproduce styles similar to the original image. 

A painting from Ortiz later became the first image publicly released with Glaze on it: a young woman, surrounded by flying eagles, holding up a wreath. Its title is Musa Victoriosa, “victorious muse.” 

It’s the one currently hanging on the SAND Lab’s walls. 

Despite many artists’ initial enthusiasm, Zhao says, Glaze’s launch caused significant backlash. Some artists were skeptical because they were worried this was a scam or yet another data-harvesting campaign. 

The lab had to take several steps to build trust, such as offering the option to download the Glaze app so that it adds the protective layer offline, which meant no data was being transferred anywhere. (The images are then shielded when artists upload them.)  

Soon after Glaze’s launch, Shan also led the development of the second tool, Nightshade. Where Glaze is a defensive mechanism, Nightshade was designed to act as an offensive deterrent to nonconsensual training. It works by changing the pixels of images in ways that are not noticeable to the human eye but manipulate machine-learning models so they interpret the image as something different from what it actually shows. If poisoned samples are scraped into AI training sets, these samples trick the AI models: Dogs become cats, handbags become toasters. The researchers say only a relatively few examples are enough to permanently damage the way a generative AI model produces images.

Currently, both tools are available as free apps or can be applied through the project’s website. The lab has also recently expanded its reach by offering integration with the new artist-supported social network Cara, which was born out of a backlash to exploitative AI training and forbids AI-produced content.

In dozens of conversations with Zhao and the lab’s researchers, as well as a handful of their artist-collaborators, it’s become clear that both groups now feel they are aligned in one mission. “I never expected to become friends with scientists in Chicago,” says Eva Toorenent, a Dutch artist who worked closely with the team on Nightshade. “I’m just so happy to have met these people during this collective battle.” 

Belladonna artwork shows a central character with a skull head in a dark forest illuminated around them by the belladonna flower slung over their shoulder
Images online of Toorenent’s Belladonna have been treated with the SAND Lab’s Nightshade tool.
EVA TOORENENT

Her painting Belladonna, which is also another name for the nightshade plant, was the first image with Nightshade’s poison on it. 

“It’s so symbolic,” she says. “People taking our work without our consent, and then taking our work without consent can ruin their models. It’s just poetic justice.” 

No perfect solution

The reception of the SAND Lab’s work has been less harmonious across the AI community.

After Glaze was made available to the public, Zhao tells me, someone reported it to sites like VirusTotal, which tracks malware, so that it was flagged by antivirus programs. Several people also started claiming on social media that the tool had quickly been broken. Nightshade similarly got a fair share of criticism when it launched; as TechCrunch reported in January, some called it a “virus” and, as the story explains, “another Reddit user who inadvertently went viral on X questioned Nightshade’s legality, comparing it to ‘hacking a vulnerable computer system to disrupt its operation.’” 

“We had no idea what we were up against,” Zhao tells me. “Not knowing who or what the other side could be meant that every single new buzzing of the phone meant that maybe someone did break Glaze.” 

Both tools, though, have gone through rigorous academic peer review and have won recognition from the computer security community. Nightshade was accepted at the IEEE Symposium on Security and Privacy, and Glaze received a distinguished paper award and the 2023 Internet Defense Prize at the Usenix Security Symposium, a top conference in the field. 

“In my experience working with poison, I think [Nightshade is] pretty effective,” says Nathalie Baracaldo, who leads the AI security and privacy solutions team at IBM and has studied data poisoning. “I have not seen anything yet—and the word yet is important here—that breaks that type of defense that Ben is proposing.” And the fact that the team has released the source code for Nightshade for others to probe, and it hasn’t been broken, also suggests it’s quite secure, she adds. 

At the same time, at least one team of researchers does claim to have penetrated the protections of Glaze, or at least an old version of it. 

As researchers from Google DeepMind and ETH Zurich detailed in a paper published in June, they found various ways Glaze (as well as similar but less popular protection tools, such as Mist and Anti-DreamBooth) could be circumvented using off-the-shelf techniques that anyone could access—such as image upscaling, meaning filling in pixels to increase the resolution of an image as it’s enlarged. The researchers write that their work shows the “brittleness of existing protections” and warn that “artists may believe they are effective. But our experiments show they are not.”

Florian Tramèr, an associate professor at ETH Zurich who was part of the study, acknowledges that it is “very hard to come up with a strong technical solution that ends up really making a difference here.” Rather than any individual tool, he ultimately advocates for an almost certainly unrealistic ideal: stronger policies and laws to help create an environment in which people commit to buying only human-created art. 

What happened here is common in security research, notes Baracaldo: A defense is proposed, an adversary breaks it, and—ideally—the defender learns from the adversary and makes the defense better. “It’s important to have both ethical attackers and defenders working together to make our AI systems safer,” she says, adding that “ideally, all defenses should be publicly available for scrutiny,” which would both “allow for transparency” and help avoid creating a false sense of security. (Zhao, though, tells me the researchers have no intention to release Glaze’s source code.)

Still, even as all these researchers claim to support artists and their art, such tests hit a nerve for Zhao. In Discord chats that were later leaked, he claimed that one of the researchers from the ETH Zurich–Google DeepMind team “doesn’t give a shit” about people. (That researcher did not respond to a request for comment, but in a blog post he said it was important to break defenses in order to know how to fix them. Zhao says his words were taken out of context.) 

Zhao also emphasizes to me that the paper’s authors mainly evaluated an earlier version of Glaze; he says its new update is more resistant to tampering. Messing with images that have current Glaze protections would harm the very style that is being copied, he says, making such an attack useless. 

This back-and-forth reflects a significant tension in the computer security community and, more broadly, the often adversarial relationship between different groups in AI. Is it wrong to give people the feeling of security when the protections you’ve offered might break? Or is it better to have some level of protection—one that raises the threshold for an attacker to inflict harm—than nothing at all? 

Yves-Alexandre de Montjoye, an associate professor of applied mathematics and computer science at Imperial College London, says there are plenty of examples where similar technical protections have failed to be bulletproof. For example, in 2023, de Montjoye and his team probed a digital mask for facial recognition algorithms, which was meant to protect the privacy of medical patients’ facial images; they were able to break the protections by tweaking just one thing in the program’s algorithm (which was open source). 

Using such defenses is still sending a message, he says, and adding some friction to data profiling. “Tools such as TrackMeNot”—which protects users from data profiling—“have been presented as a way to protest; as a way to say I do not consent.”  

“But at the same time,” he argues, “we need to be very clear with artists that it is removable and might not protect against future algorithms.”

While Zhao will admit that the researchers pointed out some of Glaze’s weak spots, he unsurprisingly remains confident that Glaze and Nightshade are worth deploying, given that “security tools are never perfect.” Indeed, as Baracaldo points out, the Google DeepMind and ETH Zurich researchers showed how a highly motivated and sophisticated adversary will almost certainly always find a way in.

Yet it is “simplistic to think that if you have a real security problem in the wild and you’re trying to design a protection tool, the answer should be it either works perfectly or don’t deploy it,” Zhao says, citing spam filters and firewalls as examples. Defense is a constant cat-and-mouse game. And he believes most artists are savvy enough to understand the risk. 

Offering hope

The fight between creators and AI companies is fierce. The current paradigm in AI is to build bigger and bigger models, and there is, at least currently, no getting around the fact that they require vast data sets hoovered from the internet to train on. Tech companies argue that anything on the public internet is fair game, and that it is “impossible” to build advanced AI tools without copyrighted material; many artists argue that tech companies have stolen their intellectual property and violated copyright law, and that they need ways to keep their individual works out of the models—or at least receive proper credit and compensation for their use. 

So far, the creatives aren’t exactly winning. A number of companies have already replaced designers, copywriters, and illustrators with AI systems. In one high-profile case, Marvel Studios used AI-generated imagery instead of human-created art in the title sequence of its 2023 TV series Secret Invasion. In another, a radio station fired its human presenters and replaced them with AI. The technology has become a major bone of contention between unions and film, TV, and creative studios, most recently leading to a strike by video-game performers. There are numerous ongoing lawsuits by artists, writers, publishers, and record labels against AI companies. It will likely take years until there is a clear-cut legal resolution. But even a court ruling won’t necessarily untangle the difficult ethical questions created by generative AI. Any future government regulation is not likely to either, if it ever materializes. 

That’s why Zhao and Zheng see Glaze and Nightshade as necessary interventions—tools to defend original work, attack those who would help themselves to it, and, at the very least, buy artists some time. Having a perfect solution is not really the point. The researchers need to offer something now because the AI sector moves at breakneck speed, Zheng says, means that companies are ignoring very real harms to humans. “This is probably the first time in our entire technology careers that we actually see this much conflict,” she adds.

On a much grander scale, she and Zhao tell me they hope that Glaze and Nightshade will eventually have the power to overhaul how AI companies use art and how their products produce it. It is eye-wateringly expensive to train AI models, and it’s extremely laborious for engineers to find and purge poisoned samples in a data set of billions of images. Theoretically, if there are enough Nightshaded images on the internet and tech companies see their models breaking as a result, it could push developers to the negotiating table to bargain over licensing and fair compensation. 

That’s, of course, still a big “if.” MIT Technology Review reached out to several AI companies, such as Midjourney and Stability AI, which did not reply to requests for comment. A spokesperson for OpenAI, meanwhile, did not confirm any details about encountering data poison but said the company takes the safety of its products seriously and is continually improving its safety measures: “We are always working on how we can make our systems more robust against this type of abuse.”

In the meantime, the SAND Lab is moving ahead and looking into funding from foundations and nonprofits to keep the project going. They also say there has also been interest from major companies looking to protect their intellectual property (though they decline to say which), and Zhao and Zheng are exploring how the tools could be applied in other industries, such as gaming, videos, or music. In the meantime, they plan to keep updating Glaze and Nightshade to be as robust as possible, working closely with the students in the Chicago lab—where, on another wall, hangs Toorenent’s Belladonna. The painting has a heart-shaped note stuck to the bottom right corner: “Thank you! You have given hope to us artists.”

This story has been updated with the latest download figures for Glaze and Nightshade.

Google DeepMind has a new way to look inside an AI’s “mind”

AI has led to breakthroughs in drug discovery and robotics and is in the process of entirely revolutionizing how we interact with machines and the web. The only problem is we don’t know exactly how it works, or why it works so well. We have a fair idea, but the details are too complex to unpick. That’s a problem: It could lead us to deploy an AI system in a highly sensitive field like medicine without understanding that it could have critical flaws embedded in its workings.

A team at Google DeepMind that studies something called mechanistic interpretability has been working on new ways to let us peer under the hood. At the end of July, it released Gemma Scope, a tool to help researchers understand what is happening when AI is generating an output. The hope is that if we have a better understanding of what is happening inside an AI model, we’ll be able to control its outputs more effectively, leading to better AI systems in the future.

“I want to be able to look inside a model and see if it’s being deceptive,” says Neel Nanda, who runs the mechanistic interpretability team at Google DeepMind. “It seems like being able to read a model’s mind should help.”

Mechanistic interpretability, also known as “mech interp,” is a new research field that aims to understand how neural networks actually work. At the moment, very basically, we put inputs into a model in the form of a lot of data, and then we get a bunch of model weights at the end of training. These are the parameters that determine how a model makes decisions. We have some idea of what’s happening between the inputs and the model weights: Essentially, the AI is finding patterns in the data and making conclusions from those patterns, but these patterns can be incredibly complex and often very hard for humans to interpret.

It’s like a teacher reviewing the answers to a complex math problem on a test. The student—the AI, in this case—wrote down the correct answer, but the work looks like a bunch of squiggly lines. This example assumes the AI is always getting the correct answer, but that’s not always true; the AI student may have found an irrelevant pattern that it’s assuming is valid. For example, some current AI systems will give you the result that 9.11 is bigger than 9.8. Different methods developed in the field of mechanistic interpretability are beginning to shed a little bit of light on what may be happening, essentially making sense of the squiggly lines.

“A key goal of mechanistic interpretability is trying to reverse-engineer the algorithms inside these systems,” says Nanda. “We give the model a prompt, like ‘Write a poem,’ and then it writes some rhyming lines. What is the algorithm by which it did this? We’d love to understand it.”

To find features—or categories of data that represent a larger concept—in its AI model, Gemma, DeepMind ran a tool known as a “sparse autoencoder” on each of its layers. You can think of a sparse autoencoder as a microscope that zooms in on those layers and lets you look at their details. For example, if you prompt Gemma about a chihuahua, it will trigger the “dogs” feature, lighting up what the model knows about “dogs.” The reason it is considered “sparse” is that it’s limiting the number of neurons used, basically pushing for a more efficient and generalized representation of the data.

The tricky part of sparse autoencoders is deciding how granular you want to get. Think again about the microscope. You can magnify something to an extreme degree, but it may make what you’re looking at impossible for a human to interpret. But if you zoom too far out, you may be limiting what interesting things you can see and discover. 

DeepMind’s solution was to run sparse autoencoders of different sizes, varying the number of features they want the autoencoder to find. The goal was not for DeepMind’s researchers to thoroughly analyze the results on their own. Gemma and the autoencoders are open-source, so this project was aimed more at spurring interested researchers to look at what the sparse autoencoders found and hopefully make new insights into the model’s internal logic. Since DeepMind ran autoencoders on each layer of their model, a researcher could map the progression from input to output to a degree we haven’t seen before.

“This is really exciting for interpretability researchers,” says Josh Batson, a researcher at Anthropic. “If you have this model that you’ve open-sourced for people to study, it means that a bunch of interpretability research can now be done on the back of those sparse autoencoders. It lowers the barrier to entry to people learning from these methods.”

Neuronpedia, a platform for mechanistic interpretability, partnered with DeepMind in July to build a demo of Gemma Scope that you can play around with right now. In the demo, you can test out different prompts and see how the model breaks up your prompt and what activations your prompt lights up. You can also mess around with the model. For example, if you turn the feature about dogs way up and then ask the model a question about US presidents, Gemma will find some way to weave in random babble about dogs, or the model may just start barking at you.

One interesting thing about sparse autoencoders is that they are unsupervised, meaning they find features on their own. That leads to surprising discoveries about how the models break down human concepts. “My personal favorite feature is the cringe feature,” says Joseph Bloom, science lead at Neuronpedia. “It seems to appear in negative criticism of text and movies. It’s just a great example of tracking things that are so human on some level.” 

You can search for concepts on Neuronpedia and it will highlight what features are being activated on specific tokens, or words, and how strongly each one is activated. “If you read the text and you see what’s highlighted in green, that’s when the model thinks the cringe concept is most relevant. The most active example for cringe is somebody preaching at someone else,” says Bloom.

Some features are proving easier to track than others. “One of the most important features that you would want to find for a model is deception,” says Johnny Lin, founder of Neuronpedia. “It’s not super easy to find: ‘Oh, there’s the feature that fires when it’s lying to us.’ From what I’ve seen, it hasn’t been the case that we can find deception and ban it.”

DeepMind’s research is similar to what another AI company, Anthropic, did back in May with Golden Gate Claude. It used sparse autoencoders to find the parts of Claude, their model, that lit up when discussing the Golden Gate Bridge in San Francisco. It then amplified the activations related to the bridge to the point where Claude literally identified not as Claude, an AI model, but as the physical Golden Gate Bridge and would respond to prompts as the bridge.

Although it may just seem quirky, mechanistic interpretability research may prove incredibly useful. “As a tool for understanding how the model generalizes and what level of abstraction it’s working at, these features are really helpful,” says Batson.

For example, a team lead by Samuel Marks, now at Anthropic, used sparse autoencoders to find features that showed a particular model was associating certain professions with a specific gender. They then turned off these gender features to reduce bias in the model. This experiment was done on a very small model, so it’s unclear if the work will apply to a much larger model.

Mechanistic interpretability research can also give us insights into why AI makes errors. In the case of the assertion that 9.11 is larger than 9.8, researchers from Transluce saw that the question was triggering the parts of an AI model related to Bible verses and September 11. The researchers concluded the AI could be interpreting the numbers as dates, asserting the later date, 9/11, as greater than 9/8. And in a lot of books like religious texts, section 9.11 comes after section 9.8, which may be why the AI thinks of it as greater. Once they knew why the AI made this error, the researchers tuned down the AI’s activations on Bible verses and September 11, which led to the model giving the correct answer when prompted again on whether 9.11 is larger than 9.8.

There are also other potential applications. Currently, a system-level prompt is built into LLMs to deal with situations like users who ask how to build a bomb. When you ask ChatGPT a question, the model is first secretly prompted by OpenAI to refrain from telling you how to make bombs or do other nefarious things. But it’s easy for users to jailbreak AI models with clever prompts, bypassing any restrictions. 

If the creators of the models are able to see where in an AI the bomb-building knowledge is, they can theoretically turn off those nodes permanently. Then even the most cleverly written prompt wouldn’t elicit an answer about how to build a bomb, because the AI would literally have no information about how to build a bomb in its system.

This type of granularity and precise control are easy to imagine but extremely hard to achieve with the current state of mechanistic interpretability. 

“A limitation is the steering [influencing a model by adjusting its parameters] is just not working that well, and so when you steer to reduce violence in a model, it ends up completely lobotomizing its knowledge in martial arts. There’s a lot of refinement to be done in steering,” says Lin. The knowledge of “bomb making,” for example, isn’t just a simple on-and-off switch in an AI model. It most likely is woven into multiple parts of the model, and turning it off would probably involve hampering the AI’s knowledge of chemistry. Any tinkering may have benefits but also significant trade-offs.

That said, if we are able to dig deeper and peer more clearly into the “mind” of AI, DeepMind and others are hopeful that mechanistic interpretability could represent a plausible path to alignment—the process of making sure AI is actually doing what we want it to do.

What’s on the table at this year’s UN climate conference

This article is from The Spark, MIT Technology Review’s weekly climate newsletter. To receive it in your inbox every Wednesday, sign up here.

It’s time for a party—the Conference of the Parties, that is. Talks kicked off this week at COP29 in Baku, Azerbaijan. Running for a couple of weeks each year, the global summit is the largest annual meeting on climate change.

The issue on the table this time around: Countries need to agree to set a new goal on how much money should go to developing countries to help them finance the fight against climate change. Complicating things? A US president-elect whose approach to climate is very different from that of the current administration (understatement of the century).

This is a big moment that could set the tone for what the next few years of the international climate world looks like. Here’s what you need to know about COP29 and how Donald Trump’s election is coloring things.

The UN COP meetings are an annual chance for nearly 200 nations to get together to discuss (and hopefully act on) climate change. Greatest hits from the talks include the Paris Agreement, a 2015 global accord that set a goal to limit global warming to 1.5 °C (2.7 °F) above preindustrial levels.

This year, the talks are in Azerbaijan, a petrostate if there ever was one. Oil and gas production makes up over 90% of the country’s export revenue and nearly half its GDP as of 2022. A perfectly ironic spot for a global climate summit!

The biggest discussion this year centers on global climate finance—specifically, how much of it is needed to help developing countries address climate change and adapt to changing conditions. The current goal, set in 2009, is for industrialized countries to provide $100 billion each year to developing nations. The deadline was 2020, and that target was actually met for the first time in 2022, according to the Organization for Economic Cooperation and Development, which keeps track of total finance via reports from contributing countries. Currently, most of that funding is in the form of public loans and grants.

The thing is, that $100 billion number was somewhat arbitrary—in Paris in 2015, countries agreed that a new, larger target should be set in 2025 to take into account how much countries actually need.

It’s looking as if the magic number is somewhere around $1 trillion each year. However, it remains to be seen how this goal will end up shaking out, because there are disagreements about basically every part of this. What should the final number be? What kind of money should count—just public funds, or private investments as well? Which nations should pay? How long will this target stand? What, exactly, would this money be going toward?

Working out all those details is why nations are gathering right now. But one shadow looming over these negotiations is the impending return of Donald Trump.

As I covered last week, Trump’s election will almost certainly result in less progress on cutting emissions than we might have seen under a more climate-focused administration. But arguably an even bigger deal than domestic progress (or lack thereof) will be how Trump shifts the country’s climate position on the international stage.

The US has emitted more carbon pollution into the atmosphere than any other country, it currently leads the world in per capita emissions, and it’s the world’s richest economy. If anybody should be a leader at the table in talks about climate finance, it’s the US. And yet, Trump is coming into power soon, and we’ve all seen this film before. 

Last time Trump was in office, he pulled the US out of the Paris Agreement. He’s made promises to do it again—and could go one step further by backing out of the UN Framework Convention on Climate Change (UNFCCC) altogether. If leaving the Paris Agreement is walking away from the table, withdrawing from the UNFCCC is like hopping on a rocket and blasting in a different direction. It’s a more drastic action and could be tougher to reverse in the future, though experts also aren’t sure if Trump could technically do this on his own.

The uncertainty of what happens next in the US is a cloud hanging over these negotiations. “This is going to be harder because we don’t have a dynamic and pushy and confident US helping us on climate action,” said Camilla Born, an independent climate advisor and former UK senior official at COP26, during an online event last week hosted by Carbon Brief.

Some experts are confident that others will step up to fill the gap. “There are many drivers of climate action beyond the White House,” said Mohamed Adow, founding director of Power Shift Africa, at the CarbonBrief event.

If I could characterize the current vibe in the climate world, it’s uncertainty. But the negotiations over the next couple of weeks could provide clues to what we can expect for the next few years. Just how much will a Trump presidency slow global climate action? Will the European Union step up? Could this cement the rise of China as a climate leader? We’ll be watching it all.


Now read the rest of The Spark

Related reading

In case you want some additional context from the last few years of these meetings, here’s my coverage of last year’s fight at COP28 over a transition away from fossil fuels, and a newsletter about negotiations over the “loss and damages” fund at COP27.

For the nitty-gritty details about what’s on the table at COP29, check out this very thorough explainer from Carbon Brief.

The White House in Washington DC under dark stormy clouds

DAN THORNBERG/ADOBE STOCK

Another thing

Trump’s election will have significant ripple effects across the economy and our lives. His victory is a tragic loss for climate progress, as my colleague James Temple wrote in an op-ed last week. Give it a read, if you haven’t already, to dig into some of the potential impacts we might see over the next four years and beyond. 

Keeping up with climate  

The US Environmental Protection Agency finalized a rule to fine oil and gas companies for methane emissions. The fee was part of the Inflation Reduction Act of 2022. (Associated Press)
→ This rule faces a cloudy future under the Trump administration; industry groups are already talking about repealing it. (NPR)

Speaking of the EPA, Donald Trump chose Lee Zeldin, a former Republican congressman from New York, to lead the agency. Zeldin isn’t particularly known for climate or economic policy. (New York Times)

Oil giant BP is scaling back its early-stage hydrogen projects. The company revealed in an earnings report that it’s canceling 18 such projects and currently plans to greenlight between five and 10. (TechCrunch)

Investors betting against renewable energy scored big last week, earning nearly $1.2 billion as stocks in that sector tumbled. (Financial Times)

Lithium iron phosphate batteries are taking over the world, or at least electric vehicles. These lithium-ion batteries are cheaper and longer-lasting than their nickel-containing cousins, though they also tend to be heavier. (Canary Media
→ I wrote about this trend last year in a newsletter about batteries and their ingredients. (MIT Technology Review)

The US unveiled plans to triple its nuclear energy capacity by 2050. That’s an additional 200 gigawatts’ worth of consistently available power. (Bloomberg)

Five subsea cables that can help power millions of homes just got the green light in Great Britain. The projects will help connect the island to other power grids, as well as to offshore wind farms in Dutch and Belgian waters. (The Guardian)

How this grassroots effort could make AI voices more diverse

We are on the cusp of a voice AI boom, with tech companies such as Apple and OpenAI rolling out the next generation of artificial-intelligence-powered assistants. But the default voices for these assistants are often white American—British, if you’re lucky—and most definitely speak English. They represent only a tiny proportion of the many dialects and accents in the English language, which spans many regions and cultures. And if you’re one of the billions of people who don’t speak English, bad luck: These tools don’t sound nearly as good in other languages.

This is because the data that has gone into training these models is limited. In AI research, most data used to train models is extracted from the English-language internet, which reflects Anglo-American culture. But there is a massive grassroots effort underway to change this status quo and bring more transparency and diversity to what AI sounds like: Mozilla’s Common Voice initiative. 

The data set Common Voice has created over the past seven years is one of the most useful resources for people wanting to build voice AI. It has seen a massive spike in downloads, partly thanks to the current AI boom; it recently hit the 5 million mark, up from 38,500 in 2020. Creating this data set has not been easy, mainly because the data collection relies on an army of volunteers. Their numbers have also jumped, from just under 500,000 in 2020 to over 900,000 in 2024. But by giving its data away, some members of this community argue, Mozilla is encouraging volunteers to effectively do free labor for Big Tech. 

Since 2017, volunteers for the Common Voice project have collected a total of 31,000 hours of voice data in around 180 languages as diverse as Russian, Catalan, and Marathi. If you’ve used a service that uses audio AI, it’s likely been trained at least partly on Common Voice. 

Mozilla’s cause is a noble one. As AI is integrated increasingly into our lives and the ways we communicate, it becomes more important that the tools we interact with sound like us. The technology could break down communication barriers and help convey information in a compelling way to, for example, people who can’t read. But instead, an intense focus on English risks entrenching a new colonial world order and wiping out languages entirely.

“It would be such an own goal if, rather than finally creating truly multimodal, multilingual, high-performance translation models and making a more multilingual world, we actually ended up forcing everybody to operate in, like, English or French,” says EM Lewis-Jong, a director for Common Voice. 

Common Voice is open source, which means anyone can see what has gone into the data set, and users can do whatever they want with it for free. This kind of transparency is unusual in AI data governance. Most large audio data sets simply aren’t publicly available, and many consist of data that has been scraped from sites like YouTube, according to research conducted by a team from the University of Washington, and Carnegie Mellon andNorthwestern universities. 

The vast majority of language data is collected by volunteers such as Bülent Özden, a researcher from Turkey. Since 2020, he has been not only donating his voice but also raising awareness around the project to get more people to donate. He recently spent two full-time months correcting data and checking for typos in Turkish. For him, improving AI models is not the only motivation to do this work. 

“I’m doing it to preserve cultures, especially low-resource [languages],” Özden says. He tells me he has recently started collecting samples of Turkey’s smaller languages, such as Circassian and Zaza.

However, as I dug into the data set, I noticed that the coverage of languages and accents is very uneven. There are only 22 hours of Finnish voices from 231 people. In comparison, the data set contains 3,554 hours of English from 94,665 speakers. Some languages, such as Korean and Punjabi, are even less well represented. Even though they have tens of millions of speakers, they account for only a couple of hours of recorded data. 

This imbalance has emerged because data collection efforts are started from the bottom up by language communities themselves, says Lewis-Jong. 

“We’re trying to give communities what they need to create their own AI training data sets. We have a particular focus on doing this for language communities where there isn’t any data, or where maybe larger tech organizations might not be that interested in creating those data sets,” Lewis-Jong says. They hope that with the help of volunteers and various bits of grant funding, the Common Voice data set will have close to 200 languages by the end of the year.

Common Voice’s permissive license means that many companies rely on it—for example, the Swedish startup Mabel AI, which builds translation tools for health-care providers. One of the first languages the company used was Ukrainian; it built a translation tool to help Ukrainian refugees interact with Swedish social services, says Karolina Sjöberg, Mabel AI’s founder and CEO. The team has since expanded to other languages, such as Arabic and Russian. 

The problem with a lot of other audio data is that it consists of people reading from books or texts. The result is very different from how people really speak, especially when they are distressed or in pain, Sjöberg says. Because anyone can submit sentences to Common Voice for others to read aloud, Mozilla’s data set also includes sentences that are more colloquial and feel more natural, she says.

Not that it is perfectly representative. The Mabel AI team soon found out that most voice data in the languages it needed was donated by younger men, which is fairly typical for the data set. 

“The refugees that we intended to use the app with were really anything but younger men,” Sjöberg says. “So that meant that the voice data that we needed did not quite match the voice data that we had.” The team started collecting its own voice data from Ukrainian women, as well as from elderly people. 

Unlike other data sets, Common Voice asks participants to share their gender and details about their accent. Making sure different genders are represented is important to fight bias in AI models, says Rebecca Ryakitimbo, a Common Voice fellow who created the project’s gender action plan. More diversity leads not only to better representation but also to better models. Systems that are trained on narrow and homogenous data tend to spew stereotyped and harmful results.

“We don’t want a case where we have a chatbot that is named after a woman but does not give the same response to a woman as it would a man,” she says. 

Ryakitimbo has collected voice data in Kiswahili in Tanzania, Kenya, and the Democratic Republic of Congo. She tells me she wanted to collect voices from a socioeconomically diverse set of Kiswahili speakers and has reached out to women young and old living in rural areas, who might not always be literate or even have access to devices. 

This kind of data collection is challenging. The importance of collecting AI voice data can feel abstract to many people, especially if they aren’t familiar with the technologies. Ryakitimbo and volunteers would approach women in settings where they felt safe to begin with, such as presentations on menstrual hygiene, and explain how the technology could, for example, help disseminate information about menstruation. For women who did not know how to read, the team read out sentences that they would repeat for the recording. 

The Common Voice project is bolstered by the belief that languages form a really important part of identity. “We think it’s not just about language, but about transmitting culture and heritage and treasuring people’s particular cultural context,” says Lewis-Jong. “There are all kinds of idioms and cultural catchphrases that just don’t translate,” they add. 

Common Voice is the only audio data set where English doesn’t dominate, says Willie Agnew, a researcher at Carnegie Mellon University who has studied audio data sets. “I’m very impressed with how well they’ve done that and how well they’ve made this data set that is actually pretty diverse,” Agnew says. “It feels like they’re way far ahead of almost all the other projects we looked at.” 

I spent some time verifying the recordings of other Finnish speakers on the Common Voice platform. As their voices echoed in my study, I felt surprisingly touched. We had all gathered around the same cause: making AI data more inclusive, and making sure our culture and language was properly represented in the next generation of AI tools. 

But I had some big questions about what would happen to my voice if I donated it. Once it was in the data set, I would have no control about how it might be used afterwards. The tech sector isn’t exactly known for giving people proper credit, and the data is available for anyone’s use. 

“As much as we want it to benefit the local communities, there’s a possibility that also Big Tech could make use of the same data and build something that then comes out as the commercial product,” says Ryakitimbo. Though Mozilla does not share who has downloaded Common Voice, Lewis-Jong tells me Meta and Nvidia have said that they have used it.

Open access to this hard-won and rare language data is not something all minority groups want, says Harry H. Jiang, a researcher at Carnegie Mellon University, who was part of the team doing audit research. For example, Indigenous groups have raised concerns. 

“Extractivism” is something that Mozilla has been thinking about a lot over the past 18 months, says Lewis-Jong. Later this year the project will work with communities to pilot alternative licenses including Nwulite Obodo Open Data License, which was created by researchers at the University of Pretoria for sharing African data sets more equitably. For example, people who want to download the data might be asked to write a request with details on how they plan to use it, and they might be allowed to license it only for certain products or for a limited time. Users might also be asked to contribute to community projects that support poverty reduction, says Lewis-Jong.  

Lewis-Jong says the pilot is a learning exercise to explore whether people will want data with alternative licenses, and whether they are sustainable for communities managing them. The hope is that it could lead to something resembling “open source 2.0.”

In the end, I decided to donate my voice. I received a list of phrases to say, sat in front of my computer, and hit Record. One day, I hope, my effort will help a company or researcher build voice AI that sounds less generic, and more like me. 

This story has been updated.