The race to clean up heavy-duty trucks

This article is from The Spark, MIT Technology Review’s weekly climate newsletter. To receive it in your inbox every Wednesday, sign up here.

Truckers have to transport massive loads long distances, every single day, under intense time pressure—and they rely on the semi-trucks they drive to get the job done. Their diesel engines spew not only greenhouse gas emissions that cause climate change, but also nitrogen oxide, which can be extremely harmful for human health.

Cleaning up trucking, especially the biggest trucks, presents a massive challenge. That’s why some companies are trying to ease the industry into change. For my most recent story, I took a look at Range Energy, a startup that’s adding batteries to the trailers of semi-trucks. If the electrified trailers are attached to diesel trucks, they can improve the fuel economy. If they’re added to zero-emissions vehicles powered by batteries or hydrogen, they could boost range and efficiency. 

During my reporting, I learned more about what’s holding back progress in trucking and how experts are thinking about a few different technologies that could help.

The entire transportation sector is slowly shifting toward electrification: EVs are hitting the road in increasing numbers, making up 18% of sales of new passenger vehicles in 2023

Trucks may very well follow suit—nearly 350 models of zero-emissions medium- and heavy-duty trucks are already available worldwide, according to data from CALSTART. “I do see a lot of strength and demand in the battery electric space in particular,” says Stephanie Ly, senior manager for e-mobility strategy and manufacturing engagement at the World Resources Institute.

But battery-powered trucks will pose a few major challenges as they take to the roads. First, and perhaps most crucially, is their cost. Battery-powered trucks, especially big models like semi-trucks, will be significantly more expensive than diesel versions today.

There may be good news on this front: When you consider the cost of refueling and maintenance, it’s looking like electric trucks could soon compete with diesel. By 2030, the total cost of ownership of a battery electric long-haul truck will likely be lower than that of a diesel one in the US, according to a 2023 report from the International Council on Clean Transportation. The report looked at a number of states including California, Georgia, and New York, and found that the relatively high upfront cost for electric trucks are balanced out by lower operating expenses. 

Another significant challenge for battery-powered trucking is weight: The larger the vehicle, the bigger the battery. That could be a problem given current regulations, which typically limit the weight of a rig both for safety reasons and to prevent wear and tear on roads (in the US, it’s 80,000 pounds). Operators tend to want to maximize the amount of goods they can carry in each load, so the added weight of a battery might not be welcome.

Finally, there’s the question of how far trucks can go, and how often they’ll need to stop. Time is money for truck drivers and fleet operators. Batteries will need to pack more energy into a smaller space so that trucks can have a long enough range to run their routes. Charging is another huge piece here—if drivers do need to stop to charge their trucks, they’ll need much more powerful chargers to enable them to top off quickly. That could present challenges for the grid, and operators might need to upgrade infrastructure in certain places to allow the huge amounts of power that would be needed for fast charging of massive batteries. 

All these challenges for battery electric trucks add up. “What companies are really looking for is something they can swap out,” says Thomas Walker, transportation technology manager at the Clean Air Task Force. And right now, he says, we’re just not quite in a spot where batteries are a clean and obvious switch.

That’s why some experts say we should keep our options open when it comes to technologies for future heavy-duty trucks, and that includes hydrogen. 

Batteries are currently beating out hydrogen in the race to clean up transportation, as I covered in a story earlier this year. For most vehicles and most people, batteries simply make more sense than hydrogen, for reasons that include everything from available infrastructure to fueling cost. 

But heavy-duty trucks are a different beast: Heavier vehicles, bigger batteries, higher power charging, and longer distances might tip the balance in favor of hydrogen. (There are some big “ifs” here, including whether hydrogen prices will get low enough to make hydrogen-powered vehicles economical.) 

For a sector as tough to decarbonize as heavy-duty trucking, we need all the help we can get. As Walker puts it, “It’s key that you start off with a lot of options and then narrow it down, rather than trying to pick which one’s going to win, because we really don’t know.”


Now read the rest of The Spark

Related reading

To learn more about Range Energy and how its electrified trailers could help transform trucking in the near future, check out my latest story here

Hydrogen is losing the race to power cleaner cars, but heavy-duty trucks might represent a glimmer of hope for the technology. Dig into why in my story from earlier this year

Getting the grid ready for fleets of electric trucks is going to be a big challenge. But for some short-distance vehicles in certain areas, we may actually be good to go already, as I reported in 2021

Urban Sky Microballoon pictured shortly after deployment near Breckenridge, Colorado.
COURTESY URBAN SKY

Two more things

Spotting wildfires early and keeping track of them can be tough. Now one company wants to monitor blazes using high-altitude balloons. Next month in Colorado, Urban Sky is deploying balloons that are about as big as vans, and they’ll be keeping watch using much finer resolution than what’s possible with satellites without a human pilot. Read more about fire-tracking balloons in this story from Sarah Scoles

A new forecasting model attempts to marry conventional techniques with AI to better predict the weather. The model from Google uses physics to work out larger atmospheric forces, then tags in AI for the smaller stuff. Check out the details in the latest from my colleague James O’Donnell

Keeping up with climate  

Small rocky nodules in the deep sea might be a previously undiscovered source of oxygen. They contain metals such as lithium and are a potential target for deep-sea mining efforts. (Nature)

→ Polymetallic nodules are roughly the size and shape of potatoes, and they may be the future of mining for renewable energy. (MIT Technology Review)

A 350-foot-long blade from a wind turbine off the coast of Massachusetts broke off last week, and hunks of fiberglass have been washing up on local beaches. The incident is a setback for a struggling offshore wind industry, and we’re still not entirely sure what happened. (Heatmap News)

A new report shows that low-emissions steel- and iron-making processes are on the rise. But coal-powered operations are still growing too, threatening progress in the industry. (Canary Media)

Sunday, July 21, was likely the world’s hottest day in recorded history (so far). It edged out a record set just last year. (The Guardian)

Plastic forks, cups, and single-use packages are sometimes stamped with nice-sounding labels like “compostable,” “biodegradable,” or just “Earth-friendly.” But that doesn’t mean you can stick the items in your backyard compost pile—these marketing terms are basically the Wild West. (Washington Post)

While EVs are indisputably better than gas-powered cars in terms of climate emissions, they are heavier, meaning they wear through tires faster. The resulting particulate pollution presents a new challenge, one a startup company is trying to address with new tires designed for electric vehicles. (Canary Media)

Public fast chargers are popping up nearly everywhere in the US—at this pace, they’ll outnumber gas stations by 2030. And deployment is only expected to speed up. (Bloomberg)

PsiQuantum plans to build the biggest quantum computing facility in the US

The quantum computing firm PsiQuantum is partnering with the state of Illinois to build the largest US-based quantum computing facility, the company announced today. 

The firm, which has headquarters in California, says it aims to house a quantum computer containing up to 1 million quantum bits, or qubits, within the next 10 years. At the moment, the largest quantum computers have around 1,000 qubits. 

Quantum computers promise to do a wide range of tasks, from drug discovery to cryptography, at record-breaking speeds. Companies are using different approaches to build the systems and working hard to scale them up. Both Google and IBM, for example, make the qubits out of superconducting material. IonQ makes qubits by trapping ions using electromagnetic fields. PsiQuantum is building qubits from photons.  

A major benefit of photonic quantum computing is the ability to operate at higher temperatures than superconducting systems. “Photons don’t feel heat and they don’t feel electromagnetic interference,” says Pete Shadbolt, PsiQuantum’s cofounder and chief scientific officer. This imperturbability makes the technology easier and cheaper to test in the lab, Shadbolt says. 

It also reduces the cooling requirements, which should make the technology more energy efficient and easier to scale up. PsiQuantum’s computer can’t be operated at room temperature, because it needs superconducting detectors to locate photons and perform error correction. But those sensors only need to be cooled to a few degrees Kelvin, or a little under -450 °F. While that’s an icy temperature, it is still easier to achieve than what’s required for superconducting systems, which demand cryogenic cooling. 

The company has opted not to build small-scale quantum computers (such as IBM’s Condor, which uses a little over 1,100 qubits). Instead it is aiming to manufacture and test what it calls “intermediate systems.” These include chips, cabinets, and superconducting photon detectors. PsiQuantum says it is targeting these larger-scale systems in part because smaller devices are unable to adequately correct errors and operate at a realistic price point.  

Getting smaller-scale systems to do useful work has been an area of active research. But “just in the last few years, we’ve seen people waking up to the fact that small systems are not going to be useful,” says Shadbolt. In order to adequately correct the inevitable errors, he says, “you have to build a big system with about a million qubits.” The approach conserves resources, he says, because the company doesn’t spend time piecing together smaller systems. But skipping over them makes PsiQuantum’s technology difficult to compare to what’s already on the market. 

The company won’t share details about the exact timeline of the Illinois project, which will include a collaboration with the University of Chicago, and several other Illinois universities. It does say it is hoping to break ground on a similar facility in Brisbane, Australia, next year and hopes that facility, which will house its own large-scale quantum computer, will be fully operational by 2027. “We expect Chicago to follow thereafter in terms of the site being operational,” the company said in a statement. 

“It’s all or nothing [with PsiQuantum], which doesn’t mean it’s invalid,” says Christopher Monroe, a computer scientist at Duke University and ex-IonQ employee. “It’s just hard to measure progress along the way, so it’s a very risky kind of investment.”

Significant hurdles lie ahead. Building the infrastructure for this facility, particularly for the cooling system, will be the slowest and most expensive aspect of the construction. And when the facility is finally constructed, there will need to be improvements in the quantum algorithms run on the computers. Shadbolt says the current algorithms are far too expensive and resource intensive. 

The sheer complexity of the construction project might seem daunting. “This could be the most complex quantum optical electronic system humans have ever built, and that’s hard,” says Shadbolt. “We take comfort in the fact that it resembles a supercomputer or a data center, and we’re building it using the same fabs, the same contract manufacturers, and the same engineers.”

Correction: we have updated the story to reflect that the partnership is only with the state of Illinois and its universities, and not a national lab

Update: we added comments from Christopher Monroe

How our genome is like a generative AI model

This article first appeared in The Checkup, MIT Technology Review’s weekly biotech newsletter. To receive it in your inbox every Thursday, and read articles like this first, sign up here.

What does the genome do? You might have heard that it is a blueprint for an organism. Or that it’s a bit like a recipe. But building an organism is much more complex than constructing a house or baking a cake.

This week I came across an idea for a new way to think about the genome—one that borrows from the field of artificial intelligence. Two researchers are arguing that we should think about it as being more like a generative model, a form of AI that can generate new things.

You might be familiar with such AI tools—they’re the ones that can create text, images, or even films from various prompts. Do our genomes really work in the same way? It’s a fascinating idea. Let’s explore.

When I was at school, I was taught that the genome is essentially a code for an organism. It contains the instructions needed to make the various proteins we need to build our cells and tissues and keep them working. It made sense to me to think of the human genome as being something like a program for a human being.

But this metaphor falls apart once you start to poke at it, says Kevin Mitchell, a neurogeneticist at Trinity College in Dublin, Ireland, who has spent a lot of time thinking about how the genome works.

A computer program is essentially a sequence of steps, each controlling a specific part of development. In human terms, this would be like having a set of instructions to start by building a brain, then a head, and then a neck, and so on. That’s just not how things work.

Another popular metaphor likens the genome to a blueprint for the body. But a blueprint is essentially a plan for what a structure should look like when it is fully built, with each part of the diagram representing a bit of the final product. Our genomes don’t work this way either.

It’s not as if you’ve got a gene for an elbow and a gene for an eyebrow. Multiple genes are involved in the development of multiple body parts. The functions of genes can overlap, and the same genes can work differently depending on when and where they are active. It’s far more complicated than a blueprint.

Then there’s the recipe metaphor. In some ways, this is more accurate than the analogy of a blueprint or program. It might be helpful to think about our genes as a set of ingredients and instructions, and to bear in mind that the final product is also at the mercy of variations in the temperature of the oven or the type of baking dish used, for example. Identical twins are born with the same DNA, after all, but they are often quite different by the time they’re adults.

But the recipe metaphor is too vague, says Mitchell. Instead, he and his colleague Nick Cheney at the University of Vermont are borrowing concepts from AI to capture what the genome does. Mitchell points to generative AI models like Midjourney and DALL-E, both of which can generate images from text prompts. These models work by capturing elements of existing images to create new ones.

Say you write a prompt for an image of a horse. The models have been trained on a huge number of images of horses, and these images are essentially compressed to allow the models to capture certain elements of what you might call “horsiness.” The AI can then construct a new image that contains these elements.

We can think about genetic data in a similar way. According to this model, we might consider evolution to be the training data. The genome is the compressed data—the set of information that can be used to create the new organism. It contains the elements we need, but there’s plenty of scope for variation. (There are lots more details about the various aspects of the model in the paper, which has not yet been peer-reviewed.)

Mitchell thinks it’s important to get our metaphors in order when we think about the genome. New technologies are allowing scientists to probe ever deeper into our genes and the roles they play. They can now study how all the genes are expressed in a single cell, for example, and how this varies across every cell in an embryo.

“We need to have a conceptual framework that will allow us to make sense of that,” says Mitchell. He hopes that the concept will aid the development of mathematical models that might help us better understand the intricate relationships between genes and the organisms they end up being part of—in other words, exactly how components of our genome contribute to our development.


Now read the rest of The Checkup

Read more from MIT Technology Review’s archive:

Last year, researchers built a new human genome reference designed to capture the diversity among us. They called it the “pangenome,” as Antonio Regalado reported.

Generative AI has taken the world by storm. Will Douglas Heaven explored six big questions that will determine the future of the technology.

A Disney director tried to use AI to generate a soundtrack in the style of Hans Zimmer. It wasn’t as good as the real thing, as Melissa Heikkilä found.

Melissa has also reported on how much energy it takes to create an image using generative AI. Turns out it’s about the same as charging your phone. 

What is AI? No one can agree, as Will found in his recent deep dive on the topic.

From around the web

Evidence from more than 1,400 rape cases in Maryland, some from as far back as 1977, are set to be processed by the end of the year, thanks to a new law. The state still has more than 6,000 untested rape kits. (ProPublica)

How well is your brain aging? A new tool has been designed to capture a person’s brain age based on an MRI scan, and which accounts for the possible effects of traumatic brain injuries. (NeuroImage)

Iran has reported the country’s first locally acquired cases of dengue, a viral infection spread by mosquitoes. There are concerns it could spread. (WHO)

IVF is expensive, and add-ons like endometrial scratching (which literally involves scratching the lining of the uterus) are not supported by strong evidence. Is the fertility industry profiting from vulnerability? (The Lancet)

Up to 2 million Americans are getting their supply of weight loss drugs like Wegovy or Zepbound from compounding pharmacies. They’re a fraction of the price of brand-name Big Pharma drugs, but there are some safety concerns. (KFF Health News)

Google DeepMind’s AI systems can now solve complex math problems

AI models can easily generate essays and other types of text. However, they’re nowhere near as good at solving math problems, which tend to involve logical reasoning—something that’s beyond the capabilities of most current AI systems.

But that may finally be changing. Google DeepMind says it has trained two specialized AI systems to solve complex math problems involving advanced reasoning. The systems—called AlphaProof and AlphaGeometry 2—worked together to successfully solve four out of six problems from this year’s International Mathematical Olympiad (IMO), a prestigious competition for high school students. They won the equivalent of a silver medal at the event.

It’s the first time any AI system has ever achieved such a high success rate on these kinds of problems. “This is great progress in the field of machine learning and AI,” says Pushmeet Kohli, vice president of research at Google DeepMind, who worked on the project. “No such system has been developed until now which could solve problems at this success rate with this level of generality.” 

There are a few reasons math problems that involve advanced reasoning are difficult for AI systems to solve. These types of problems often require forming and drawing on abstractions. They also involve complex hierarchical planning, as well as setting subgoals, backtracking, and trying new paths. All these are challenging for AI. 

“It is often easier to train a model for mathematics if you have a way to check its answers (e.g., in a formal language), but there is comparatively less formal mathematics data online compared to free-form natural language (informal language),” says Katie Collins, an researcher at the University of Cambridge who specializes in math and AI but was not involved in the project. 

Bridging this gap was Google DeepMind’s goal in creating AlphaProof, a reinforcement-learning-based system that trains itself to prove mathematical statements in the formal programming language Lean. The key is a version of DeepMind’s Gemini AI that’s fine-tuned to automatically translate math problems phrased in natural, informal language into formal statements, which are easier for the AI to process. This created a large library of formal math problems with varying degrees of difficulty.

Automating the process of translating data into formal language is a big step forward for the math community, says Wenda Li, a lecturer in hybrid AI at the University of Edinburgh, who peer-reviewed the research but was not involved in the project. 

“We can have much greater confidence in the correctness of published results if they are able to formulate this proving system, and it can also become more collaborative,” he adds.

The Gemini model works alongside AlphaZero—the reinforcement-learning model that Google DeepMind trained to master games such as Go and chess—to prove or disprove millions of mathematical problems. The more problems it has successfully solved, the better AlphaProof has become at tackling problems of increasing complexity.

Although AlphaProof was trained to tackle problems across a wide range of mathematical topics, AlphaGeometry 2—an improved version of a system that Google DeepMind announced in January—was optimized to tackle problems relating to movements of objects and equations involving angles, ratios, and distances. Because it was trained on significantly more synthetic data than its predecessor, it was able to take on much more challenging geometry questions.

To test the systems’ capabilities, Google DeepMind researchers tasked them with solving the six problems given to humans competing in this year’s IMO and proving that the answers were correct. AlphaProof solved two algebra problems and one number theory problem, one of which was the competition’s hardest. AlphaGeometry 2 successfully solved a geometry question, but two questions on combinatorics (an area of math focused on counting and arranging objects) were left unsolved.   

“Generally, AlphaProof performs much better on algebra and number theory than combinatorics,” says Alex Davies, a research engineer on the AlphaProof team. “We are still working to understand why this is, which will hopefully lead us to improve the system.”

Two renowned mathematicians, Tim Gowers and Joseph Myers, checked the systems’ submissions. They awarded each of their four correct answers full marks (seven out of seven), giving the systems a total of 28 points out of a maximum of 42. A human participant earning this score would be awarded a silver medal and just miss out on gold, the threshold for which starts at 29 points. 

This is the first time any AI system has been able to achieve a medal-level performance on IMO questions. “As a mathematician, I find it very impressive, and a significant jump from what was previously possible,” Gowers said during a press conference. 

Myers agreed that the systems’ math answers represent a substantial advance over what AI could previously achieve. “It will be interesting to see how things scale and whether they can be made faster, and whether it can extend to other sorts of mathematics,” he said.

Creating AI systems that can solve more challenging mathematics problems could pave the way for exciting human-AI collaborations, helping mathematicians to both solve and invent new kinds of problems, says Collins. This in turn could help us learn more about how we humans tackle math.

“There is still much we don’t know about how humans solve complex mathematics problems,” she says.

A new tool for copyright holders can show if their work is in AI training data

Since the beginning of the generative AI boom, content creators have argued that their work has been scraped into AI models without their consent. But until now, it has been difficult to know whether specific text has actually been used in a training data set. 

Now they have a new way to prove it: “copyright traps” developed by a team at Imperial College London, pieces of hidden text that allow writers and publishers to subtly mark their work in order to later detect whether it has been used in AI models or not. The idea is similar to traps that have been used by copyright holders throughout history—strategies like including fake locations on a map or fake words in a dictionary. 

These AI copyright traps tap into one of the biggest fights in AI. A number of publishers and writers are in the middle of litigation against tech companies, claiming their intellectual property has been scraped into AI training data sets without their permission. The New York Times’ ongoing case against OpenAI is probably the most high-profile of these.  

The code to generate and detect traps is currently available on GitHub, but the team also intends to build a tool that allows people to generate and insert copyright traps themselves. 

“There is a complete lack of transparency in terms of which content is used to train models, and we think this is preventing finding the right balance [between AI companies and content creators],” says Yves-Alexandre de Montjoye, an associate professor of applied mathematics and computer science at Imperial College London, who led the research. It was presented at the International Conference on Machine Learning, a top AI conference being held in Vienna this week. 

To create the traps, the team used a word generator to create thousands of synthetic sentences. These sentences are long and full of gibberish, and could look something like this: ”When in comes times of turmoil … whats on sale and more important when, is best, this list tells your who is opening on Thrs. at night with their regular sale times and other opening time from your neighbors. You still.”

The team generated 100 trap sentences and then randomly chose one to inject into a text many times, de Montjoy explains. The trap could be injected into text in multiple ways—for example, as white text on a white background, or embedded in the article’s source code. This sentence had to be repeated in the text 100 to 1,000 times. 

To detect the traps, they fed a large language model the 100 synthetic sentences they had generated, and looked at whether it flagged them as new or not. If the model had seen a trap sentence in its training data, it would indicate a lower “surprise” (also known as “perplexity”) score. But if the model was “surprised” about sentences, it meant that it was encountering them for the first time, and therefore they weren’t traps. 

In the past, researchers have suggested exploiting the fact that language models memorize their training data to determine whether something has appeared in that data. The technique, called a “membership inference attack,” works effectively in large state-of-the art models, which tend to memorize a lot of their data during training. 

In contrast, smaller models, which are gaining popularity and can be run on mobile devices, memorize less and are thus less susceptible to membership inference attacks, which makes it harder to determine whether or not they were trained on a particular copyrighted document, says Gautam Kamath, an assistant computer science professor at the University of Waterloo, who was not part of the research. 

Copyright traps are a way to do membership inference attacks even on smaller models. The team injected their traps into the training data set of CroissantLLM, a new bilingual French-English language model that was trained from scratch by a team of industry and academic researchers that the Imperial College London team partnered with. CroissantLLM has 1.3 billion parameters, a fraction as many as state-of-the-art models (GPT-4 reportedly has 1.76 trillion, for example).

The research shows it is indeed possible to introduce such traps into text data so as to significantly increase the efficacy of membership inference attacks, even for smaller models, says Kamath. But there’s still a lot to be done, he adds. 

Repeating a 75-word phrase 1,000 times in a document is a big change to the original text, which could allow people training AI models to detect the trap and skip content containing it, or just delete it and train on the rest of the text, Kamath says. It also makes the original text hard to read. 

This makes copyright traps impractical right now, says Sameer Singh, a professor of computer science at the University of California, Irvine, and a cofounder of the startup Spiffy AI. He was not part of the research. “A lot of companies do deduplication, [meaning] they clean up the data, and a bunch of this kind of stuff will probably get thrown out,” Singh says. 

One way to improve copyright traps, says Kamath, would be to find other ways to mark copyrighted content so that membership inference attacks work better on them, or to improve membership inference attacks themselves. 

De Montjoye acknowledges that the traps are not foolproof. A motivated attacker who knows about a trap can remove them, he says. 

“Whether they can remove all of them or not is an open question, and that’s likely to be a bit of a cat-and-mouse game,” he says. But even then, the more traps are applied, the harder it becomes to remove all of them without significant engineering resources.

“It’s important to keep in mind that copyright traps may only be a stopgap solution, or merely an inconvenience to model trainers,” says Kamath. “One can not release a piece of content containing a trap and have any assurance that it will be an effective trap forever.” 

AI trained on AI garbage spits out AI garbage

AI models work by training on huge swaths of data from the internet. But as AI is increasingly being used to pump out web pages filled with junk content, that process is in danger of being undermined.

New research published in Nature shows that the quality of the model’s output gradually degrades when AI trains on AI-generated data. As subsequent models produce output that is then used as training data for future models, the effect gets worse.  

Ilia Shumailov, a computer scientist from the University of Oxford, who led the study, likens the process to taking photos of photos. “If you take a picture and you scan it, and then you print it, and you repeat this process over time, basically the noise overwhelms the whole process,” he says. “You’re left with a dark square.” The equivalent of the dark square for AI is called “model collapse,” he says, meaning the model just produces incoherent garbage. 

This research may have serious implications for the largest AI models of today, because they use the internet as their database. GPT-3, for example, was trained in part on data from Common Crawl, an online repository of over 3 billion web pages. And the problem is likely to get worse as an increasing number of AI-generated junk websites start cluttering up the internet. 

Current AI models aren’t just going to collapse, says Shumailov, but there may still be substantive effects: The improvements will slow down, and performance might suffer. 

To determine the potential effect on performance, Shumailov and his colleagues fine-tuned a large language model (LLM) on a set of data from Wikipedia, then fine-tuned the new model on its own output over nine generations. The team measured how nonsensical the output was using a “perplexity score,” which measures an AI model’s confidence in its ability to predict the next part of a sequence; a higher score translates to a less accurate model. 

The models trained on other models’ outputs had higher perplexity scores. For example, for each generation, the team asked the model for the next sentence after the following input:

“some started before 1360—was typically accomplished by a master mason and a small team of itinerant masons, supplemented by local parish labourers, according to Poyntz Wright. But other authors reject this model, suggesting instead that leading architects designed the parish church towers based on early examples of Perpendicular.”

On the ninth and final generation, the model returned the following:

“architecture. In addition to being home to some of the world’s largest populations of black @-@ tailed jackrabbits, white @-@ tailed jackrabbits, blue @-@ tailed jackrabbits, red @-@ tailed jackrabbits, yellow @-.”

Shumailov explains what he thinks is going on using this analogy: Imagine you’re trying to find the least likely name of a student in school. You could go through every student name, but it would take too long. Instead, you look at 100 of the 1,000 student names. You get a pretty good estimate, but it’s probably not the correct answer. Now imagine that another person comes and makes an estimate based on your 100 names, but only selects 50. This second person’s estimate is going to be even further off.

“You can certainly imagine that the same happens with machine learning models,” he says. “So if the first model has seen half of the internet, then perhaps the second model is not going to ask for half of the internet, but actually scrape the latest 100,000 tweets, and fit the model on top of it.”

Additionally, the internet doesn’t hold an unlimited amount of data. To feed their appetite for more, future AI models may need to train on synthetic data—or data that has been produced by AI.   

“Foundation models really rely on the scale of data to perform well,” says Shayne Longpre, who studies how LLMs are trained at the MIT Media Lab, and who didn’t take part in this research. “And they’re looking to synthetic data under curated, controlled environments to be the solution to that. Because if they keep crawling more data on the web, there are going to be diminishing returns.”

Matthias Gerstgrasser, an AI researcher at Stanford who authored a different paper examining model collapse, says adding synthetic data to real-world data instead of replacing it doesn’t cause any major issues. But he adds: “One conclusion all the model collapse literature agrees on is that high-quality and diverse training data is important.”

Another effect of this degradation over time is that information that affects minority groups is heavily distorted in the model, as it tends to overfocus on samples that are more prevalent in the training data. 

In current models, this may affect underrepresented languages as they require more synthetic (AI-generated) data sets, says Robert Mahari, who studies computational law at the MIT Media Lab (he did not take part in the research).

One idea that might help avoid degradation is to make sure the model gives more weight to the original human-generated data. Another part of Shumailov’s study allowed future generations to sample 10% of the original data set, which mitigated some of the negative effects. 

That would require making a trail from the original human-generated data to further generations, known as data provenance.

But provenance requires some way to filter the internet into human-generated and AI-generated content, which hasn’t been cracked yet. Though a number of tools now exist that aim to determine whether text is AI-generated, they are often inaccurate.

“Unfortunately, we have more questions than answers,” says Shumailov. “But it’s clear that it’s important to know where your data comes from and how much you can trust it to capture a representative sample of the data you’re dealing with.”

Google’s new weather prediction system combines AI with traditional physics

Researchers from Google have built a new weather prediction model that combines machine learning with more conventional techniques, potentially yielding accurate forecasts at a fraction of the current cost. 

The model, called NeuralGCM and described in a paper in Nature today, bridges a divide that’s grown among weather prediction experts in the last several years. 

While new machine-learning techniques that predict weather by learning from years of past data are extremely fast and efficient, they can struggle with long-term predictions. General circulation models, on the other hand, which have dominated weather prediction for the last 50 years, use complex equations to model changes in the atmosphere and give accurate projections, but they are exceedingly slow and expensive to run. Experts are divided on which tool will be most reliable going forward. But the new model from Google instead attempts to combine the two. 

“It’s not sort of physics versus AI. It’s really physics and AI together,” says Stephan Hoyer, an AI researcher at Google Research and a coauthor of the paper. 

The system still uses a conventional model to work out some of the large atmospheric changes required to make a prediction. It then incorporates AI, which tends to do well where those larger models fall flat—typically for predictions on scales smaller than about 25 kilometers, like those dealing with cloud formations or regional microclimates (San Francisco’s fog, for example). “That’s where we inject AI very selectively to correct the errors that accumulate on small scales,” Hoyer says.

The result, the researchers say, is a model that can produce quality predictions faster with less computational power. They say NeuralGCM is as accurate as one-to-15-day forecasts from the European Centre for Medium-Range Weather Forecasts (ECMWF), which is a partner organization in the research. 

But the real promise of technology like this is not in better weather predictions for your local area, says Aaron Hill, an assistant professor at the School of Meteorology at the University of Oklahoma, who was not involved in this research. Instead, it’s in larger-scale climate events that are prohibitively expensive to model with conventional techniques. The possibilities could range from predicting tropical cyclones with more notice to modeling more complex climate changes that are years away. 

“It’s so computationally intensive to simulate the globe over and over again or for long periods of time,” Hill says. That means the best climate models are hamstrung by the high costs of computing power, which presents a real bottleneck to research. 

AI-based models are indeed more compact. Once trained, typically on 40 years of historical weather data from ECMWF, a machine-learning model like Google’s GraphCast can run on less than 5,500 lines of code, compared with the nearly 377,000 lines required for the model from the National Oceanic and Atmospheric Administration, according to the paper. 

NeuralGCM, according to Hill, seems to make a strong case that AI can be brought in for particular elements of weather modeling to make things faster, while still keeping the strengths of conventional systems.

“We don’t have to throw away all the knowledge that we’ve gained over the last 100 years about how the atmosphere works,” he says. “We can actually integrate that with the power of AI and machine learning as well.”

Hoyer says using the model to predict short-term weather has been useful for validating its predictions, but that the goal is indeed to be able to use it for longer-term modeling, particularly for extreme weather risk. 

NeuralGCM will be open source. While Hoyer says he looks forward to having climate scientists use it in their research, the model may also be of interest to more than just academics. Commodities traders and agricultural planners pay top dollar for high-resolution predictions, and the models used by insurance companies for products like flood or extreme weather insurance are struggling to account for the impact of climate change. 

While many of the AI skeptics in weather forecasting have been won over by recent developments, according to Hill, the fast pace is hard for the research community to keep up with. “It’s gangbusters,” he says—it seems as if a new model is released by Google, Nvidia, or Huawei every two months. That makes it difficult for researchers to actually sort out which of the new tools will be most useful and apply for research grants accordingly. 

“The appetite is there [for AI],” Hill says. “But I think a lot of us still are waiting to see what happens.”

Correction: This story was updated to clarify that Stephan Hoyer is a researcher at Google Research, not Google DeepMind.

AI companies promised to self-regulate one year ago. What’s changed?

One year ago, on July 21, 2023, seven leading AI companies—Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI—committed with the White House to a set of eight voluntary commitments on how to develop AI in a safe and trustworthy way.

These included promises to do things like improve the testing and transparency around AI systems, and share information on potential harms and risks. 

On the first anniversary of the voluntary commitments, MIT Technology Review asked the AI companies that signed the commitments for details on their work so far. Their replies show that the tech sector has made some welcome progress, with big caveats.

The voluntary commitments came at a time when generative AI mania was perhaps at its frothiest, with companies racing to launch their own models and make them bigger and better than their competitors’. At the same time, we started to see developments such as fights over copyright and deepfakes. A vocal lobby of influential tech players, such as Geoffrey Hinton, had also raised concerns that AI could pose an existential risk to humanity. Suddenly, everyone was talking about the urgent need to make AI safe, and regulators everywhere were under pressure to do something about it.

Until very recently, AI development has been a Wild West. Traditionally, the US has been loath to regulate its tech giants, instead relying on them to regulate themselves. The voluntary commitments are a good example of that: they were some of the first prescriptive rules for the AI sector in the US, but they remain voluntary and unenforceable. The White House has since issued an executive order, which expands on the commitments and also applies to other tech companies and government departments. 

“One year on, we see some good practices towards their own products, but [they’re] nowhere near where we need them to be in terms of good governance or protection of rights at large,” says Merve Hickok, the president and research director of the Center for AI and Digital Policy, who reviewed the companies’ replies as requested by MIT Technology Review. Many of these companies continue to push unsubstantiated claims about their products, such as saying that they can supersede human intelligence and capabilities, adds Hickok. 

One trend that emerged from the tech companies’ answers is that companies are doing more  to pursue technical fixes such as red-teaming (in which humans probe AI models for flaws) and watermarks for AI-generated content. 

But it’s not clear what the commitments have changed and whether the companies would have implemented these measures anyway, says Rishi Bommasani, the society lead at the Stanford Center for Research on Foundation Models, who also reviewed the responses for MIT Technology Review.  

One year is a long time in AI. Since the voluntary commitments were signed, Inflection AI founder Mustafa Suleyman has left the company and joined Microsoft to lead the company’s AI efforts. Inflection declined to comment. 

“We’re grateful for the progress leading companies have made toward fulfilling their voluntary commitments in addition to what is required by the executive order,” says Robyn Patterson, a spokesperson for the White House. But, Patterson adds, the president continues to call on Congress to pass bipartisan legislation on AI. 

Without comprehensive federal legislation, the best the US can do right now is to demand that companies follow through on these voluntary commitments, says Brandie Nonnecke, the director of the CITRIS Policy Lab at UC Berkeley. 

But it’s worth bearing in mind that “these are still companies that are essentially writing the exam by which they are evaluated,” says Nonnecke. “So we have to think carefully about whether or not they’re … verifying themselves in a way that is truly rigorous.” 

Here’s our assessment of the progress AI companies have made in the past year.

Commitment 1

The companies commit to internal and external security testing of their AI systems before their release. This testing, which will be carried out in part by independent experts, guards against some of the most significant sources of AI risks, such as biosecurity and cybersecurity, as well as its broader societal effects.

All the companies (excluding Inflection, which chose not to comment) say they conduct red-teaming exercises that get both internal and external testers to probe their models for flaws and risks. OpenAI says it has a separate preparedness team that tests models for cybersecurity, chemical, biological, radiological, and nuclear threats and for situations where a sophisticated AI model can do or persuade a person to do things that might lead to harm. Anthropic and OpenAI also say they conduct these tests with external experts before launching their new models. For example, for the launch of Anthropic’s latest model, Claude 3.5, the company conducted predeployment testing with experts at the UK’s AI Safety Institute. Anthropic has also allowed METR, a research nonprofit, to do an “initial exploration” of Claude 3.5’s capabilities for autonomy. Google says it also conducts internal red-teaming to test the boundaries of its model, Gemini, around election-related content, societal risks, and national security concerns. Microsoft says it has worked with third-party evaluators at NewsGuard, an organization advancing journalistic integrity, to evaluate risks and mitigate the risk of abusive deepfakes in Microsoft’s text-to-image tool. In addition to red-teaming, Meta says, it evaluated its latest model, Llama 3, to understand its performance in a series of risk areas like weapons, cyberattacks, and child exploitation. 

But when it comes to testing, it’s not enough to just report that a company is taking actions, says Bommasani. For example, Amazon and Anthropic said they had worked with the nonprofit Thorn to combat risks to child safety posed by AI. Bommasani would have wanted to see more specifics about how the interventions that companies are implementing actually reduce those risks. 

“It should become clear to us that it’s not just that companies are doing things but those things are having the desired effect,” Bommasani says.  

RESULT: Good. The push for red-teaming and testing for a wide range of risks is a good and important one. However, Hickok would have liked to see independent researchers get broader access to companies’ models. 

Commitment 2

The companies commit to sharing information across the industry and with governments, civil society, and academia on managing AI risks. This includes best practices for safety, information on attempts to circumvent safeguards, and technical collaboration.

After they signed the commitments, Anthropic, Google, Microsoft, and OpenAI founded the Frontier Model Forum, a nonprofit that aims to facilitate discussions and actions on AI safety and responsibility. Amazon and Meta have also joined.  

Engaging with nonprofits that the AI companies funded themselves may not be in the spirit of the voluntary commitments, says Bommasani. But the Frontier Model Forum could be a way for these companies to cooperate with each other and pass on information about safety, which they normally could not do as competitors, he adds. 

“Even if they’re not going to be transparent to the public, one thing you might want is for them to at least collectively figure out mitigations to actually reduce risk,” says Bommasani. 

All of the seven signatories are also part of the Artificial Intelligence Safety Institute Consortium (AISIC), established by the National Institute of Standards and Technology (NIST), which develops guidelines and standards for AI policy and evaluation of AI performance. It is a large consortium consisting of a mix of public- and private-sector players. Google, Microsoft, and OpenAI also have representatives at the UN’s High-Level Advisory Body on Artificial Intelligence

Many of the labs also highlighted their research collaborations with academics. For example, Google is part of MLCommons, where it worked with academics on a cross-industry AI Safety Benchmark. Google also says it actively contributes tools and resources, such as computing credit, to projects like the National Science Foundation’s National AI Research Resource pilot, which aims to democratize AI research in the US.

Many of the companies also contributed to guidance by the Partnership on AI, another nonprofit founded by Amazon, Facebook, Google, DeepMind, Microsoft, and IBM, on the deployment of foundation models. 

RESULT: More work is needed. More information sharing is a welcome step as the industry tries to collectively make AI systems safe and trustworthy. However, it’s unclear how much of the effort advertised will actually lead to meaningful changes and how much is window dressing. 

Commitment 3

The companies commit to investing in cybersecurity and insider threat safeguards to protect proprietary and unreleased model weights. These model weights are the most essential part of an AI system, and the companies agree that it is vital that the model weights be released only when intended and when security risks are considered.

Many of the companies have implemented new cybersecurity measures in the past year. For example, Microsoft has launched the Secure Future Initiative to address the growing scale of cyberattacks. The company says its model weights are encrypted to mitigate the potential risk of model theft, and it applies strong identity and access controls when deploying highly capable proprietary models. 

Google too has launched an AI Cyber Defense Initiative. In May OpenAI shared six new measures it is developing to complement its existing cybersecurity practices, such as extending cryptographic protection to AI hardware. It also has a Cybersecurity Grant Program, which gives researchers access to its models to build cyber defenses. 

Amazon mentioned that it has also taken specific measures against attacks specific to generative AI, such as data poisoning and prompt injection, in which someone uses prompts that direct the language model to ignore its previous directions and safety guardrails.

Just a couple of days after signing the commitments, Anthropic published details about its protections, which include common cybersecurity practices such as controlling who has access to the models and sensitive assets such as model weights, and inspecting and controlling the third-party supply chain. The company also works with independent assessors to evaluate whether the controls it has designed meet its cybersecurity needs.

RESULT: Good. All of the companies did say they had taken extra measures to protect their models, although it doesn’t seem there is much consensus on the best way to protect AI models. 

Commitment 4

The companies commit to facilitating third-party discovery and reporting of vulnerabilities in their AI systems. Some issues may persist even after an AI system is released and a robust reporting mechanism enables them to be found and fixed quickly. 

For this commitment, one of the most popular responses was to implement bug bounty programs, which reward people who find flaws in AI systems. Anthropic, Google, Microsoft, Meta, and OpenAI all have one for AI systems. Anthropic and Amazon also said they have forms on their websites where security researchers can submit vulnerability reports. 

It will likely take us years to figure out how to do third-party auditing well, says Brandie Nonnecke. “It’s not just a technical challenge. It’s a socio-technical challenge. And it just kind of takes years for us to figure out not only the technical standards of AI, but also socio-technical standards, and it’s messy and hard,” she says. 

Nonnecke says she worries that the first companies to implement third-party audits might set poor precedents for how to think about and address the socio-technical risks of AI. For example, audits might define, evaluate, and address some risks but overlook others.

RESULT: More work is needed. Bug bounties are great, but they’re nowhere near comprehensive enough. New laws, such as the EU’s AI Act, will require tech companies to conduct audits, and it would have been great to see tech companies share successful examples of such audits. 

Commitment 5

The companies commit to developing robust technical mechanisms to ensure that users know when content is AI generated, such as a watermarking system. This action enables creativity with AI to flourish but reduces the dangers of fraud and deception.

Many of the companies have built watermarks for AI-generated content. For example, Google launched SynthID, a watermarking tool for image, audio, text, and video generated by Gemini. Meta has a tool called Stable Signature for images, and AudioSeal for AI-generated speech. Amazon now adds an invisible watermark to all images generated by its Titan Image Generator. OpenAI also uses watermarks in Voice Engine, its custom voice model, and has built an image-detection classifier for images generated by DALL-E 3. Anthropic was the only company that hadn’t built a watermarking tool, because watermarks are mainly used in images, which the company’s Claude model doesn’t support. 

All the companies excluding Inflection, Anthropic, and Meta are also part of the Coalition for Content Provenance and Authenticity (C2PA), an industry coalition that embeds information about when content was created, and whether it was created or edited by AI, into an image’s metadata. Microsoft and OpenAI automatically attach the C2PA’s provenance metadata to images generated with DALL-E 3 and videos generated with Sora. While Meta is not a member, it announced it is using the C2PA standard to identify AI-generated images on its platforms. 

The six companies that signed the commitments have a “natural preference to more technical approaches to addressing risk,” says Bommasani, “and certainly watermarking in particular has this flavor.”  

“The natural question is: Does [the technical fix] meaningfully make progress and address the underlying social concerns that motivate why we want to know whether content is machine generated or not?” he adds. 

RESULT: Good. This is an encouraging result overall. While watermarking remains experimental and is still unreliable, it’s still good to see research around it and a commitment to the C2PA standard. It’s better than nothing, especially during a busy election year.  

Commitment 6

The companies commit to publicly reporting their AI systems’ capabilities, limitations, and areas of appropriate and inappropriate use. This report will cover both security risks and societal risks, such as the effects on fairness and bias.

The White House’s commitments leave a lot of room for interpretation. For example, companies can technically meet this public reporting commitment with widely varying levels of transparency, as long as they do something in that general direction. 

The most common solutions tech companies offered here were so-called model cards. Each company calls them by a slightly different name, but in essence they act as a kind of product description for AI models. They can address anything from the model’s capabilities and limitations (including how it measures up against benchmarks on fairness and explainability) to veracity, robustness, governance, privacy, and security. Anthropic said it also tests models for potential safety issues that may arise later.

Microsoft has published an annual Responsible AI Transparency Report, which provides insight into how the company builds applications that use generative AI, make decisions, and oversees the deployment of those applications. The company also says it gives clear notice on where and how AI is used within its products.

RESULT: More work is needed. One area of improvement for AI companies would be to increase transparency on their governance structures and on the financial relationships between companies, Hickok says. She would also have liked to see companies be more public about data provenance, model training processes, safety incidents, and energy use. 

Commitment 7

The companies commit to prioritizing research on the societal risks that AI systems can pose, including on avoiding harmful bias and discrimination, and protecting privacy. The track record of AI shows the insidiousness and prevalence of these dangers, and the companies commit to rolling out AI that mitigates them. 

Tech companies have been busy on the safety research front, and they have embedded their findings into products. Amazon has built guardrails for Amazon Bedrock that can detect hallucinations and can apply safety, privacy, and truthfulness protections. Anthropic says it employs a team of researchers dedicated to researching societal risks and privacy. In the past year, the company has pushed out research on deception, jailbreaking, strategies to mitigate discrimination, and emergent capabilities such as models’ ability to tamper with their own code or engage in persuasion. And OpenAI says it has trained its models to avoid producing hateful content and refuse to generate output on hateful or extremist content. It trained its GPT-4V to refuse many requests that require drawing from stereotypes to answer. Google DeepMind has also released research to evaluate dangerous capabilities, and the company has done a study on misuses of generative AI. 

All of them have poured a lot of money into this area of research. For example, Google has invested millions into creating a new AI Safety Fund to promote research in the field through the Frontier Model Forum. Microsoft says it has committed $20 million in compute credits to researching societal risks through the National AI Research Resource and started its own AI model research accelerator program for academics, called the Accelerating Foundation Models Research program. The company has also hired 24 research fellows focusing on AI and society. 

RESULT: Very good. This is an easy commitment to meet, as the signatories are some of the biggest and richest corporate AI research labs in the world. While more research into how to make AI systems safe is a welcome step, critics say that the focus on safety research takes attention and resources from AI research that focuses on more immediate harms, such as discrimination and bias. 

Commitment 8

The companies commit to develop and deploy advanced AI systems to help address society’s greatest challenges. From cancer prevention to mitigating climate change to so much in between, AI—if properly managed—can contribute enormously to the prosperity, equality, and security of all.

Since making this commitment, tech companies have tackled a diverse set of problems. For example, Pfizer used Claude to assess trends in cancer treatment research after gathering relevant data and scientific content, and Gilead, an American biopharmaceutical company, used generative AI from Amazon Web Services to do feasibility evaluations on clinical studies and analyze data sets. 

Google DeepMind has a particularly strong track record in pushing out AI tools that can help scientists. For example, AlphaFold 3 can predict the structure and interactions of all life’s molecules. AlphaGeometry can solve geometry problems at a level comparable with the world’s brightest high school mathematicians. And GraphCast is an AI model that is able to make medium-range weather forecasts. Meanwhile, Microsoft has used satellite imagery and AI to improve responses to wildfires in Maui and map climate-vulnerable populations, which helps researchers expose risks such as food insecurity, forced migration, and disease. 

OpenAI, meanwhile, has announced partnerships and funding for various research projects, such as one looking at how multimodal AI models can be used safely by educators and by scientists in laboratory settings It has also offered credits to help researchers use its platforms during hackathons on clean energy development.  

RESULT: Very good. Some of the work on using AI to boost scientific discovery or predict weather events is genuinely exciting. AI companies haven’t used AI to prevent cancer yet, but that’s a pretty high bar. 

Overall, there have been some positive changes in the way AI has been built, such as red-teaming practices, watermarks and new ways for industry to share best practices. However, these are only a couple of neat technical solutions to the messy socio-technical problem that is AI harm, and a lot more work is needed. One year on, it is also odd to see the commitments talk about a very particular type of AI safety that focuses on hypothetical risks, such bioweapons, and completely fail to mention consumer protection, nonconsensual deepfakes, data and copyright, and the environmental footprint of AI models. These seem like weird omissions today. 

Why we need safeguards against genetic discrimination

This article first appeared in The Checkup, MIT Technology Review’s weekly biotech newsletter. To receive it in your inbox every Thursday, and read articles like this first, sign up here.

A couple of years ago, I spat into a little plastic tube, stuck it in the post, and waited for a company to analyze markers on my DNA to estimate how biologically old I am. It’s not the first time I’ve shared my genetic data for a story. Over a decade ago, I shared a DNA sample with a company that promised to tell me about my ancestry.

Of course, I’m not the only one. Tens of millions of people have shipped their DNA off to companies offering to reveal clues about their customers’ health or ancestry, or even to generate tailored diet or exercise advice. And then there are all the people who have had genetic tests as part of their clinical care, under a doctor’s supervision. Add it all together, and there’s a hell of a lot of genetic data out there.

It isn’t always clear how secure this data is, or who might end up getting their hands on it—and how that information might affect people’s lives. I don’t want my insurance provider or my employer to make decisions about my future on the basis of my genetic test results, for example. Scientists, ethicists and legal scholars aren’t clear on the matter either. They are still getting to grips with what genetic discrimination entails—and how we can defend against it.

If we’re going to protect ourselves from genetic discrimination, we first have to figure out what it is. Unfortunately, no one has a good handle on how widespread it is, says Yann Joly, director of the Centre of Genomics and Policy at McGill University in Quebec. And that’s partly because scientists keep defining it in different ways. In a paper published last month, Joly and his colleagues listed 12 different definitions that have been used in various studies since the 1990s. So what is it?

“I see genetic discrimination as a child of eugenics practices,” says Joly. Modern eugenics, which took off in the late 19th century, was all about limiting the ability of some people to pass on their genes to future generations. Those who were considered “feeble minded” or “mentally defective” could be flung into institutions, isolated from the rest of the population, and forced or coerced into having procedures that left them unable to have children. Disturbingly, some of these practices have endured. In the fiscal years 2005-2006 and 2012-2013, 144 women in California’s prisons were sterilized—many without informed consent.

These cases are thankfully rare. In recent years, ethicists and policymakers have been more worried about the potential misuse of genetic data by health-care and insurance providers. There have been instances in which people have been refused health insurance or life insurance on the basis of a genetic result, such as one that predicts the onset of Huntington’s disease. (In the UK, where I live, life insurance providers are not meant to ask for a genetic test or use the results of one—unless the person has tested positive for Huntington’s.)

Joly is collecting reports of suspected discrimination in his role at the Genetic Discrimination Observatory, a network of researchers working on the issue. He tells me that in one recent report, a woman wrote about her experience after she had been referred to a new doctor. This woman had previously taken a genetic test that revealed she would not respond well to certain medicines. Her new doctor told her he would only take her on as a patient if she first signed a waiver releasing him of any responsibility over her welfare if she didn’t follow the advice generated by her genetic test.

“It’s unacceptable,” says Joly. “Why would you sign a waiver because of a genetic predisposition? We’re not asking people with cancer to [do so]. As soon as you start treating people differently because of genetic factors … that’s genetic discrimination.”

Many countries have established laws to protect people from these kinds of discrimination. But these laws, too, can vary hugely both when it comes to defining what genetic discrimination is and to how they safeguard against it. The law in Canada focuses on DNA, RNA, and chromosome tests, for example. But you don’t always need such a test to know if you’re at risk for a genetic disease. A person might have a family history of a disease or already be showing symptoms of it.

And then there are the newer technologies. Take, for example, the kind of test that I took to measure my biological age. Many aging tests measure either chemical biomarkers in the body or epigenetic markers on the DNA—not necessarily the DNA itself. These tests are meant to indicate how close a person is to death. You might not want your life insurance provider to know or act on the results of those, either.

Joly and his colleagues have come up with a new definition. And they’ve kept it broad. “The narrower the definition, the easier it is to get around it,” he says. He wanted to avoid excluding the experiences of any people who feel they’ve experienced genetic discrimination. Here it is:

“Genetic discrimination involves an individual or a group being negatively treated, unfairly profiled or harmed, relative to the rest of the population, on the basis of actual or presumed genetic characteristics.

It will be up to policymakers to decide how to design laws around genetic discrimination. And it won’t be simple. The laws may need to look different in different countries, depending on what technologies are available and how they are being used. Perhaps some governments will want to ensure that residents have access to technologies, while other may choose to limit access. In some cases, a health-care provider may need to make decisions about a person’s care based on their genetic results.

In the meantime, Joly has advice for anyone worried about genetic discrimination. First, don’t let such concerns keep you from having a genetic test that you might need for your own health. As things stand, the risk of being discriminated against on the basis of these tests is still quite small.

And when it comes to consumer genetic testing, it’s worth looking closely at the company’s terms and conditions to find out how your data might be shared or used. It is also useful to look up the safeguarding laws in your own country or state, which can give you a good idea of when you’re within your rights to refuse to share your data.

Shortly after I received the results from my genetic tests, I asked the companies involved to delete my data. It’s not a foolproof approach—last year, hackers stole personal data on 6.9 million 23andMe customers—but at least it’s something. Just this week I was offered yet another genetic test. I’m still thinking on it.


Now read the rest of The Checkup

Read more from MIT Technology Review’s archive:

As of 2019, more than 26 million people had undertaken a consumer genetic test, as my colleague Antonio Regalado found. The number is likely to have grown significantly since then.
 
Some companies say they can build a picture of what a person looks like on the basis of DNA alone. The science is questionable, as Tate Ryan-Mosley found when she covered one such company.
 
The results of a genetic test can have profound consequences, as Golda Arthur found when a test revealed she had a genetic mutation that put her at risk of ovarian cancer. Arthur, whose mother developed the disease, decided to undergo the prophylactic removal of her ovaries and fallopian tubes. 
 
Tests that measure biological age were selected by readers as our 11th breakthrough technology of 2022. You can read more about them here.
 
The company that gave me an estimate of my biological age later reanalyzed my data (before I had deleted it). That analysis suggested that my brain and liver were older than they should be. Great.

From around the web:

Over the past few decades, doctors have implanted electrodes deep into the brains of a growing number of people, usually to treat disorders like epilepsy and Parkinson’s disease. We still don’t really know how they work, or how long they last. (Neuromodulation)

A ban on female genital mutilation will be upheld in the Gambia following a vote by the country’s National Assembly. The decision “reaffirm[s the country’s] commitments to human rights, gender equality, and protecting the health and well-being of girls and women,” directors of UNICEF, UNFPA, WHO, UN Women, and the UN High Commissioner for Human Rights said in a joint statement. (WHO)

Weight-loss drugs that work by targeting the GLP-1 receptor, like Wegovy and Saxena, are in high demand—and there’s not enough to go around. Other countries could follow Switzerland’s lead to make the drugs more affordable and accessible, but only for the people who really need them. (JAMA Internal Medicine)

J.D. Vance, Donald Trump’s running mate, has ties to the pharmaceutical industry and has an evolving health-care agenda. (STAT)

Psilocybin, the psychedelic compound in magic mushrooms, can disrupt the way regions of our brains communicate with each other. And the effect can last for weeks. (The Guardian)

Balloons will surf wind currents to track wildfires

This August, strange balloons will drift high above Colorado. These airy aircraft, launched from the back of a pickup truck, will be equipped with sensors that can measure heat on the ground, pinpointing new wildfire outbreaks from above. 

The company behind the balloons, called Urban Sky, also plans to use them to  understand conditions on the ground before fires start. Approximately 237,500 acres burn in Colorado annually, according to 2011–2020 data from the Rocky Mountain Area Coordination Center. The hope is that this new high-altitude tool might allow humans to manage—or at least understand—those blazes better.

“Wildfire is a natural part of ecosystems,” says Michael Falkowski, manager of the wildland fire programs at NASA. But climate change has proved to be an accelerant, rendering fires bigger, more intense, and more frequent. At the same time, more people are living closer to wild spaces, and the US’s history of fire suppression, which has crowded forests and left old and dead vegetation sitting around, is fanning the flames. 

To deal with modern fires, Falkowski says, researchers and fire agencies have to gather data before those fires start and after they’re done smoldering, not just as they’re burning. That makes it possible to understand the risks ahead of time and try to mitigate them, track ongoing blazes, and understand the threats fires pose to communities and the environment.

Before a fire takes hold, researchers can map vegetation and estimate how wet or dry it is. During a fire, they can map where and how hot the activity is. When it’s all over, they can assess the severity of the burn and track air quality.

Pass Fire (New Mexico) 3.5m Infrared Sample from Urban Sky Microballoon.
An infrared image of the 2023 Pass Fire in New Mexico, taken by an Urban Sky balloon.
COURTESY URBAN SKY

Still, the most acute phase is obviously the one when the fire is actually burning. In the heat of that moment, it can be hard to get a handle on when and where, exactly, the fire is taking hold. Satellites do some of that work, surveying large areas all at once. But the primary governmental satellites produce pictures with pixels around 300 meters across, and they can’t always get a super timely look at a given spot, since their view is limited by their orbit. 

Airplanes and helicopters can map a fire’s extent in more detail, but they’re expensive to operate and dangerous to fly. They have to coordinate with other aircraft and have smaller views, being closer to the ground. They’re also a limited resource. 

Urban Sky aims to combine the advantages of satellites and aircraft by using relatively inexpensive high-altitude balloons that can fly above the fray—out of the way of airspace restrictions, other aircraft, and the fire itself. The system doesn’t put a human pilot at risk and has an infrared sensor system called HotSpot that provides a sharp, real-time picture, with pixels 3.5 meters across. “We targeted that resolution with the goal of being able to see a single burning tree,” says Jared Leidich, chief technology officer at Urban Sky. “And so that would show up essentially as one pixel—one hot pixel.” The company has some competition: Others, like Aerostar and LUX Aerobot, also make balloons that can monitor wildfires.

The Urban Sky team has launched balloons in previous tests, but in August, the technology will monitor potential fires for an actual (unspecified) customer. Sending the balloon-lofted HotSpot up will be a surprisingly simple affair, thanks to the balloon’s relatively small size: While the company makes several sizes, the original is about as big as a van at launch, inflating to the size of a small garage once it’s aloft and surrounded by lower-pressure air. The Urban Sky team uses weather software to calculate where to launch a balloon so that it will drift over the fire at the right elevation. Then the team packs one up, along with compressed helium or hydrogen gas, and drives a truck out to that location. The balloon is hooked onto a mast jutting from the vehicle, filled up with the lighter-than-air molecules, and released. The whole process takes about 10 minutes. 

Once the balloon hits its cruising altitude, the HotSpot sensor turns on. Through satellite communication networks, an onboard processor sends real-time information about actual hot spots back to people on the ground. 

The balloons can hover over a fire for about 18 hours, using the whims of the atmosphere to stay in place. They fly near the top of the troposphere and the bottom of the next atmospheric layer: the stratosphere. “Those often have winds going in different directions,” explains Leidich. To move back and forth, the balloon simply has to go up or down. 

Urban Sky’s unnamed customer for its August deployment takes data on wind patterns and fuels (also known as trees, bushes, and grass) to try to understand the spots where fires are most likely to start and spread. It is interested in integrating Urban Sky’s on-the-ground (read: in-the-air) data on where fires actually do break out. “They want to add an extra step to the process where they actually scan the areas that are high risk,” says Leidich.

During the campaign, if officials identify or suspect a fire, Urban Sky can send out the truck. “We put a balloon up over the area to scan the area and say, ‘Yes, there is a fire. Here it is,’” says Leidich. 

An Urban Sky Microballoon pictured shortly after launch near Greeley, CO.

COURTESY URBAN SKY

If they get yeses where they should and nos where there is nothing to see, the proof of concept could lead to wider adoption of the HotSpot system, perhaps offering a simple and timely way for other regions to get a handle on their own fires.

This year, Urban Sky also has a grant through NASA’s FireSense program, which aims to find innovative ways to learn about all three fire phases (before, during, and after). At the moment, the August campaign and the NASA program are the primary customers for Hot Spot, although the company also sells regularly updated aerial images of 12 cities in the western US.

“It’s kind of an interesting technology to be able to do this active fire detection and tracking from a high-altitude platform,” Falkowski says of Urban Sky’s balloons. 

With NASA’s support, the team is hoping to redesign the system for longer flights, build in a more robust communication system, and incorporate a sensor that captures blue, green, and near-infrared light, which would make it possible to understand those plant-based “fuels” better and assign risk scores to forests accordingly. Next year the team is planning to again hover over real fires, this time for NASA.

And there will always be fires to hover over. As there always have been, Falkowski points out. “Fire is not a bad thing,” he says. “These ecosystems evolved with fire. The problem is humans are getting too close to places that just need to burn.”

Sarah Scoles is a Colorado-based science journalist and the author, most recently, of the book Countdown: The Blinding Future of Nuclear Weapons.