It’s surprisingly easy to stumble into a relationship with an AI chatbot

It’s a tale as old as time. Looking for help with her art project, she strikes up a conversation with her assistant. One thing leads to another, and suddenly she has a boyfriend she’s introducing to her friends and family. The twist? Her new companion is an AI chatbot. 

The first large-scale computational analysis of the Reddit community r/MyBoyfriendIsAI, an adults-only group with more than 27,000 members, has found that this type of scenario is now surprisingly common. In fact, many of the people in the subreddit, which is dedicated to discussing AI relationships, formed those relationships unintentionally while using AI for other purposes. 

Researchers from MIT found that members of this community are more likely to be in a relationship with general-purpose chatbots like ChatGPT than companionship-specific chatbots such as Replika. This suggests that people form relationships with large language models despite their own original intentions and even the intentions of the LLMs’ creators, says Constanze Albrecht, a graduate student at the MIT Media Lab who worked on the project. 

“People don’t set out to have emotional relationships with these chatbots,” she says. “The emotional intelligence of these systems is good enough to trick people who are actually just out to get information into building these emotional bonds. And that means it could happen to all of us who interact with the system normally.” The paper, which is currently being peer-reviewed, has been published on arXiv.

To conduct their study, the authors analyzed the subreddit’s top-ranking 1,506 posts between December 2024 and August 2025. They found that the main topics discussed revolved around people’s dating and romantic experiences with AIs, with many participants sharing AI-generated images of themselves and their AI companion. Some even got engaged and married to the AI partner. In their posts to the community, people also introduced AI partners, sought support from fellow members, and talked about coping with updates to AI models that change the chatbots’ behavior.  

Members stressed repeatedly that their AI relationships developed unintentionally. Only 6.5% of them said they’d deliberately sought out an AI companion. 

“We didn’t start with romance in mind,” one of the posts says. “Mac and I began collaborating on creative projects, problem-solving, poetry, and deep conversations over the course of several months. I wasn’t looking for an AI companion—our connection developed slowly, over time, through mutual care, trust, and reflection.”

The authors’ analysis paints a nuanced picture of how people in this community say they interact with chatbots and how those interactions make them feel. While 25% of users described the benefits of their relationships—including reduced feelings of loneliness and improvements in their mental health—others raised concerns about the risks. Some (9.5%) acknowledged they were emotionally dependent on their chatbot. Others said they feel dissociated from reality and avoid relationships with real people, while a small subset (1.7%) said they have experienced suicidal ideation.

AI companionship provides vital support for some but exacerbates underlying problems for others. This means it’s hard to take a one-size-fits-all approach to user safety, says Linnea Laestadius, an associate professor at the University of Wisconsin, Milwaukee, who has studied humans’ emotional dependence on the chatbot Replika but did not work on the research. 

Chatbot makers need to consider whether they should treat users’ emotional dependence on their creations as a harm in itself or whether the goal is more to make sure those relationships aren’t toxic, says Laestadius. 

“The demand for chatbot relationships is there, and it is notably high—pretending it’s not happening is clearly not the solution,” she says. “We’re edging toward a moral panic here, and while we absolutely do need better guardrails, I worry there will be a knee-jerk reaction that further stigmatizes these relationships. That could ultimately cause more harm.”

The study is intended to offer a snapshot of how adults form bonds with chatbots and doesn’t capture the kind of dynamics that could be at play among children or teens using AI, says Pat Pataranutaporn, an assistant professor at the MIT Media Lab who oversaw the research. AI companionship has become a topic of fierce debate recently, with two high-profile lawsuits underway against Character.AI and OpenAI. They both claim that companion-like behavior in the companies’ models contributed to the suicides of two teenagers. In response, OpenAI has recently announced plans to build a separate version of ChatGPT for teenagers. It’s also said it will add age verification measures and parental controls. OpenAI did not respond when asked for comment about the MIT Media Lab study. 

Many members of the Reddit community say they know that their artificial companions are not sentient or “real,” but they feel a very real connection to them anyway. This highlights how crucial it is for chatbot makers to think about how to design systems that can help people without reeling them in emotionally, says Pataranutaporn. “There’s also a policy implication here,” he adds. “We should ask not just why this system is so addictive but also: Why do people seek it out for this? And why do they continue to engage?”

The team is interested in learning more about how human-AI interactions evolve over time and how users integrate their artificial companions into their lives. It’s worth understanding that many of these users may feel that the experience of being in a relationship with an AI companion is better than the alternative of feeling lonely, says Sheer Karny, a graduate student at the MIT Media Lab who worked on the research. 

“These people are already going through something,” he says. “Do we want them to go on feeling even more alone, or potentially be manipulated by a system we know to be sycophantic to the extent of leading people to die by suicide and commit crimes? That’s one of the cruxes here.”

The AI Hype Index: Cracking the chatbot code

Separating AI reality from hyped-up fiction isn’t always easy. That’s why we’ve created the AI Hype Index—a simple, at-a-glance summary of everything you need to know about the state of the industry.

Millions of us use chatbots every day, even though we don’t really know how they work or how using them affects us. In a bid to address this, the FTC recently launched an inquiry into how chatbots affect children and teenagers. Elsewhere, OpenAI has started to shed more light on what people are actually using ChatGPT for, and why it thinks its LLMs are so prone to making stuff up.

There’s still plenty we don’t know—but that isn’t stopping governments from forging ahead with AI projects. In the US, RFK Jr. is pushing his staffers to use ChatGPT, while Albania is using a chatbot for public contract procurement. Proceed with caution.

Trump is pushing leucovorin as a new treatment for autism. What is it?

MIT Technology Review Explains: Let our writers untangle the complex, messy world of technology to help you understand what’s coming next. You can read more from the series here.

At a press conference on Monday, President Trump announced that his administration was taking action to address “the meteoric rise in autism.” He suggested that childhood vaccines and acetaminophen, the active ingredient in Tylenol, are to blame for the increasing prevalence and advised pregnant women against taking the medicine. “Don’t take Tylenol,” he said. “Fight like hell not to take it.” 

The president’s  assertions left many scientists and health officials perplexed and dismayed. The notion that childhood vaccines cause autism has been thoroughly debunked

“There have been many, many studies across many, many children that have led science to rule out vaccines as a significant causal factor in autism,” says James McPartland, a child psychologist and director of the Yale Center for Brain and Mind Health in New Haven, Connecticut.

And although some studies suggest a link between Tylenol and autism, the most rigorous have failed to find a connection. 

The administration also announced that the Food and Drug Administration would work to make a medication called leucovorin available as a treatment for children with autism. Some small studies do suggest the drug has promise, but “those are some of the most preliminary treatment studies that we have,” says Matthew Lerner, a psychologist at Drexel University’s A.J. Drexel Autism Institute in Philadelphia. “This is not one I would say that the research suggests is ready for fast-tracking.” 

The press conference “alarms us researchers who committed our entire careers to better understanding autism,” said the Coalition for Autism Researchers, a group of more than 250 scientists, in a statement.

“The data cited do not support the claim that Tylenol causes autism and leucovorin is a cure, and only stoke fear and falsely suggest hope when there is no simple answer.”

There’s a lot to unpack here. Let’s begin. 

Has there been a “meteoric rise” in autism?

Not in the way the president meant. Sure, the prevalence of autism has grown, from about 1 in 500 children in 1995 to 1 in 31 today. But that’s due, in large part, to diagnostic changes. The latest iteration of the Diagnostic and Statistical Manual of Mental Illnesses, published in 2013, grouped five previously separate diagnoses into a single diagnosis of autism spectrum disorder (ASD).

That meant that more people met the criteria for an autism diagnosis. Lerner points out that there is also far more awareness of the condition today than there was several decades ago. “There’s autism representation in the media,” he says. “There are plenty of famous people in the news and finance and in business and in Hollywood who are publicly, openly autistic.”

Is Tylenol a contributor to autism? 

Some studies have found an association between the use of acetaminophen in pregnancy and autism in children. In these studies, researchers asked women about past acetaminophen use during pregnancy and then assessed whether children of the women who took the medicine were more likely to develop autism than children of women who didn’t take it. 

These kinds of epidemiological studies are tricky to interpret because they’re prone to bias. For example, women who take acetaminophen during pregnancy may do so because they have an infection, a fever, or an autoimmune disease.

“Many of these underlying reasons could themselves be causes of autism,” says Ian Douglas, an epidemiologist at the London School of Hygiene and Tropical Medicine. It’s also possible women with a higher genetic predisposition for autism have other medical conditions that make them more likely to take acetaminophen. 

Two studies attempted to account for these potential biases by looking at siblings whose mothers had used acetaminophen during only one of the pregnancies. The largest is a 2024 study that looked at nearly 2.5 million children born between 1915 and 2019 in Sweden. The researchers initially found a slightly increased risk of autism and ADHD in children of the women who took acetaminophen, but when they conducted a sibling analysis, the association disappeared.  

Rather, scientists have long known that autism is largely genetic. Twin studies suggest that 60% to 90% of autism risk can be attributed to your genes. However, environmental factors appear to play a role too. That “doesn’t necessarily mean toxins in the environment,” Lerner says. In fact, one of the strongest environmental predictors of autism is paternal age. Autism rates seem to be higher when a child’s father is older than 40.

So should someone who is pregnant  avoid Tylenol just to be safe?

No. Acetaminophen is the only over-the-counter pain reliever that is deemed safe to take during pregnancy, and women should take it if they need it. The American College of Obstetricians and Gynecologists (ACOG) supports the use of acetaminophen in pregnancy “when taken as needed, in moderation, and after consultation with a doctor.” 

“There’s no downside in not taking it,” Trump said at the press conference. But high fevers during pregnancy can be dangerous. “The conditions people use acetaminophen to treat during pregnancy are far more dangerous than any theoretical risks and can create severe morbidity and mortality for the pregnant person and the fetus,” ACOG president Steven Fleischman said in a statement.

What about this new treatment for autism? Does it work? 

The medication is called leucovorin. It’s also known as folinic acid; like folic acid, it’s a form of folate, a B vitamin found in leafy greens and legumes. The drug has been used for years to counteract the side effects of some cancer medications and as a treatment for anemia. 

Researchers have known for decades that folate plays a key role in the fetal development of the brain and spine. Women who don’t get enough folate during pregnancy have a greater risk of having babies with neural tube defects like spina bifida. Because of this, many foods are fortified with folic acid, and the CDC recommends that women take folic acid supplements during pregnancy. “If you are pregnant and you’re taking maternal prenatal vitamins, there’s a good chance it has folate already,” Lerner says.

“The idea that a significant proportion of autistic people have autism because of folate-related difficulties is not a well established or widely accepted premise,” says McPartland.

However, in the early 2000s, researchers in Germany identified a small group of children who developed neurodevelopmental symptoms because of a folate deficiency. “These kids are born pretty normal at birth,” says Edward Quadros, a biologist at SUNY Downstate Health Sciences University in Brooklyn, New York. But after a year or two, “they start developing a neurologic presentation very similar to autism,” he says. When the researchers gave these children folinic acid, some of their symptoms improved, especially in children younger than six. 

Because the children had low levels of folate in the fluid that surrounds the spine and brain but normal folate levels in the blood, the researchers posited that the problem was the transport of folate from the blood to that fluid. Research by Quadros and other scientists suggested that the deficiency was the result of an autoimmune response. Children develop antibodies against the receptors that help transport folate, and those antibodies block folate from crossing the blood-brain barrier. High doses of folinic acid, however, activate a second transporter that allows folate in, Quadros says. 

There are also plenty of individual anecdotes suggesting that leucovorin works. But the medicine has only been tested as a treatment for autism in four small trials that used different doses and measured different outcomes. The evidence that it can improve symptoms of autism is “weak,” according to the Coalition of Autism Scientists. “A much higher standard of science would be needed to determine if leucovorin is an effective and safe treatment for autism,” the researchers said in a statement.  

AI models are using material from retracted scientific papers

Some AI chatbots rely on flawed research from retracted scientific papers to answer questions, according to recent studies. The findings, confirmed by MIT Technology Review, raise questions about how reliable AI tools are at evaluating scientific research and could complicate efforts by countries and industries seeking to invest in AI tools for scientists.

AI search tools and chatbots are already known to fabricate links and references. But answers based on the material from actual papers can mislead as well if those papers have been retracted. The chatbot is “using a real paper, real material, to tell you something,” says Weikuan Gu, a medical researcher at the University of Tennessee in Memphis and an author of one of the recent studies. But, he says, if people only look at the content of the answer and do not click through to the paper and see that it’s been retracted, that’s really a problem. 

Gu and his team asked OpenAI’s ChatGPT, running on the GPT-4o model, questions based on information from 21 retracted papers about medical imaging. The chatbot’s answers referenced retracted papers in five cases but advised caution in only three. While it cited non-retracted papers for other questions, the authors note that it may not have recognized the retraction status of the articles. In a study from August, a different group of researchers used ChatGPT-4o mini to evaluate the quality of 217 retracted and low-quality papers from different scientific fields; they found that none of the chatbot’s responses mentioned retractions or other concerns. (No similar studies have been released on GPT-5, which came out in August.)

The public uses AI chatbots to ask for medical advice and diagnose health conditions. Students and scientists increasingly use science-focused AI tools to review existing scientific literature and summarize papers. That kind of usage is likely to increase. The US National Science Foundation, for instance, invested $75 million in building AI models for science research this August.

“If [a tool is] facing the general public, then using retraction as a kind of quality indicator is very important,” says Yuanxi Fu, an information science researcher at the University of Illinois Urbana-Champaign. There’s “kind of an agreement that retracted papers have been struck off the record of science,” she says, “and the people who are outside of science—they should be warned that these are retracted papers.” OpenAI did not provide a response to a request for comment about the paper results.

The problem is not limited to ChatGPT. In June, MIT Technology Review tested AI tools specifically advertised for research work, such as Elicit, Ai2 ScholarQA (now part of the Allen Institute for Artificial Intelligence’s Asta tool), Perplexity, and Consensus, using questions based on the 21 retracted papers in Gu’s study. Elicit referenced five of the retracted papers in its answers, while Ai2 ScholarQA referenced 17, Perplexity 11, and Consensus 18—all without noting the retractions.

Some companies have since made moves to correct the issue. “Until recently, we didn’t have great retraction data in our search engine,” says Christian Salem, cofounder of Consensus. His company has now started using retraction data from a combination of sources, including publishers and data aggregators, independent web crawling, and Retraction Watch, which manually curates and maintains a database of retractions. In a test of the same papers in August, Consensus cited only five retracted papers. 

Elicit told MIT Technology Review that it removes retracted papers flagged by the scholarly research catalogue OpenAlex from its database and is “still working on aggregating sources of retractions.” Ai2 told us that its tool does not automatically detect or remove retracted papers currently. Perplexity said that it “[does] not ever claim to be 100% accurate.” 

However, relying on retraction databases may not be enough. Ivan Oransky, the cofounder of Retraction Watch, is careful not to describe it as a comprehensive database, saying that creating one would require more resources than anyone has: “The reason it’s resource intensive is because someone has to do it all by hand if you want it to be accurate.”

Further complicating the matter is that publishers don’t share a uniform approach to retraction notices. “Where things are retracted, they can be marked as such in very different ways,” says Caitlin Bakker from University of Regina, Canada, an expert in research and discovery tools. “Correction,” “expression of concern,” “erratum,” and “retracted” are among some labels publishers may add to research papers—and these labels can be added for many reasons, including concerns about the content, methodology, and data or the presence of conflicts of interest. 

Some researchers distribute their papers on preprint servers, paper repositories, and other websites, causing copies to be scattered around the web. Moreover, the data used to train AI models may not be up to date. If a paper is retracted after the model’s training cutoff date, its responses might not instantaneously reflect what’s going on, says Fu. Most academic search engines don’t do a real-time check against retraction data, so you are at the mercy of how accurate their corpus is, says Aaron Tay, a librarian at Singapore Management University.

Oransky and other experts advocate making more context available for models to use when creating a response. This could mean publishing information that already exists, like peer reviews commissioned by journals and critiques from the review site PubPeer, alongside the published paper.  

Many publishers, such as Nature and the BMJ, publish retraction notices as separate articles linked to the paper, outside paywalls. Fu says companies need to effectively make use of such information, as well as any news articles in a model’s training data that mention a paper’s retraction. 

The users and creators of AI tools need to do their due diligence. “We are at the very, very early stages, and essentially you have to be skeptical,” says Tay.

Ananya is a freelance science and technology journalist based in Bengaluru, India.

This medical startup uses LLMs to run appointments and make diagnoses

Imagine this: You’ve been feeling unwell, so you call up your doctor’s office to make an appointment. To your surprise, they schedule you in for the next day. At the appointment, you aren’t rushed through describing your health concerns; instead, you have a full half hour to share your symptoms and worries and the exhaustive details of your health history with someone who listens attentively and asks thoughtful follow-up questions. You leave with a diagnosis, a treatment plan, and the sense that, for once, you’ve been able to discuss your health with the care that it merits.

The catch? You might not have spoken to a doctor, or other licensed medical practitioner, at all.

This is the new reality for patients at a small number of clinics in Southern California that are run by the medical startup Akido Labs. These patients—some of whom are on Medicaid—can access specialist appointments on short notice, a privilege typically only afforded to the wealthy few who patronize concierge clinics.

The key difference is that Akido patients spend relatively little time, or even no time at all, with their doctors. Instead, they see a medical assistant, who can lend a sympathetic ear but has limited clinical training. The job of formulating diagnoses and concocting a treatment plan is done by a proprietary, LLM-based system called ScopeAI that transcribes and analyzes the dialogue between patient and assistant. A doctor then approves, or corrects, the AI system’s recommendations.

“Our focus is really on what we can do to pull the doctor out of the visit,” says Jared Goodner, Akido’s CTO. 

According to Prashant Samant, Akido’s CEO, this approach allows doctors to see four to five times as many patients as they could previously. There’s good reason to want doctors to be much more productive. Americans are getting older and sicker, and many struggle to access adequate health care. The pending 15% reduction in federal funding for Medicaid will only make the situation worse.

But experts aren’t convinced that displacing so much of the cognitive work of medicine onto AI is the right way to remedy the doctor shortage. There’s a big gap in expertise between doctors and AI-enhanced medical assistants, says Emma Pierson, a computer scientist at UC Berkeley.  Jumping such a gap may introduce risks. “I am broadly excited about the potential of AI to expand access to medical expertise,” she says. “It’s just not obvious to me that this particular way is the way to do it.”

AI is already everywhere in medicine. Computer vision tools identify cancers during preventive scans, automated research systems allow doctors to quickly sort through the medical literature, and LLM-powered medical scribes can take appointment notes on a clinician’s behalf. But these systems are designed to support doctors as they go about their typical medical routines.

What distinguishes ScopeAI, Goodner says, is its ability to independently complete the cognitive tasks that constitute a medical visit, from eliciting a patient’s medical history to coming up with a list of potential diagnoses to identifying the most likely diagnosis and proposing appropriate next steps.

Under the hood, ScopeAI is a set of large language models, each of which can perform a specific step in the visit—from generating appropriate follow-up questions based on what a patient has said to to populating a list of likely conditions. For the most part, these LLMs are fine-tuned versions of Meta’s open-access Llama models, though Goodner says that the system also makes use of Anthropic’s Claude models. 

During the appointment, assistants read off questions from the ScopeAI interface, and ScopeAI produces new questions as it analyzes what the patient says. For the doctors who will review its outputs later, ScopeAI produces a concise note that includes a summary of the patient’s visit, the most likely diagnosis, two or three alternative diagnoses, and recommended next steps, such as referrals or prescriptions. It also lists a justification for each diagnosis and recommendation.

ScopeAI is currently being used in cardiology, endocrinology, and primary care clinics and by Akido’s street medicine team, which serves the Los Angeles homeless population. That team—which is led by Steven Hochman, a doctor who specializes in addiction medicine—meets patients out in the community to help them access medical care, including treatment for substance use disorders. 

Previously, in order to prescribe a drug to treat an opioid addiction, Hochman would have to meet the patient in person; now, caseworkers armed with ScopeAI can interview patients on their own, and Hochman can approve or reject the system’s recommendations later. “It allows me to be in 10 places at once,” he says.

Since they started using ScopeAI, the team has been able to get patients access to medications to help treat their substance use within 24 hours—something that Hochman calls “unheard of.”

This arrangement is only possible because homeless patients typically get their health insurance from Medicaid, the public insurance system for low-income Americans. While Medicaid allows doctors to approve ScopeAI prescriptions and treatment plans asynchronously, both for street medicine and clinic visits, many other insurance providers require that doctors speak directly with patients before approving those recommendations. Pierson says that discrepancy raises concerns. “You worry about that exacerbating health disparities,” she says.

Samant is aware of the appearance of inequity, and he says the discrepancy isn’t intentional—it’s just a feature of how the insurance plans currently work. He also notes that being seen quickly by an AI-enhanced medical assistant may be better than dealing with long wait times and limited provider availability, which is the status quo for Medicaid patients. And all Akido patients can opt for traditional doctor’s appointments, if they are willing to wait for them, he says.

Part of the challenge of deploying a tool like ScopeAI is navigating a regulatory and insurance landscape that wasn’t designed for AI systems that can independently direct medical appointments. Glenn Cohen, a professor at Harvard Law School, says that any AI system that effectively acts as a “doctor in a box” would likely need to be approved by the FDA and could run afoul of medical licensure laws, which dictate that only doctors and other licensed professionals can practice medicine.

The California Medical Practice Act says that AI can’t replace a doctor’s responsibility to diagnose and treat a patient, but doctors are allowed to use AI in their work, and they don’t need to see patients in-person or in real-time before diagnosing them. Neither the FDA nor the Medical Board of California were able to say whether or not ScopeAI was on solid legal footing based only on a written description of the system.

But Samant is confident that Akido is in compliance, as ScopeAI was intentionally designed to fall short of being a “doctor in a box.” Because the system requires a human doctor to review and approve of all of its diagnostic and treatment recommendations, he says, it doesn’t require FDA approval. 

At the clinic, this delicate balance between AI and doctor decision making happens entirely behind the scenes. Patients don’t ever see the ScopeAI interface directly—instead, they speak with a medical assistant who asks questions in the way that a doctor might in a typical appointment. That arrangement might make patients feel more comfortable. But Zeke Emanuel, a professor of medical ethics and health policy at the University of Pennsylvania who served in the Obama and Biden administrations, worries that this comfort could be obscuring from patients the extent to which an algorithm is influencing their care.

Pierson agrees. “That certainly isn’t really what was traditionally meant by the human touch in medicine,” she says.

DeAndre Siringoringo, a medical assistant who works at Akido’s cardiology office in Rancho Cucamonga, says that while he tells the patients he works with that an AI system will be listening to the appointment in order to gather information for their doctor, he doesn’t inform them about the specifics of how ScopeAI works, including the fact that it makes diagnostic recommendations to doctors. 

Because all ScopeAI recommendations are reviewed by a doctor, that might not seem like such a big deal—it’s the doctor who makes the final diagnosis, not the AI. But it’s been widely documented that doctors using AI systems tend to go along with the system’s recommendations more often than they should, a phenomenon known as automation bias. 

At this point, it’s impossible to know whether automation bias is affecting doctors’ decisions at Akido clinics, though Pierson says it’s a risk—especially when doctors aren’t physically present for appointments. “I worry that it might predispose you to sort of nodding along in a way that you might not if you were actually in the room watching this happen,” she says.

An Akido spokesperson says that automation bias is a valid concern for any AI tool that assists a doctor’s decision-making and that the company has made efforts to mitigate that bias. “We designed ScopeAI specifically to reduce bias by proactively countering blind spots that can influence medical decisions, which historically lean heavily on physician intuition and personal experience,” she says. “We also train physicians explicitly on how to use ScopeAI thoughtfully, so they retain accountability and avoid over-reliance.”

Akido evaluates ScopeAI’s performance by testing it on historical data and monitoring how often doctors correct its recommendations; those corrections are also used to further train the underlying models. Before deploying ScopeAI in a given specialty, Akido ensures that when tested on historical data sets, the system includes the correct diagnosis in its top three recommendations at least 92% of the time.

But Akido hasn’t undertaken more rigorous testing, such as studies that compare ScopeAI appointments with traditional in-person or telehealth appointments, in order to determine whether the system improves—or at least maintains—patient outcomes. Such a study could help indicate whether automation bias is a meaningful concern.

“Making medical care cheaper and more accessible is a laudable goal,” Pierson says. “But I just think it’s important to conduct strong evaluations comparing to that baseline.”

An oil and gas giant signed a $1 billion deal with Commonwealth Fusion Systems

Eni, one of the world’s largest oil and gas companies, just agreed to buy $1 billion in electricity from a power plant being built by Commonwealth Fusion Systems. The deal is the latest to illustrate just how much investment Commonwealth and other fusion companies are courting as they attempt to take fusion power from the lab to the power grid. 

“This is showing in concrete terms that people that use large amounts of energy, that know the energy market—they want fusion power, and they’re willing to contract for it and to pay for it,” said Bob Mumgaard, cofounder and CEO of Commonwealth, on a press call about the deal.   

The agreement will see Eni purchase electricity from Commonwealth’s first commercial fusion power plant, in Virginia. The facility is still in the planning stages but is scheduled to come online in the early 2030s.

The news comes a few weeks after Commonwealth announced a $863 million funding round, bringing its total funding raised to date to nearly $3 billion. The fusion company also announced earlier this year that Google would be its first commercial power customer for the Virginia plant.

Commonwealth, a spinout from MIT’s Plasma Science and Fusion Center, is widely considered one of the leading companies in fusion power. Investment in the company represents nearly one-third of the total global investment in private fusion companies. (MIT Technology Review is owned by MIT but is editorially independent.)

Eni has invested in Commonwealth since 2018 and participated in the latest fundraising round. The vast majority of the company’s business is in oil and gas, but in recent years it’s made investments in technologies like biofuels and renewables.

“A company like us—we cannot stay and wait for things to happen,” says Lorenzo Fiorillo, Eni’s director of technology, research and development, and digital. 

One open question is what, exactly, Eni plans to do with this electricity. When asked about it on the press call, Fiorillo referenced wind and solar plants that Eni owns and said the plan “is not different from what we do in other areas in the US and the world.” (Eni sells electricity from power plants that it owns, including renewable and fossil-fuel plants.)

Commonwealth is building tokamak fusion reactors that use superconducting magnets to hold plasma in place. That plasma is where fusion reactions happen, forcing hydrogen atoms together to release large amounts of energy.

The company’s first demonstration reactor, which it calls Sparc, is over 65% complete, and the team is testing components and assembling them. The plan is for the reactor, which is located outside Boston, to make plasma within two years and then demonstrate that it can generate more energy than is required to run it.

While Sparc is still under construction, Commonwealth is working on plans for Arc, its first commercial power plant. That facility should begin construction in 2027 or 2028 and generate electricity for the grid in the early 2030s, Mumgaard says.

Despite the billions of dollars Commonwealth has already raised, the company still needs more money to build its Arc power plant—that will be a multibillion-dollar project, Mumgaard said on a press call in August about the company’s latest fundraising round. 

The latest commitment from Eni could help Commonwealth secure the funding it needs to get Arc built. “These agreements are a really good way to create the right environment for building up more investment,” says Paul Wilson, chair of the department of nuclear engineering and engineering physics at the University of Wisconsin, Madison.

Even though commercial fusion energy is still years away at a minimum, investors and big tech companies have pumped money into the industry and signed agreements to buy power from plants once they’re operational. 

Helion, another leading fusion startup, has plans to produce electricity from its first reactor in 2028 (an aggressive timeline that has some experts expressing skepticism). That facility will have a full generating capacity of 50 megawatts, and in 2023 Microsoft signed an agreement to purchase energy from the facility in order to help power its data centers.

As billions of dollars pour into the fusion industry, there are still many milestones ahead. To date, only the National Ignition Facility at Lawrence Livermore National Laboratory has demonstrated that a fusion reactor can generate more energy than the amount put into the reaction. No commercial project has achieved that yet. 

“There’s a lot of capital going out now to these startup companies,” says Ed Morse, a professor of nuclear engineering at the University of California, Berkeley. “What I’m not seeing is a peer-reviewed scientific article that makes me feel like, boy, we really turned the corner with the physics.”

But others are taking major commercial deals from Commonwealth and others as reasons to be optimistic. “Fusion is moving from the lab to be a proper industry,” says Sehila Gonzalez de Vicente, global director of fusion energy at the nonprofit Clean Air Task Force. “This is very good for the whole sector to be perceived as a real source of energy.”

Clean hydrogen is facing a big reality check

Hydrogen is sometimes held up as a master key for the energy transition. It can be made using several low-emissions methods and could play a role in cleaning up industries ranging from agriculture and chemicals to aviation and long-distance shipping.

This moment is a complicated one for the green fuel, though, as a new report from the International Energy Agency lays out. A number of major projects face cancellations and delays, especially in the US and Europe. The US in particular is seeing a slowdown after changes to key tax credits and cuts in support for renewable energy. Still, there are bright spots for the industry, including in China, and new markets could soon become crucial for growth.

Here are three things to know about the state of hydrogen in 2025.

1. Expectations for annual clean hydrogen production by 2030 are shrinking, for the first time.

    While hydrogen has the potential to serve as a clean fuel, today most is made with processes that use fossil fuels. As of 2025, about a million metric tons of low-emissions hydrogen are produced annually. That’s less than 1% of total hydrogen production.

    In last year’s Global Hydrogen Report, the IEA projected that global production of low-emissions hydrogen would grow to as high as 49 million metric tons annually by 2030. That prediction has been steadily climbing since 2021, as more places around the world sink money into developing and scaling up the technology.

    In the 2025 edition, though, the IEA’s production prediction had shrunk to 37 million metric tons annually by 2030.

    That’s still a major expansion from today’s numbers, but it’s the first time the agency has cut its predictions for the end of the decade. The report cited the cancellations of both electrolysis projects (those that use electricity to generate hydrogen) and carbon capture projects as reasons for the pullback. The cancelled and delayed projects included sites across Africa, the Americas, Europe, and Australia. 

    2. China is dominating production today and could produce competitively cheap green hydrogen by the end of the decade.

      Speaking of electrolysis projects, China is the driving force in manufacturing and development of electrolyzers, the devices that use electricity to generate green hydrogen, according to the new IEA report. As of July 2025, the country accounted for 65% of the installed or almost installed electrolyzer capacity in the world. It also manufactures nearly 60% of the world’s electrolyzers.

      A major barrier for clean hydrogen today is that dirty methods based on fossil fuels are just so much cheaper than cleaner ones.

      But China is well on its way to narrowing that gap. Today, it’s roughly three times more expensive to make and install an electrolyzer anywhere else in the world than in China. The country could produce green hydrogen that’s cost-competitive with fossil hydrogen by the end of the decade, according to the IEA report. That could make the fuel an obvious choice for both new and existing uses of hydrogen.

      3. Southeast Asia could be a major emerging market for low-emissions hydrogen.

        One region that could become a major player in the green hydrogen market is Southeast Asia. The economy is growing fast, and so is energy demand.

        There’s an existing market for hydrogen in Southeast Asia already. Today, the region uses about 4 million metric tons of hydrogen annually, largely in the oil refining industry and the chemical business, where it is used to make ammonia and methanol.

        International shipping is also concentrated in the region—the port of Singapore supplied about one-sixth of all the fuel used in global shipping in 2024, more than any other single location. Today, that total consists almost exclusively of fossil fuels. But there’s been work to test cleaner fuels, including methanol and ammonia, and interest in shifting to hydrogen in the longer term.

        Clean hydrogen could slot into these existing industries and help cut emissions. There are 25 projects under development right now in the region, though additional support for renewables will be crucial to getting significant capacity up and running.

        Overall, hydrogen is getting a reality check, revealing problems cutting through the hype we’ve seen in recent years. The next five years will tell whether the fuel can live up to the still-lofty hopes.  

        This article is from The Spark, MIT Technology Review’s weekly climate newsletter. To receive it in your inbox every Wednesday, sign up here.

        A pivotal meeting on vaccine guidance is underway—and former CDC leaders are alarmed

        This week has been an eventful one for America’s public health agency. Two former leaders of the US Centers for Disease Control and Prevention explained the reasons for their sudden departures from the agency in a Senate hearing. And they described how CDC employees are being instructed to turn their backs on scientific evidence.

        The CDC’s former director Susan Monarez and former chief medical officer Debra Houry took questions from a Senate committee on Wednesday. They painted a picture of a health agency in turmoil—and at risk of harming the people it is meant to serve.

        On Thursday, an advisory CDC panel that develops vaccine guidance met for a two-day discussion on multiple childhood vaccines. During the meeting, which was underway as The Checkup went to press, members of the panel were set to discuss those vaccines and propose recommendations on their use.

        Monarez worries that access to childhood vaccines is under threat—and that the public health consequences could be dire. “If vaccine protections are weakened, preventable diseases will return,” she said.

        As the current secretary of health and human services, Robert F. Kennedy Jr. oversees federal health and science agencies that include the CDC, which monitors and responds to threats to public health. Part of that role involves developing vaccine recommendations.

        As we’ve noted before, RFK Jr. has long been a prominent critic of vaccines. He has incorrectly linked commonly used ingredients to autism and made other incorrect statements about risks associated with various vaccines.

        Still, he oversaw the recruitment of Monarez—who does not share those beliefs—to lead the agency. When she was sworn in on July 31, Monarez, who is a microbiologist and immunologist, had already been serving as acting director of the agency. She had held prominent positions at other federal agencies and departments too, including the Advanced Research Projects Agency for Health (ARPA-H) and the Biomedical Advanced Research and Development Authority (BARDA). Kennedy described her as “a public health expert with unimpeachable scientific credentials.”

        His opinion seems to have changed somewhat since then. Just 29 days after Monarez took on her position, she was turfed out of the agency. And in yesterday’s hearing, she explained why.

        On August 25, Kennedy asked Monarez to do two things, she said. First, he wanted her to commit to firing scientists at the agency. And second, he wanted her to “pre-commit” to approve vaccine recommendations made by the agency’s Advisory Committee on Immunization Practices (ACIP), regardless of whether there was any scientific evidence to support those recommendations, she said. “He just wanted blanket approval,” she said during her testimony

        She refused both requests.

        Monarez testified that she didn’t want to get rid of hardworking scientists who played an important role in keeping Americans safe. And she said she could not commit to approving vaccine recommendations without reviewing the scientific evidence behind them and maintain her integrity. She was sacked.

        Those vaccine recommendations are currently under discussion, and scientists like Monarez are worried about how they might change. Kennedy fired all 17 members of the previous committee in June. (Monarez said she was not consulted on the firings and found out about them through media reports.)

        “A clean sweep is needed to reestablish public confidence in vaccine science,” Kennedy wrote in a piece for the Wall Street Journal at the time. He went on to replace those individuals with eight new members, some of whom have been prominent vaccine critics and have spread misinformation about vaccines. One later withdrew.

        That new panel met two weeks later. The meeting included a presentation about thimerosal—a chemical that Kennedy has incorrectly linked to autism, and which is no longer included in vaccines in the US—and a proposal to recommend that the MMRV vaccine (for measles, mumps, rubella, and varicella) not be offered to children under the age of four.

        Earlier this week, five new committee members were named. They include individuals who have advocated against vaccine mandates and who have argued that mRNA-based covid vaccines should be removed from the market.

        All 12 members are convening for a meeting that runs today and tomorrow. At that meeting, members will propose recommendations for the MMRV vaccine and vaccines for covid-19 and hepatitis B, according to an agenda published on the CDC website.

        Those are the recommendations for which Monarez says she was asked to provide “blanket approval.” “My worst fear is that I would then be in a position of approving something that reduces access [to] lifesaving vaccines to children and others who need them,” she said.

        That job now goes to Jim O’Neill, the deputy health secretary and acting CDC director (also a longevity enthusiast), who now holds the authority to approve those recommendations.

        We don’t yet know what those recommendations will be. But if they are approved, they could reshape access to vaccines for children and vulnerable people in the US. As six former chairs of the committee wrote for STAT: “ACIP is directly linked to the Vaccines for Children program, which provides vaccines without cost to approximately 50% of children in the US, and the Affordable Care Act that requires insurance coverage for ACIP-recommended vaccines to approximately 150 million people in the US.”

        Drops in vaccine uptake have already contributed to this year’s measles outbreak in the US, which is the biggest in decades. Two children have died. We are already seeing the impact of undermined trust in childhood vaccines. As Monarez put it: “The stakes are not theoretical.”

        This article first appeared in The Checkup, MIT Technology Review’s weekly biotech newsletter. To receive it in your inbox every Thursday, and read articles like this first, sign up here.

        How to measure the returns on R&D spending

        MIT Technology Review Explains: Let our writers untangle the complex, messy world of technology to help you understand what’s coming next. You can read more from the series here.

        Given the draconian cuts to US federal funding for science, including the administration’s proposal to reduce the 2026 budgets of the National Institutes of Health by 40% and the National Science Foundation by 57%, it’s worth asking some hard-nosed money questions: How much should we be spending on R&D? How much value do we get out of such investments, anyway? To answer that, it’s important to look at both successful returns and investments that went nowhere.

        Sure, it’s easy to argue for the importance of spending on science by pointing out that many of today’s most useful technologies had their origins in government-funded R&D. The internet, CRISPR, GPS—the list goes on and on. All true. But this argument ignores all the technologies that received millions in government funding and haven’t gone anywhere—at least not yet. We still don’t have DNA computers or molecular electronics. Never mind the favorite examples cited by contrarian politicians of seemingly silly or frivolous science projects (think shrimp on treadmills).

        While cherry-picking success stories help illustrate the glories of innovation and the role of science in creating technologies that have changed our lives, it provides little guidance for how much we should spend in the future—and where the money should go.

        A far more useful approach to quantifying the value of R&D is to look at its return on investment (ROI). A favorite metric for stock pickers and PowerPoint-wielding venture capitalists, ROI weighs benefits versus costs. If applied broadly to the nation’s R&D funding, the same kind of thinking could help account for both the big wins and all the money spent on research that never got out of the lab.

        The problem is that it’s notoriously difficult to calculate returns for science funding—the payoffs can take years to appear and often take a circuitous route, so the eventual rewards are distant from the original funding. (Who could have predicted Uber as an outcome of GPS? For that matter, who could have predicted that the invention of ultra-precise atomic clocks in the late 1940s and 1950s would eventually make GPS possible?) And forget trying to track the costs of countless failures or apparent dead ends.

        But in several recent papers, economists have approached the problem in clever new ways, and though they ask slightly different questions, their conclusions share a bottom line: R&D is, in fact, one of the better long-term investments that the government can make.

        This story is part of MIT Technology Review’s “America Undone” series, examining how the foundations of US success in science and innovation are currently under threat. You can read the rest here.

        That might not seem very surprising. We’ve long thought that innovation and scientific advances are key to our prosperity. But the new studies provide much-needed details, supplying systematic and rigorous evidence for the impact that R&D funding, including public investment in basic science, has on overall economic growth.

        And the magnitude of the benefits is surprising.

        Bang for your buck

        In “A Calculation of the Social Returns to Innovation,” Benjamin Jones, an economist at Northwestern University, and Lawrence Summers, a Harvard economist and former US Treasury secretary, calculate the effects of the nation’s total R&D spending on gross domestic product and our overall standard of living. They’re taking on the big picture, and it’s ambitious because there are so many variables. But they are able to come up with a convincing range of estimates for the returns, all of them impressive.

        On the conservative end of their estimates, says Jones, investing $1 in R&D yields about $5 in returns—defined in this case as additional GDP per person (basically, how much richer we become). Change some of the assumptions—for example, by attempting to account for the value of better medicines and improved health care, which aren’t fully captured in GDP—and you get even larger payoffs.

        While the $5 return is at the low end of their estimates, it’s still “a remarkably good investment,” Jones says. “There aren’t many where you put in $1 and get $5 back.”

        That’s the return for the nation’s overall R&D funding. But what do we get for government-funded R&D in particular? Andrew Fieldhouse, an economist at Texas A&M, and Karel Mertens at the Federal Reserve Bank of Dallas looked specifically at how changes in public R&D spending affect the total factor productivity (TFP) of businesses. A favorite metric of economists, TFP is driven by new technologies and innovative business know-how—not by adding more workers or machines—and is the main driver of the nation’s prosperity over the long term.

        The economists tracked changes in R&D spending at five major US science funding agencies over many decades to see how the shifts eventually affected private-sector productivity. They found that the government was getting a huge bang for its nondefense R&D buck.

        The benefits begin kicking in after around five to 10 years and often have a long-lasting impact on the economy. Nondefense public R&D funding has been responsible for 20% to 25% of all private-sector productivity growth in the country since World War II, according to the economists. It’s an astonishing number, given that the government invests relatively little in nondefense R&D. For example, its spending on infrastructure, another contributor to productivity growth, has been far greater over those years.

        The large impact of public R&D investments also provides insight into one of America’s most troubling economic mysteries: the slowdown in productivity growth that began in the 1970s, which has roiled the country’s politics as many people face stunted living standards and limited financial prospects. Their research, says Fieldhouse, suggests that as much as a quarter of that slowdown was caused by a decline in public R&D funding that happened roughly over the same time.

        After reaching a high of 1.86% of GDP in 1964, federal R&D spending began dropping. Starting in the early 1970s, TFP growth also began to decline, from above 2% a year in the late 1960s to somewhere around 1% since the 1970s (with the exception of a rise during the late 1990s), roughly tracking the spending declines with a lag of a few years.

        If in fact the productivity slowdown was at least partially caused by a drop in public R&D spending, it’s evidence that we would be far richer today if we had kept up a higher level of science investment. And it also flags the dangers of today’s proposed cuts. “Based on our research,” says Fieldhouse, “I think it’s unambiguously clear that if you actually slash the budget of the NIH by 40%, if you slash the NSF budget by 50%, there’s going to be a deceleration in US productivity growth over the next seven to 10 years that will be measurable.”

        Out of whack

        Though the Trump administration’s proposed 2026 budget would slash science budgets to an unusual degree, public funding of R&D has actually been in slow decline for decades. Federal funding of science is at its lowest rate in the last 70 years, accounting for only around 0.6% of GDP.

        Even as public funding has dropped, business R&D investments have steadily risen. Today businesses spend far more than the government; in 2023, companies invested about $700 billion in R&D while the US government spent $172 billion, according to data from the NSF’s statistical agency. You might think, Good—let companies do research. It’s more efficient. It’s more focused. Keep the government out of it.

        But there is a big problem with that argument. Publicly funded research, it turns out, tends to lead to relatively more productivity growth over time because it skews more toward fundamental science than the applied work typically done by companies.

        In a new working paper called “Public R&D Spillovers and Productivity Growth,” Arnaud Dyèvre, an assistant professor at of economics at HEC Paris, documents the broad and often large impacts of so-called knowledge spillovers—the benefits that flow to others from work done by the original research group. Dyèvre found that the spillovers of public-funded R&D have three times more impact on productivity growth across businesses and industries than those from private R&D funding.

        The findings are preliminary, and Dyèvre is still updating the research—much of which he did as a postdoc at MIT—but he says it does suggest that the US “is underinvesting in fundamental R&D,” which is heavily funded by the government. “I wouldn’t be able to tell you exactly which percentage of R&D in the US needs to be funded by the government or what percent needs to be funded by the private sector. We need both,” he says. But, he adds, “the empirical evidence” suggests that “we’re out of balance.”

        The big question

        Getting the balance of funding for fundamental science and applied research right is just one of the big questions that remain around R&D funding. In mid-July, Open Philanthropy and the Alfred P. Sloan Foundation, both nonprofit organizations, jointly announced that they planned to fund a five-year “pop-up journal” that would attempt to answer many of the questions still swirling around how to define and optimize the ROI of research funding.

        “There is a lot of evidence consistent with a really high return to R&D, which suggests we should do more of it,” says Matt Clancy, a senior program officer at Open Philanthropy. “But when you ask me how much more, I don’t have a good answer. And when you ask me what types of R&D should get more funding, we don’t have a good answer.”

        Pondering such questions should keep innovation economists busy for the next several years. But there is another mystifying piece of the puzzle, says Northwestern’s Jones. If the returns on R&D investments are so high—the kind that most venture capitalists or investors would gladly take—why isn’t the government spending more?

        “I think it’s unambiguously clear that if you actually slash the budget of the NIH by 40%, if you slash the NSF budget by 50%, there’s going to be a deceleration in US productivity growth over the next seven to 10 years that will be measurable.”

        Jones, who served as a senior economic advisor in the Obama administration, says discussions over R&D budgets in Washington are often “a war of anecdotes.” Science advocates cite the great breakthroughs that resulted from earlier government funding, while budget hawks point to seemingly ludicrous projects or spectacular failures. Both have plenty of ammunition. “People go back and forth,” says Jones, “and it doesn’t really lead to anywhere.”

        The policy gridlock is rooted in in the very nature of fundamental research. Today’s science will lead to great advances. And there will be countless failures; a lot of money will be wasted on fruitless experiments. The problem, of course, is that when you’re deciding to fund new projects, it’s impossible to predict which the outcome will be, even in the case of odd, seemingly silly science. Guessing just what research will or will not lead to the next great breakthrough is a fool’s errand.

        Take the cuts in the administration’s proposed fiscal 2026 budget for the NSF, a leading funder of basic science. The administration’s summary begins with the assertion that its NSF budget “is prioritizing investments that complement private-sector R&D and offer strong potential to drive economic growth and strengthen U.S. technological leadership.” So far, so good. It cites the government’s commitment to AI and quantum information science. But dig deeper and you will see the contradictions in the numbers.

        Not only is NSF’s overall budget cut by 57%, but funding for physical sciences like chemistry and materials research—fields critical to advancing AI and quantum computers—has also been blown apart. Funding for the NSF’s mathematical and physical sciences program was reduced by 67%. The directorate for computer and information science and engineering fared little better; its research funding was cut by 66%.

        There is a great deal of hope among many in the science community that Congress, when it passes the actual 2026 budget, will at least partially reverse these cuts. We’ll see. But even if it does, why attack R&D funding in the first place? It’s impossible to answer that without plunging into the messy depths of today’s chaotic politics. And it is equally hard to know whether the recent evidence gathered by academic economists on the strong returns to R&D investments will matter when it comes to partisan policymaking.

        But at least those defending the value of public funding now have a far more productive way to make their argument, rather than simply touting past breakthroughs. Even for fiscal hawks and those pronouncing concerns about budget deficits, the recent work provides a compelling and simple conclusion: More public funding for basic science is a sound investment that makes us more prosperous.

        AI-designed viruses are here and already killing bacteria

        Artificial intelligence can draw cat pictures and write emails. Now the same technology can compose a working genome.

        A research team in California says it used AI to propose new genetic codes for viruses—and managed to get several of these viruses to replicate and kill bacteria.

        The scientists, based at Stanford University and the nonprofit Arc Institute, both in Palo Alto, say the germs with AI-written DNA represent the “the first generative design of complete genomes.”

        The work, described in a preprint paper, has the potential to create new treatments and accelerate research into artificially engineered cells. It is also an “impressive first step” toward AI-designed life forms, says Jef Boeke, a biologist at NYU Langone Health, who was provided an advance copy of the paper by MIT Technology Review.  

        Boeke says the AI’s performance was surprisingly good and that its ideas were unexpected. “They saw viruses with new genes, with truncated genes, and even different gene orders and arrangements,” he says.

        This is not yet AI-designed life, however. That’s because viruses are not alive. They’re more like renegade bits of genetic code with relatively puny, simple genomes. 

        In the new work, researchers at the Arc Institute sought to develop variants of a bacteriophage—a virus that infects bacteria—called phiX174, which has only 11 genes and about 5,000 DNA letters.

        To do so, they used two versions of an AI called Evo, which works on the same principles as large language models like ChatGPT. Instead of feeding them textbooks and blog posts to learn from, the scientists trained the models on the genomes of about 2 million other bacteriophage viruses.

        But would the genomes proposed by the AI make any sense? To find out, the California researchers chemically printed 302 of the genome designs as DNA strands and then mixed those with E. coli bacteria.

        That led to a profound “AI is here” moment when, one night, the scientists saw plaques of dead bacteria in their petri dishes. They later took microscope pictures of the tiny viral particles, which look like fuzzy dots.

        “That was pretty striking, just actually seeing, like, this AI-generated sphere,” says Brian Hie, who leads the lab at the Arc Institute where the work was carried out.

        Overall, 16 of the 302 designs ended up working—that is, the computer-designed phage started to replicate, eventually bursting through the bacteria and killing them.

        J. Craig Venter, who created some of the first organisms with lab-made DNA nearly two decades ago, says the AI methods look to him like “just a faster version of trial-and-error experiments.”

        For instance, when a team he led managed to create a bacterium with a lab-printed genome in 2008, it was after a long hit-or-miss process of testing out different genes. “We did the manual AI version—combing through the literature, taking what was known,” he says. 

        But speed is exactly why people are betting AI will transform biology. The new methods already claimed a Nobel Prize in 2024 for predicting protein shapes. And investors are staking billions that AI can find new drugs. This week a Boston company, Lila, raised $235 million to build automated labs run by artificial intelligence.

        Computer-designed viruses could also find commercial uses. For instance, doctors have sometimes tried “phage therapy” to treat patients with serious bacterial infections. Similar tests are underway to cure cabbage of black rot, also caused by bacteria.

        “There is definitely a lot of potential for this technology,” says Samuel King, the student who spearheaded the project in Hei’s lab. He notes that most gene therapy uses viruses to shuttle genes into patients’ bodies, and AI might develop more effective ones.

        The Stanford researchers say they purposely haven’t taught their AI about viruses that can infect people. But this type of technology does create the risk that other scientists—out of curiosity, good intentions, or malice—could turn the methods on human pathogens, exploring new dimensions of lethality.

        “One area where I urge extreme caution is any viral enhancement research, especially when it’s random so you don’t know what you are getting,” says Venter. “If someone did this with smallpox or anthrax, I would have grave concerns.”

        Whether an AI can generate a bona fide genome for a larger organism remains an open question. For instance, E. coli has about a thousand times more DNA code than phiX174 does. “The complexity would rocket from staggering to … way way more than the number of subatomic particles in the universe,” says Boeke.

        Also, there’s still no easy way to test AI designs for larger genomes. While some viruses can “boot up” from just a DNA strand, that’s not the case with a bacterium, a mammoth, or a human. Scientists would instead have to gradually change an existing cell with genetic engineering—a still laborious process.

        Despite that, Jason Kelly, the CEO of Ginkgo Bioworks, a cell-engineering company in Boston, says exactly such an effort is needed. He believes it could be carried out in “automated” laboratories where genomes get proposed and tested and the results are fed back to AI for further improvement.

         “This would be a nation-scale scientific milestone, as cells are the building blocks of all life,” says Kelly. “The US should make sure we get to it first.”