Mustafa Suleyman: AI development won’t hit a wall anytime soon—here’s why

We evolved for a linear world. If you walk for an hour, you cover a certain distance. Walk for two hours and you cover double that distance. This intuition served us well on the savannah. But it catastrophically fails when confronting AI and the core exponential trends at its heart.

From the time I began work on AI in 2010 to now, the amount of training data that goes into frontier AI models has grown by a staggering 1 trillion times—from roughly 10¹⁴ flops (floating-point operations‚ the core unit of computation) for early systems to over 10²⁶ flops for today’s largest models. This is an explosion. Everything else in AI follows from this fact.

The skeptics keep predicting walls. And they keep being wrong in the face of this epic generational compute ramp. Often, they point out that Moore’s Law is slowing. They also mention a lack of data, or they cite limitations on energy.

But when you look at the combined forces driving this revolution, the exponential trend seems quite predictable. To understand why, it’s worth looking at the complex and fast-moving reality beneath the headlines.

Think of AI training as a room full of people working calculators. For years, adding computational power meant adding more people with calculators to that room. Much of the time those workers sat idle, drumming their fingers on desks, waiting for the numbers to come through for their next calculation. Every pause was wasted potential. Today’s revolution goes beyond more and better calculators (although it delivers those); it is actually about ensuring that all those calculators never stop, and that they work together as one.

Three advances are now converging to enable this. First, the basic calculators got faster. Nvidia’s chips have delivered an over sevenfold increase in raw performance in just six years, from 312 teraflops in 2020 to 2,250 teraflops today. Our own Maia 200 chip, launched this January, delivers 30% better performance per dollar than any other hardware in our fleet. Second, the numbers arrive faster thanks to a technology called HBM, or high bandwidth memory, which stacks chips vertically like tiny skyscrapers; the latest generation, HBM3, triples the bandwidth of its predecessor, feeding data to processors fast enough to keep them busy all the time. Third, the room of people with calculators became an office and then a whole campus or city. Technologies like NVLink and InfiniBand connect hundreds of thousands of GPUs into warehouse-size supercomputers that function as single cognitive entities. A few years ago this was impossible.

These gains all come together to deliver dramatically more compute. Where training a language model took 167 minutes on eight GPUs in 2020, it now takes under four minutes on equivalent modern hardware. To put this in perspective: Moore’s Law would predict only about a 5x improvement over this period. We saw 50x. We’ve gone from two GPUs training AlexNet, the image recognition model that kicked off the modern boom in deep learning in 2012, to over 100,000 GPUs in today’s largest clusters, each one individually far more powerful than its predecessors.

Then there’s the revolution in software. Research from Epoch AI suggests that the compute required to reach a fixed performance level halves approximately every eight months, much faster than the traditional 18-to-24-month doubling of Moore’s Law. The costs of serving some recent models have collapsed by a factor of up to 900 on an annualized basis. AI is becoming radically cheaper to deploy.

The numbers for the near future are just as staggering. Consider that leading labs are growing capacity at nearly 4x annually. Since 2020, the compute used to train frontier models has grown 5x every year. Global AI-relevant compute is forecast to hit 100 million H100-equivalents by 2027, a tenfold increase in three years. Put all this together and we’re looking at something like another 1,000x in effective compute by the end of 2028. It’s plausible that by 2030 we’ll bring an additional 200 gigawatts of compute online every year—akin to the peak energy use of the UK, France, Germany, and Italy put together.

What does all this get us? I believe it will drive the transition from chatbots to nearly human-level agents—semiautonomous systems capable of writing code for days, carrying out weeks- and months-long projects, making calls, negotiating contracts, managing logistics. Forget basic assistants that answer questions. Think teams of AI workers that deliberate, collaborate, and execute. Right now we’re only in the foothills of this transition, and the implications stretch far beyond tech. Every industry built on cognitive work will be transformed.

The obvious constraint here is energy. A single refrigerator-size AI rack consumes 120 kilowatts, equivalent to 100 homes. But this hunger collides with another exponential: Solar costs have fallen by a factor of nearly 100 over 50 years; battery prices have dropped 97% over three decades. There is a pathway to clean scaling coming into view.

The capital is deployed. The engineering is delivering. The $100 billion clusters, the 10-gigawatt power draws, the warehouse-scale supercomputers … these are no longer science fiction. Ground is being broken for these projects now across the US and the world. As a result, we are heading toward true cognitive abundance. At Microsoft AI, this is the world our superintelligence lab is planning for and building.

Skeptics accustomed to a linear world will continue predicting diminishing returns. They will continue being surprised. The compute explosion is the technological story of our time, full stop. And it is still only just beginning.

Mustafa Suleyman is CEO of Microsoft AI.

Roundtables: The Next Era of Space Exploration

Listen to the session or watch below

Whether it’s the race to find life on Mars, the campaign to outsmart killer asteroids, or the quest to make the moon a permanent home to astronauts, scientists’ efforts in space can tell us more about where humanity is headed. This subscriber-only discussion examines the progress and possibilities ahead.

Speakers: Amanda Silverman, features & investigations editor, and Robin George Andrews, award-winning science journalist and author

Recorded on March 25, 2026

Related Stories:

Exclusive eBook: Are we ready to hand AI agents the keys?

We’re starting to give AI agents real autonomy, but are we prepared for what could happen next?

This subscriber-only eBook explores this and angles from experts, such as “If we continue on the current path … we are basically playing Russian roulette with humanity.”

by Grace Huckins June 12, 2025

Related Stories:

Access all subscriber-only eBooks:

Roundtables: Why 2026 Is the Year for Sodium-Ion Batteries

Listen to the session or watch below

Sodium-based batteries could be a cheaper, safer alternative to lithium-ion, and the technology is finally making its way into cars—and energy storage arrays on the grid. Sodium-ion batteries are one of MIT Technology Review’s 10 Breakthrough Technologies of 2026 list, and this subscriber-only discussion explains why.

Watch a discussion exploring the present moment for sodium-ion batteries—and what’s coming next.

Speakers: Mary Beth Griggs, Science Editor; Casey Crownhart, Senior Climate Reporter; and Caiwei Chen, China Reporter

Recorded on February 25, 2026

Related Stories:

Exclusive eBook: The great Al hype correction of 2025

2025 was a year of reckoning, including how the heads of the top AI companies made promises they couldn’t keep. In this exclusive subscriber-only eBook, you’ll learn more about why we may need to readjust our expectations.

This story is part of the Hype Correction package.

by Will Douglas Heaven December 15, 2025

Table of Contents:

  • LLMs are not everything
  • AI is not a quick fix to all your problems
  • Are we in a bubble? (If so, what kind of bubble?)
  • ChatGPT was not the beginning, and it won’t be the end

Related Stories:

Access all subscriber-only eBooks:

Roundtables: Why AI Companies Are Betting on Next-Gen Nuclear

AI is driving unprecedented investment for massive data centers and an energy supply that can support its huge computational appetite. One potential source of electricity for these facilities is next-generation nuclear power plants, which could be cheaper to construct and safer to operate than their predecessors.

Watch a discussion with our editors and reporters on hyperscale AI data centers and next-gen nuclear—two featured technologies on the MIT Technology Review 10 Breakthrough Technologies of 2026 list.

Speakers: Amy Nordrum, Executive Editor, Operations; Casey Crownhart, Senior Climate Reporter; and Mat Honan
Editor in Chief

Recorded on January 28, 2026

Related Stories:

Exclusive eBook: How AGI Became a Consequential Conspiracy Theory

In this exclusive subscriber-only eBook, you’ll learn about how the idea that machines will be as smart as—or smarter than—humans has hijacked an entire industry.

by Will Douglas Heaven October 30, 2025

Table of Contents:

  • How Silicon Valley got AGI-pilled
  • The great AGI conspiracy
  • How AGI hijacked an industry
  • The great AGI conspiracy, concluded

Related Stories:

Access all subscriber-only eBooks:

Exclusive eBook: Aging Clocks & Understanding Why We Age

In this exclusive subscriber-only eBook, you’ll learn about a new method that scientists have uncovered to look at the ways our bodies are aging.

by  Jessica Hamzelou October 14, 2025

Table of Contents:

  • Clocks kick off
  • Black-box clocks
  • How to be young again
  • Dogs and dolphins
  • When young meets old

Related Stories:

Access all subscriber-only eBooks:

Roundtables: Surviving the New Age of Conspiracies

Everything is a conspiracy theory now. MIT Technology Review’s series, “The New Conspiracy Age,” explores how this moment is changing science and technology. Watch a discussion with our editors and Mike Rothschild, journalist and conspiracy theory expert, about how we can make sense of them all.

Speakers: Amanda Silverman, Editor, Features & Investigations; Niall Firth, Executive Editor, Newsroom; and Mike Rothschild, Journalist & Conspiracy Theory Expert.

Recorded on November 20, 2025

Related Content:

Exclusive eBook: The Math on AI’s Energy Footprint

In this exclusive subscirber-only ebook you’ll learn how the emissions from individual AI text, image, and video queries seem small—until you add up what the industry isn’t tracking and consider where it’s heading next.

by James O’Donnell and Casey Crownhart May 20, 2025

Table of contents

  • Part One: Making the model
  • Part Two: A Query
  • Part Three: Fuel and emissions
  • Part Four: The future ahead

Related Content: