Behind Microsoft CEO Satya Nadella’s push to get AI tools in developers’ hands

In San Francisco last week, everyone’s favorite surprise visitor was Microsoft CEO Satya Nadella. 

At OpenAI’s DevDay—the company’s first-ever event for developers building on its platform—Nadella bounded on stage to join OpenAI CEO Sam Altman, blowing the hair back on an already electrified audience. “You guys have built something magic,” he gushed. 

Two days later on another stage, in another venue, at another developers’ conference, Nadella made his second unannounced appearance of the week—this time at GitHub Universe. There Thomas Dohmke, GitHub’s CEO, was showing off a new version of the company’s AI programming tool, Copilot, that can generate computer code from natural language. Nadella was effusive: “I can code again!” he exclaimed. 

Today, Nadella will be onstage speaking to developers at Microsoft Ignite, where the company is announcing even more AI-based developer tools, including an Azure AI Studio that will let devs choose between model catalogs from not only Microsoft, but also the likes of Meta, OpenAI, and Hugging Face, as well as new tools for customizing Copilot for Microsoft 365. 

If it seems like Nadella is obsessed with developers, you’re not wrong. He’s making the rounds to tout all the ways they can use a new generation of AI-powered tools, like GitHub Copilot (Microsoft acquired GitHub in 2018) or the new suite of developer tools from OpenAI, a company in which Microsoft has reportedly invested some $13 billion.

Last week, Nadella took a 20-minute break from all of his onstage appearances to sit down with MIT Technology Review to talk about (you guessed it) developers. He repeatedly emphasized Microsoft’s longstanding focus on developers. But he also had a message: The way we create software is fundamentally changing. 

Nadella believes a platform shift is underway, one that will prove just as significant as the shifts from mainframe to desktop or desktop to mobile. This time, the transition is to natural language AI tools, some of which he argues will lower the barrier to entry for software development, make existing developers more productive, and ultimately lead to a new era of creativity. 

We present Nadella in his own words, below. His remarks have been edited and condensed somewhat for readability.  

ON THE RELATIONSHIP WITH OPENAI

One criticism of OpenAI is that its very business is only possible via Microsoft, which has given the startup billions of dollars and access to the resources it needs to power its computing-intensive language model. Yet Microsoft is also highly dependent on OpenAI’s technology to power services like GitHub Copilot, Bing, and Office 365. Altman even joked about the partnership onstage. We asked Nadella about this relationship.   

I’ve always felt that Microsoft is a platform-and-partner-first company, and this is not new to us. And so therefore, we both are effectively codependent, right? They depend on us to build the best systems, we depend on them to build the best models, and we go to market together. 

ON HIS MISSION TO GET IN FRONT OF DEVELOPERS

Nadella says this platform shift is different enough from previous ones that he feels the company needs to provide developers not only with tools, but also with a clear message about what it’s thinking and how devs can come along. 

Whenever you have a platform shift, the key thing is to make sure the platform is ubiquitously available for developers to build all kinds of new things. So to us, the most important task is to make the developer tools, the developer platforms, broadly available. 

The second thing is for us to also show the light, right? Whether it’s OpenAI building ChatGPT and then innovating on top of it, or us building Copilot and innovating on it. That will give developers an opportunity to distribute their applications. So the most important thing in any platform creation is to get the platform ubiquitously available, and then help developers reach [their] audience. 

Those are the two goals that we have across all of these [conferences].

ON WHAT IS DIFFERENT ABOUT THIS SHIFT AND PRODUCTIVITY

Productivity gains in the United States have been sluggish for the past 15 or more years. The last huge platform shift—the rise of mobile development—did little to achieve widespread prosperity. Nadella says this time will be different, largely because the shift to AI will fuel a creative revolution by making it easy for anyone to generate new work, including code. 

On the other hand, coding today is a highly skilled, well-paid job, and there’s some concern that AI could effectively automate it. Nadella argues that skilled programmers will remain in demand, but that their jobs will change and even more jobs will become available. Nadella has said he envisions 1 billion developers creating on its platforms, many of them with little to no previous experience with coding.   

Anytime you have something as disruptive as this, you have to think about the displacement and causes. And that means it’s all about upskilling and reskilling, and in an interesting way, it’s more akin to what happened when word processors and spreadsheets started showing up. Obviously, if you were a typist, it really drastically changed. But at the same time, it enabled a billion people to be able to type into word processors and create and share documents.

I don’t think professional developers are going to be any less valuable than they are today. It’s just that we’re going to have many, many gradations of developers. Each time you’re prompting a Bing chat or ChatGPT, you’re essentially programming. The conversation itself is steering a model.

I think there will be many, many new jobs, there will be many, many new types of knowledge work, or frontline work, where the drudgery is removed.

I think the mobile era was fantastic. It made ubiquitous consumption of services. It didn’t translate into ubiquitous creation of services.

The last time there was a broad spread of productivity in the United States and beyond because of information technology was the [advent of the] PC. In fact, even the critics of information technology and productivity, like Robert Gordon of Northwestern, acknowledged that the PC, when it first showed up at work, did actually translate to broad productivity stats changes.

So that’s where I think this is, where these tools, like Copilot, being used by a [beginner] software engineer in Detroit, in order to be able to write [code].… I think we’ll have a real change in the productivity of the auto industry. Same thing in retail, same thing in frontline work and knowledge work.

The barrier to entry is very low. Because it’s natural language, domain experts can build apps or workflows. That, I think, is what’s the most exciting thing about this. This is not about just a consumption-led thing. This is not about elite creation. This is about democratized creation. I’m very, very hopeful that we’ll start seeing the productivity gains much more broadly.

ON PROTECTING DEVELOPERS

Numerous intellectual property cases and class action lawsuits are before the US courts over issues of fair use. At least one singles out GitHub Copilot specifically, claiming Microsoft and OpenAI’s generative tools, which are trained on open source code, amount to software piracy. There’s a fear that people who use these tools could be subject to intellectual property claims themselves. Microsoft is trying to address these issues with a broad indemnification policy. OpenAI also announced its own indemnification policy, Copyright Shield, at its DevDay conference. 

Fundamentally these large models crawl and get content and then train on that content, right? If anybody doesn’t want their content to be crawled, we have great granular controls in our crawlers that allow anybody to stop it from crawling. In fact, we have controls where you can have it crawl just for search, but not for large language model training. That’s available today. So anybody who really wants to ensure that their content is not being taken for retraining can do so today. 

The second thing, of course, is I think the courts and the legislative process in some combination will have to decide what is fair use and what is not fair use.

We have taken a lot of control in making sure that we are only training models, and we are using data to train models that we’re allowed to and which we believe we have a legal standing on. 

If it comes to it, we’ll litigate it in the courts. We’ll take that burden on so the users of our products don’t have to worry about it. That’s as simple as that, which is to take the liability and transfer it from our users to us. And of course, we are going to be very, very mindful of making sure we’re on the right side of the law there.

Google DeepMind’s weather AI can forecast extreme weather faster and more accurately

This year the Earth has been hit by a record number of unpredictable extreme weather events made worse by climate change. Predicting them faster and with greater accuracy could enable us to prepare better for natural disasters and help save lives. A new AI model from Google DeepMind could make that easier. 

In research published in Science today, Google DeepMind’s model, GraphCast, was able to predict weather conditions up to 10 days in advance, more accurately and much faster than the current gold standard. GraphCast outperformed the model from the European Centre for Medium-Range Weather Forecasts (ECMWF) in more than 90% of over 1,300 test areas. And on predictions for Earth’s troposphere—the lowest part of the atmosphere, where most weather happens—GraphCast outperformed the ECMWF’s model on more than 99% of weather variables, such as rain and air temperature 

Crucially, GraphCast can also offer meteorologists accurate warnings, much earlier than standard models, of conditions such as extreme temperatures and the paths of cyclones. In September, GraphCast accurately predicted that Hurricane Lee would make landfall in Nova Scotia nine days in advance, says Rémi Lam, a staff research scientist at Google DeepMind. Traditional weather forecasting models pinpointed the hurricane to Nova Scotia only six days in advance.

Weather prediction is one of the most challenging problems that humanity has been working on for a long, long time. And if you look at what has happened in the last few years with climate change, this is an incredibly important problem,” says Pushmeet Kohli, the vice president of research at Google DeepMind.  

Traditionally, meteorologists use massive computer simulations to make weather predictions. They are very energy intensive and  time consuming to run, because the simulations take into account many physics-based equations and different weather variables such as temperature, precipitation, pressure, wind, humidity, and cloudiness, one by one. 

GraphCast uses machine learning to do these calculations in under a minute. Instead of using the physics-based equations, it bases its predictions on four decades of historical weather data. GraphCast uses graph neural networks, which map Earth’s surface into more than a million grid points. At each grid point, the model predicts the temperature, wind speed and direction, and mean sea-level pressure, as well as other conditions like humidity. The neural network is then able to find patterns and draw conclusions about what will happen next for each of these data points. 

For the past year, weather forecasting has been going through a revolution as models such as GraphCast, Huawei’s Pangu-Weather and Nvidia’s FourcastNet have made meteorologists rethink the role AI can play in weather forecasting. GraphCast improves on the performance of other competing models, such as Pangu-Weather, and is able to predict more weather variables, says Lam. The ECMWF is already using it.

When Google DeepMind first debuted GraphCast last December, it felt like Christmas, says Peter Dueben, head of Earth system modeling at ECMWF, who was not involved in the research. 

“It showed that these models are so good that we cannot avoid them anymore,” he says. 

GraphCast is a “reckoning moment” for weather prediction because it shows that predictions can be made using historical data, says Aditya Grover, an assistant professor of computer science at UCLA, who developed ClimaX, a foundation model that allows researchers to do different tasks relating to modeling the Earth’s weather and climate. 

DeepMind’s model is “great work and extremely exciting,” says Oliver Fuhrer, the head of the numerical prediction department at MeteoSwiss, the Swiss Federal Office of Meteorology and Climatology. Fuhrer says that other weather agencies, such as the ECMWF and the Swedish Meteorological and Hydrological Institute, have also used the graph neural network architecture proposed by Google DeepMind to build their own models. 

But GraphCast is not perfect. It still lags behind conventional weather forecasting models in some areas, such as precipitation, Dueben says. Meteorologists will still have to use conventional models alongside machine-learning models to offer better predictions. 

Google DeepMind is also making GraphCast open source. This is a good development, says UCLA’s Grover. 

“With climate change on the rise, it’s very important that big organizations, which have had the luxury of so much compute, also think about giving back [to the scientific community],” he says. 

Google DeepMind’s weather AI can forecast extreme weather faster and more accurately

This year the Earth has been hit by a record number of unpredictable extreme weather events made worse by climate change. Predicting them faster and with greater accuracy could enable us to prepare better for natural disasters and help save lives. A new AI model from Google DeepMind could make that easier. 

In research published in Science today, Google DeepMind’s model, GraphCast, was able to predict weather conditions up to 10 days in advance, more accurately and much faster than the current gold standard. GraphCast outperformed the model from the European Centre for Medium-Range Weather Forecasts (ECMWF) in more than 90% of over 1,300 test areas. And on predictions for Earth’s troposphere—the lowest part of the atmosphere, where most weather happens—GraphCast outperformed the ECMWF’s model on more than 99% of weather variables, such as rain and air temperature 

Crucially, GraphCast can also offer meteorologists accurate warnings, much earlier than standard models, of conditions such as extreme temperatures and the paths of cyclones. In September, GraphCast accurately predicted that Hurricane Lee would make landfall in Nova Scotia nine days in advance, says Rémi Lam, a staff research scientist at Google DeepMind. Traditional weather forecasting models pinpointed the hurricane to Nova Scotia only six days in advance.

Weather prediction is one of the most challenging problems that humanity has been working on for a long, long time. And if you look at what has happened in the last few years with climate change, this is an incredibly important problem,” says Pushmeet Kohli, the vice president of research at Google DeepMind.  

Traditionally, meteorologists use massive computer simulations to make weather predictions. They are very energy intensive and  time consuming to run, because the simulations take into account many physics-based equations and different weather variables such as temperature, precipitation, pressure, wind, humidity, and cloudiness, one by one. 

GraphCast uses machine learning to do these calculations in under a minute. Instead of using the physics-based equations, it bases its predictions on four decades of historical weather data. GraphCast uses graph neural networks, which map Earth’s surface into more than a million grid points. At each grid point, the model predicts the temperature, wind speed and direction, and mean sea-level pressure, as well as other conditions like humidity. The neural network is then able to find patterns and draw conclusions about what will happen next for each of these data points. 

For the past year, weather forecasting has been going through a revolution as models such as GraphCast, Huawei’s Pangu-Weather and Nvidia’s FourcastNet have made meteorologists rethink the role AI can play in weather forecasting. GraphCast improves on the performance of other competing models, such as Pangu-Weather, and is able to predict more weather variables, says Lam. The ECMWF is already using it.

When Google DeepMind first debuted GraphCast last December, it felt like Christmas, says Peter Dueben, head of Earth system modeling at ECMWF, who was not involved in the research. 

“It showed that these models are so good that we cannot avoid them anymore,” he says. 

GraphCast is a “reckoning moment” for weather prediction because it shows that predictions can be made using historical data, says Aditya Grover, an assistant professor of computer science at UCLA, who developed ClimaX, a foundation model that allows researchers to do different tasks relating to modeling the Earth’s weather and climate. 

DeepMind’s model is “great work and extremely exciting,” says Oliver Fuhrer, the head of the numerical prediction department at MeteoSwiss, the Swiss Federal Office of Meteorology and Climatology. Fuhrer says that other weather agencies, such as the ECMWF and the Swedish Meteorological and Hydrological Institute, have also used the graph neural network architecture proposed by Google DeepMind to build their own models. 

But GraphCast is not perfect. It still lags behind conventional weather forecasting models in some areas, such as precipitation, Dueben says. Meteorologists will still have to use conventional models alongside machine-learning models to offer better predictions. 

Google DeepMind is also making GraphCast open source. This is a good development, says UCLA’s Grover. 

“With climate change on the rise, it’s very important that big organizations, which have had the luxury of so much compute, also think about giving back [to the scientific community],” he says. 

Noise-canceling headphones could let you pick and choose the sounds you want to hear

This is a subscriber-only story.

Future noise-canceling headphones could let users opt back in to certain sounds they’d like to hear, such as babies crying, birds tweeting, or alarms ringing.

The technology that makes it possible, called semantic hearing, could pave the way for smarter hearing aids and earphones, allowing the wearer to filter out some sounds while boosting others. 

The system, which is still in prototype, works by connecting off-the-shelf noise-canceling headphones to a smartphone app. The microphones embedded in these headphones, which are used to cancel out noise, are repurposed to also detect the sounds in the world around the wearer. These sounds are then played back to a neural network, which is running on the smartphone; then certain sounds are boosted or suppressed in real time, depending on the user’s preferences. It was developed by researchers from the University of Washington, who presented the research at the ACM Symposium on User Interface Software and Technology (UIST) last week.

The team trained the network on thousands of audio samples from online data sets and sounds collected from various noisy environments. Then they taught it to recognize 20 everyday sounds, such as a thunderstorm, a toilet flushing, or glass breaking.

It was tested on nine participants, who wandered around offices, parks, and streets. The researchers found that their system performed well at muffling and boosting sounds, even in situations it hadn’t been trained for. However, it struggled slightly at separating human speech from background music, especially rap music.

Mimicking human ability

Researchers have long tried to solve the “cocktail party problem”—that is, to get a computer to focus on a single voice in a crowded room, as humans are able to do. This new method represents a significant step forward and demonstrates the technology’s potential, says Marc Delcroix, a senior research scientist at NTT Communication Science Laboratories, Kyoto, who studies speech enhancement and recognition and was not involved in the project. 

“This kind of achievement is very helpful for the field,” he says. “Similar ideas have been around, especially in the field of speech separation, but they are the first to propose a complete real-time binaural target sound extraction system.”

“Noise-canceling headsets today have this capability where you can still play music even when the noise canceling is turned on,” says Shyam Gollakota, an assistant professor at the University of Washington, who worked on the project. “Instead of playing music, we are playing back the actual sounds of interest from the environment, which we extracted from our machine-learning algorithms.”

Gollakota is excited by the technology’s potential for helping people with hearing loss, as hearing aids can be of limited use in noisy environments. “It’s a unique opportunity to create the future of intelligent hearables through enhanced hearing,” he says.

The ability to be more selective about what we can and can’t hear could also benefit people who require focused listening for their job, such as health-care, military, and engineering professionals, or for factory or construction workers who want to protect their hearing while still being able to communicate.

Filtering out the world

This type of system could for the first time give us a degree of control over the sounds that surround us—for better or worse, says Mack Hagood, an associate professor of media and communication at Miami University in Ohio, and author of Hush: Media and Sonic Self-Control, who did not work on the project.

“This is the dream—I’ve seen people fantasizing about this for a long time,” he says. “We’re basically getting to tick a box whether we want to hear those sounds or not, and there could be times where this narrowing of experience is really beneficial—something we really should do that might actually help promote better communication.”

But whenever we opt for control and choice, we’re pushing aside serendipity and happy accidents, he says. “We’re deciding in advance what we do and don’t want to hear,” he adds. “And that doesn’t give us the opportunity to know whether we really would have enjoyed hearing something.”

Bridging the expectation-reality gap in machine learning

Machine learning (ML) is now mission critical in every industry. Business leaders are urging their technical teams to accelerate ML adoption across the enterprise to fuel innovation and long-term growth. But there is a disconnect between business leaders’ expectations for wide-scale ML deployment and the reality of what engineers and data scientists can actually build and deliver on time and at scale.

In a Forrester study launched today and commissioned by Capital One, the majority of business leaders expressed excitement at deploying ML across the enterprise, but data scientist team members said they didn’t yet have all the necessary tools to develop ML solutions at scale. Business leaders would love to leverage ML as a plug-and-play opportunity: “just input data into a black box and valuable learnings emerge.” The engineers who wrangle company data to build ML models know it’s far more complex than that. Data may be unstructured or poor quality, and there are compliance, regulatory, and security parameters to meet.

There is no quick-fix to closing this expectation-reality gap, but the first step is to foster honest dialogue between teams. Then, business leaders can begin to democratize ML across the organization. Democratization means both technical and non-technical teams have access to powerful ML tools and are supported with continuous learning and training. Non-technical teams get user-friendly data visualization tools to improve their business decision-making, while data scientists get access to the robust development platforms and cloud infrastructure they need to efficiently build ML applications. At Capital One, we’ve used these democratization strategies to scale ML across our entire company of more than 50,000 associates.

When everyone has a stake in using ML to help the company succeed, the disconnect between business and technical teams fades. So what can companies do to begin democratizing ML? Here are several best practices to bring the power of ML to everyone in the organization.

Enable your creators

The best engineers today aren’t just technical whizzes, but also creative thinkers and vital partners to product specialists and designers. To foster greater collaboration, companies should provide opportunities for tech, product, and design to work together toward shared goals. According to the Forrester study, because ML use can be siloed, focusing on collaboration can be a key cultural component of success. It will also ensure that products are built from a business, human, and technical perspective. 

Leaders should also ask engineers and data scientists what tools they need to be successful to accelerate delivery of ML solutions to the business. According to Forrester, 67% of respondents agree that a lack of easy-to-use tools is slowing down cross-enterprise adoption of ML. These tools should be compatible with an underlying tech infrastructure that supports ML engineering. Don’t make your developers live in a “hurry up and wait” world where they develop a ML model in the sandbox staging area, but then must wait to deploy it because they don’t have the compute and infrastructure to put the model into production. A robust cloud-native multitenant infrastructure that supports ML training environments is critical.

Empower your employees

Putting the power of ML into the hands of every employee, whether they’re a marketing associate or business analyst, can turn any company into a data-driven organization. Companies can start by granting employees governed access to data. Then, offer teams no-code/low-code tools to analyze data for business decisioning. It goes without saying these tools should be developed with human-centered design, so they are easy to use. Ideally, a business analyst could upload a data set, apply ML functionality through a clickable interface, and quickly generate actionable outputs.

Many employees are eager to learn more about technology. Leaders should provide teams across the enterprise with many ways to learn new skills. At Capital One, we have found success with multiple technical upskilling programs, including our Tech College that offers courses in seven technology disciplines that align to our business imperatives; our Machine Learning Engineering Program that teaches the skills necessary to jumpstart a career in ML and AI; and the Capital One Developer Academy for recent college graduates with non-computer science degrees preparing for careers in software engineering. In the Forrester study, 64% of respondents agreed that lack of training was slowing the adoption of ML in their organizations. Thankfully, upskilling is something every company can offer by encouraging seasoned associates to mentor younger talent.

Measure and celebrate success

Democratizing ML is a powerful way to spread data-driven decision-making throughout the organization. But don’t forget to measure the success of democratization initiatives and continually improve areas that need work. To quantify the success of ML democratization, leaders can analyze which data-driven decisions made through the platforms delivered measurable business results, such as new customers or additional revenue. For example, at Capital One, we have measured the amount of money customers have saved with card fraud defense enabled by our ML innovations around anomaly and change point detection.

The success of any ML democratization program is built on collaborative teamwork and measurable accountability. Business users of ML tools can provide feedback to technical teams on what functionality would help them do their jobs better. Technical teams can share the challenges they face in building future product iterations and ask for training and tools to help them succeed.

When business leaders and technical teams coalesce around a unified, human-centered vision for ML, that ultimately benefits end-customers. A company can translate data-driven learnings into better products and services that delight their customers. Deploying a few best practices to democratize ML across the enterprise will go a long way toward building a future-forward organization that innovates with powerful data insights.

This content was produced by Capital One. It was not written by MIT Technology Review’s editorial staff.