Deploying high-performance, energy-efficient AI

Although AI is by no means a new technology there have been massive and rapid investments in it and large language models. However, the high-performance computing that powers these rapidly growing AI tools — and enables record automation and operational efficiency — also consumes a staggering amount of energy. With the proliferation of AI comes the responsibility to deploy that AI responsibly and with an eye to sustainability during hardware and software R&D as well as within data centers.

“Enterprises need to be very aware of the energy consumption of their digital technologies, how big it is, and how their decisions are affecting it,” says corporate vice president and general manager of data center platform engineering and architecture at Intel, Zane Ball.

One of the key drivers of a more sustainable AI is modularity, says Ball. Modularity breaks down subsystems of a server into standard building blocks, defining interfaces between those blocks so they can work together. This system can reduce the amount of embodied carbon in a server’s hardware components and allows for components of the overall ecosystem to be reused, subsequently reducing R&D investments.

Downsizing infrastructure within data centers, hardware, and software can also help enterprises reach greater energy efficiency without compromising function or performance. While very large AI models require megawatts of super compute power, smaller, fine-tuned models that operate within a specific knowledge domain can maintain high performance but low energy consumption.

“You give up that kind of amazing general purpose use like when you’re using ChatGPT-4 and you can ask it everything from 17th century Italian poetry to quantum mechanics, if you narrow your range, these smaller models can give you equivalent or better kind of capability, but at a tiny fraction of the energy consumption,” says Ball.

The opportunities for greater energy efficiency within AI deployment will only expand over the next three to five years. Ball forecasts significant hardware optimization strides, the rise of AI factories — facilities that train AI models on a large scale while modulating energy consumption based on its availability — as well as the continued growth of liquid cooling, driven by the need to cool the next generation of powerful AI innovations.

“I think making those solutions available to our customers is starting to open people’s eyes how energy efficient you can be while not really giving up a whole lot in terms of the AI use case that you’re looking for.”

This episode of Business Lab is produced in partnership with Intel.

Full Transcript

Laurel Ruma: From MIT Technology Review, I’m Laurel Ruma and this is Business Lab, the show that helps business leaders make sense of new technologies coming out of the lab and into the marketplace.

Our topic is building a better AI architecture. Going green isn’t for the faint of heart, but it’s also a pressing need for many, if not all enterprises. AI provides many opportunities for enterprises to make better decisions, so how can it also help them be greener?

Two words for you: sustainable AI.

My guest is Zane Ball, corporate vice president and general manager of data center platform engineering and architecture at Intel.

This podcast is produced in partnership with Intel.

Welcome Zane.

Zane Ball: Good morning.

Laurel: So to set the stage for our conversation, let’s start off with the big topic. As AI transforms businesses across industries, it brings the benefits of automation and operational efficiency, but that high-performance computing also consumes more energy. Could you give an overview of the current state of AI infrastructure and sustainability at the large enterprise level?

Zane: Absolutely. I think it helps to just kind of really zoom out big picture, and if you look at the history of IT services maybe in the last 15 years or so, obviously computing has been expanding at a very fast pace. And the good news about that history of the last 15 years or so, is while computing has been expanding fast, we’ve been able to contain the growth in energy consumption overall. There was a great study a couple of years ago in Science Magazine that talked about how compute had grown by maybe 550% over a decade, but that we had just increased electricity consumption by a few percent. So those kind of efficiency gains were really profound. So I think the way to kind of think about it is computing’s been expanding rapidly, and that of course creates all kinds of benefits in society, many of which reduce carbon emissions elsewhere.

But we’ve been able to do that without growing electricity consumption all that much. And that’s kind of been possible because of things like Moore’s Law, Big Silicon has been improving with every couple of years and make devices smaller, they consume less power, things get more efficient. That’s part of the story. Another big part of this story is the advent of these hyperscale data centers. So really, really large-scale computing facilities, finding all kinds of economies of scale and efficiencies, high utilization of hardware, not a lot of idle hardware sitting around. That also was a very meaningful energy efficiency. And then finally this development of virtualization, which allowed even more efficient utilization of hardware. So those three things together allowed us to kind of accomplish something really remarkable. And during that time, we also had AI starting to play, I think since about 2015, AI workloads started to play a pretty significant role in digital services of all kinds.

But then just about a year ago, ChatGPT happens and we have a non-linear shift in the environment and suddenly large language models, probably not news to anyone on this listening to this podcast, has pivoted to the center and there’s just a breakneck investment across the industry to build very, very fast. And what is also driving that is that not only is everyone rushing to take advantage of this amazing large language model kind of technology, but that technology itself is evolving very quickly. And in fact also quite well known, these models are growing in size at a rate of about 10x per year. So the amount of compute required is really sort of staggering. And when you think of all the digital services in the world now being infused with AI use cases with very large models, and then those models themselves growing 10x per year, we’re looking at something that’s not very similar to that last decade where our efficiency gains and our greater consumption were almost penciling out.

Now we’re looking at something I think that’s not going to pencil out. And we’re really facing a really significant growth in energy consumption in these digital services. And I think that’s concerning. And I think that means that we’ve got to take some strong actions across the industry to get on top of this. And I think just the very availability of electricity at this scale is going to be a key driver. But of course many companies have net-zero goals. And I think as we pivot into some of these AI use cases, we’ve got work to do to square all of that together.

Laurel: Yeah, as you mentioned, the challenges are trying to develop sustainable AI and making data centers more energy efficient. So could you describe what modularity is and how a modularity ecosystem can power a more sustainable AI?

Zane: Yes, I think over the last three or four years, there’ve been a number of initiatives. Intel’s played a big part of this as well of re-imagining how servers are engineered into modular components. And really modularity for servers is just exactly as it sounds. We break different subsystems of the server down into some standard building blocks, define some interfaces between those standard building blocks so that they can work together. And that has a number of advantages. Number one, from a sustainability point of view, it lowers the embodied carbon of those hardware components. Some of these hardware components are quite complex and very energy intensive to manufacture. So imagine a 30 layer circuit board, for example, is a pretty carbon intensive piece of hardware. I don’t want the entire system, if only a small part of it needs that kind of complexity. I can just pay the price of the complexity where I need it.

And by being intelligent about how we break up the design in different pieces, we bring that embodied carbon footprint down. The reuse of pieces also becomes possible. So when we upgrade a system, maybe to a new telemetry approach or a new security technology, there’s just a small circuit board that has to be replaced versus replacing the whole system. Or maybe a new microprocessor comes out and the processor module can be replaced without investing in new power supplies, new chassis, new everything. And so that circularity and reuse becomes a significant opportunity. And so that embodied carbon aspect, which is about 10% of carbon footprint in these data centers can be significantly improved. And another benefit of the modularity, aside from the sustainability, is it just brings R&D investment down. So if I’m going to develop a hundred different kinds of servers, if I can build those servers based on the very same building blocks just configured differently, I’m going to have to invest less money, less time. And that is a real driver of the move towards modularity as well.

Laurel: So what are some of those techniques and technologies like liquid cooling and ultrahigh dense compute that large enterprises can use to compute more efficiently? And what are their effects on water consumption, energy use, and overall performance as you were outlining earlier as well?

Zane: Yeah, those are two I think very important opportunities. And let’s just take them one at a  time. Emerging AI world, I think liquid cooling is probably one of the most important low hanging fruit opportunities. So in an air cooled data center, a tremendous amount of energy goes into fans and chillers and evaporative cooling systems. And that is actually a significant part. So if you move a data center to a fully liquid cooled solution, this is an opportunity of around 30% of energy consumption, which is sort of a wow number. I think people are often surprised just how much energy is burned. And if you walk into a data center, you almost need ear protection because it’s so loud and the hotter the components get, the higher the fan speeds get, and the more energy is being burned in the cooling side and liquid cooling takes a lot of that off the table.

What offsets that is liquid cooling is a bit complex. Not everyone is fully able to utilize it. There’s more upfront costs, but actually it saves money in the long run. So the total cost of ownership with liquid cooling is very favorable, and as we’re engineering new data centers from the ground up. Liquid cooling is a really exciting opportunity and I think the faster we can move to liquid cooling, the more energy that we can save. But it’s a complicated world out there. There’s a lot of different situations, a lot of different infrastructures to design around. So we shouldn’t trivialize how hard that is for an individual enterprise. One of the other benefits of liquid cooling is we get out of the business of evaporating water for cooling. A lot of North America data centers are in arid regions and use large quantities of water for evaporative cooling.

That is good from an energy consumption point of view, but the water consumption can be really extraordinary. I’ve seen numbers getting close to a trillion gallons of water per year in North America data centers alone. And then in humid climates like in Southeast Asia or eastern China for example, that evaporative cooling capability is not as effective and so much more energy is burned. And so if you really want to get to really aggressive energy efficiency numbers, you just can’t do it with evaporative cooling in those humid climates. And so those geographies are kind of the tip of the spear for moving into liquid cooling.

The other opportunity you mentioned was density and bringing higher and higher density of computing has been the trend for decades. That is effectively what Moore’s Law has been pushing us forward. And I think it’s just important to realize that’s not done yet. As much as we think about racks of GPUs and accelerators, we can still significantly improve energy consumption with higher and higher density traditional servers that allows us to pack what might’ve been a whole row of racks into a single rack of computing in the future. And those are substantial savings. And at Intel, we’ve announced we have an upcoming processor that has 288 CPU cores and 288 cores in a single package enables us to build racks with as many as 11,000 CPU cores. So the energy savings there is substantial, not just because those chips are very, very efficient, but because the amount of networking equipment and ancillary things around those systems is a lot less because you’re using those resources more efficiently with those very high dense components. So continuing, if perhaps even accelerating our path to this ultra-high dense kind of computing is going to help us get to the energy savings we need maybe to accommodate some of those larger models that are coming.

Laurel: Yeah, that definitely makes sense. And this is a good segue into this other part of it, which is how data centers and hardware as well software can collaborate to create greater energy efficient technology without compromising function. So how can enterprises invest in more energy efficient hardware such as hardware-aware software, and as you were mentioning earlier, large language models or LLMs with smaller downsized infrastructure but still reap the benefits of AI?

Zane: I think there are a lot of opportunities, and maybe the most exciting one that I see right now is that even as we’re pretty wowed and blown away by what these really large models are able to do, even though they require tens of megawatts of super compute power to do, you can actually get a lot of those benefits with far smaller models as long as you’re content to operate them within some specific knowledge domain. So we’ve often referred to these as expert models. So take for example an open source model like the Llama 2 that Meta produced. So there’s like a 7 billion parameter version of that model. There’s also, I think, a 13 and 70 billion parameter versions of that model compared to a GPT-4, maybe something like a trillion element model. So it’s far, far, far smaller, but when you fine tune that model with data to a specific use case, so if you’re an enterprise, you’re probably working on something fairly narrow and specific that you’re trying to do.

Maybe it’s a customer service application or it’s a financial services application, and you as an enterprise have a lot of data from your operations, that’s data that you own and you have the right to use to train the model. And so even though that’s a much smaller model, when you train it on that domain specific data, the domain specific results can be quite good in some cases even better than the large model. So you give up that kind of amazing general purpose use like when you’re using ChatGPT-4 and you can ask it everything from 17th century Italian poetry to quantum mechanics, if you narrow your range, these smaller models can give you equivalent or better kind of capability, but at a tiny fraction of the energy consumption.

And we’ve demonstrated a few times, even with just a standard Intel Xeon two socket server with some of the AI acceleration technologies we have in those systems, you can actually deliver quite a good experience. And that’s without even any GPUs involved in the system. So that’s just good old-fashioned servers and I think that’s pretty exciting.

That also means the technology’s quite accessible, right? So you may be an enterprise, you have a general purpose infrastructure that you use for a lot of things, you can use that for AI use cases as well. And if you’ve taken advantage of these smaller models that fit within infrastructure we already have or infrastructure that you can easily obtain. And so those smaller models are pretty exciting opportunities. And I think that’s probably one of the first things the industry will adopt to get energy consumption under control is just right sizing the model to the activity to the use case that we’re targeting. I think there’s also… you mentioned the concept of hardware-aware software. I think that the collaboration between hardware and software has always been an opportunity for significant efficiency gains.

I mentioned early on in this conversation how virtualization was one of the pillars that gave us that kind of fantastic result over the last 15 years. And that was very much exactly that. That’s bringing some deep collaboration between the operating system and the hardware to do something remarkable. And a lot of the acceleration that exists in AI today actually is a similar kind of thinking, but that’s not really the end of the hardware software collaboration. We can deliver quite stunning results in encryption and in memory utilization in a lot of areas. And I think that that’s got to be an area where the industry is ready to invest. It is very easy to have plug and play hardware where everyone programs at a super high level language, nobody thinks about the impact of their software application downstream. I think that’s going to have to change. We’re going to have to really understand how our application designs are impacting energy consumption going forward. And it isn’t purely a hardware problem. It’s got to be hardware and software working together.

Laurel: And you’ve outlined so many of these different kind of technologies. So how can enterprise adoption of things like modularity and liquid cooling and hardware aware software be incentivized to actually make use of all these new technologies?

Zane: A year ago, I worried a lot about that question. How do we get people who are developing new applications to just be aware of the downstream implications? One of the benefits of this revolution in the last 12 months is I think just availability of electricity is going to be a big challenge for many enterprises as they seek to adopt some of these energy intensive applications. And I think the hard reality of energy availability is going to bring some very strong incentives very quickly to attack these kinds of problems.

But I do think beyond that like a lot of areas in sustainability, accounting is really important. There’s a lot of good intentions. There’s a lot of companies with net-zero goals that they’re serious about. They’re willing to take strong actions against those goals. But if you can’t accurately measure what your impact is either as an enterprise or as a software developer, I think you have to kind of find where the point of action is, where does the rubber meet the road where a micro decision is being made. And if the carbon impact of that is understood at that point, then I think you can see people take the actions to take advantage of the tools and capabilities that are there to get a better result. And so I know there’s a number of initiatives in the industry to create that kind of accounting, and especially for software development, I think that’s going to be really important.

Laurel: Well, it’s also clear there’s an imperative for enterprises that are trying to take advantage of AI to curb that energy consumption as well as meet their environmental, social, and governance or ESG goals. So what are the major challenges that come with making more sustainable AI and computing transformations?

Zane: It’s a complex topic, and I think we’ve already touched on a couple of them. Just as I was just mentioning, definitely getting software developers to understand their impact within the enterprise. And if I’m an enterprise that’s procuring my applications and software, maybe cloud services, I need to make sure that accounting is part of my procurement process, that in some cases that’s gotten easier. In some cases, there’s still work to do. If I’m operating my own infrastructure, I really have to look at liquid cooling, for example, an adoption of some of these more modern technologies that let us get to significant gains in energy efficiency. And of course, really looking at the use cases and finding the most energy efficient architecture for that use case. For example, like using those smaller models that I was talking about. Enterprises need to be very aware of the energy consumption of their digital technologies, how big it is and how their decisions are affecting it.

Laurel: So could you offer an example or use case of one of those energy efficient AI driven architectures and how AI was subsequently deployed for it?

Zane: Yes. I think that some of the best examples I’ve seen in the last year were really around these smaller models where Intel did an example that we published around financial services, and we found that something like three hours of fine-tuning training on financial services data allowed us to create a chatbot solution that performed in an outstanding manner on a standard Xeon processor. And I think making those solutions available to our customers is starting to open people’s eyes how energy efficient you can be while not really giving up a whole lot in terms of the AI use case that you’re looking for. And so I think we need to just continue to get those examples out there. We have a number of collaborations such as with Hugging Face with open source models, enabling those solutions on our products like our Gaudi2 accelerator has also performed very well from a performance per watt point of view, the Xeon processor itself. So those are great opportunities.

Laurel: And then how do you envision the future of AI and sustainability in the next three to five years? There seems like so much opportunity here.

Zane: I think there’s going to be so much change in the next three to five years. I hope no one holds me to what I’m about to say, but I think there are some pretty interesting trends out there. One thing, I think, to think about is the trend of AI factories. So training a model is a little bit of an interesting activity that’s distinct from what we normally think of as real time digital services. You have real time digital service like Vinnie, the app on your iPhone that’s connected somewhere in the cloud, and that’s a real time experience. And it’s all about 99.999% uptime, short latencies to deliver that user experience that people expect. But AI training is different. It’s a little bit more like a factory. We produce models as a product and then the models are used to create the digital services. And that I think becomes an important distinction.

So I can actually build some giant gigawatt facility somewhere that does nothing but train models on a large scale. I can partner with the infrastructure of the electricity providers and utilities much like an aluminum plant or something would do today where I actually modulate my energy consumption with its availability. Or maybe I take advantage of solar or wind power’s ability, I can modulate when I’m consuming power, not consuming power. And so I think if we’re going to see some really large scale kinds of efforts like that, and those AI factories could be very, very efficient, they can be liquid cooled and they can be closely coupled to the utility infrastructure. I think that’s a pretty exciting opportunity. And while that’s kind of an acknowledgement that there’s going to be gigawatts and gigawatts of AI training going on. Second opportunity, I think in this three to five years, I do think liquid cooling will become far more pervasive.

I think that will be driven by the need to cool the next generation of accelerators and GPUs will make it a requirement, but then that will be able to build that technology out and scale it more ubiquitously for all kinds of infrastructure. And that will let us shave huge amounts of gigawatts out of the infrastructure, save hundreds of billions of gallons of water annually. I think that’s incredibly exciting. And if I just… the innovation on the model size as well, so much has changed with just the last five years with large language models like ChatGPT, let’s not assume there’s not going to be even bigger change in the next three to five years. What are the new problems that are going to be solved, new innovations? So I think as the costs and impact of AI are being felt more substantively, there’ll be a lot of innovation on the model side and people will come up with new ways of cracking some of these problems and there’ll be new exciting use cases that come about.

Finally, I think on the hardware side, there will be new AI architectures. From an acceleration point of view today, a lot of AI performance is limited by memory bandwidth, memory bandwidth and networking bandwidth between the various accelerator components. And I don’t think we’re anywhere close to having an optimized AI training system or AI inferencing systems. I think the discipline is moving faster than the hardware and there’s a lot of opportunity for optimization. So I think we’ll see significant differences in networking, significant differences in memory solutions over the next three to five years, and certainly over the next 10 years that I think can open up a substantial set of improvements.

And of course, Moore’s Law itself continues to advance advanced packaging technologies, new transistor types that allow us to build ever more ambitious pieces of silicon, which will have substantially higher energy efficiency. So all of those things I think will be important. Whether we can keep up with our energy efficiency gains with the explosion in AI functionality, I think that’s the real question and it’s just going to be a super interesting time. I think it’s going to be a very innovative time in the computing industry over the next few years.

Laurel: And we’ll have to see. Zane, thank you so much for joining us on the Business Lab.

Zane: Thank you.

Laurel: That was Zane Ball, corporate vice president and general manager of data center platform engineering and architecture at Intel, who I spoke with from Cambridge, Massachusetts, the home of MIT and MIT Technology Review.

That’s it for this episode of Business Lab. I’m your host, Laurel Ruma. I’m the director of Insights, the custom publishing division of MIT Technology Review. We were founded in 1899 at the Massachusetts Institute of Technology, and you can also find us in print, on the web, and at events each year around the world. For more information about us and the show, please check out our website at technologyreview.com.

This show is available wherever you get your podcasts. If you enjoyed this episode, we hope you’ll take a moment to rate and review us. Business Lab is a production of MIT Technology Review. This episode was produced by Giro Studios. Thanks for listening.


This content was produced by Insights, the custom content arm of MIT Technology Review. It was not written by MIT Technology Review’s editorial staff.

Bringing breakthrough data intelligence to industries

As organizations recognize the transformational opportunity presented by generative AI, they must consider how to deploy that technology across the enterprise in the context of their unique industry challenges, priorities, data types, applications, ecosystem partners, and governance requirements. Financial institutions, for example, need to ensure that data and AI governance has the built-in intelligence to fully align with strict compliance and regulatory requirements. Media and entertainment (M&E) companies seek to build AI models to drive deeper product personalization. And manufacturers want to use AI to make their internet of things (IoT) data insights readily accessible to everyone from the data scientist to the shop floor worker.

In any of these scenarios, the starting point is access to all relevant data—of any type, from any source, in real time—governed comprehensively and shared across an industry ecosystem. When organizations can achieve this with the right data and AI foundation, they have the beginnings of data intelligence: the ability to understand their data and break free from data silos that would block the most valuable AI outcomes.

But true data intelligence is about more than establishing the right data foundation. Organizations are also wrestling with how to overcome dependence on highly technical staff and create frameworks for data privacy and organizational control when using generative AI. Specifically, they are looking to enable all employees to use natural language to glean actionable insight from the company’s own data; to leverage that data at scale to train, build, deploy, and tune their own secure large language models (LLMs); and to infuse intelligence about the company’s data into every business process.

In this next frontier of data intelligence, organizations will maximize value by democratizing AI while differentiating through their people, processes, and technology within their industry context. Based on a global, cross-industry survey of 600 technology leaders as well as in-depth interviews with technology leaders, this report explores the foundations being built and leveraged across industries to democratize data and AI. Following are its key findings:

• Real-time access to data, streaming, and analytics are priorities in every industry. Because of the power of data-driven decision-making and its potential for game-changing innovation, CIOs require seamless access to all of their data and the ability to glean insights from it in real time. Seventy-two percent of survey respondents say the ability to stream data in real time for analysis and action is “very important” to their overall technology goals, while another 20% believe it is “somewhat important”—whether that means enabling real-time recommendations in retail or identifying a next best action in a critical health-care triage situation.

• All industries aim to unify their data and AI governance models. Aspirations for a single approach to governance of data and AI assets are strong: 60% of survey respondents say a single approach to built-in governance for data and AI is “very important,” and an additional 38% say it is “somewhat important,” suggesting that many organizations struggle with a fragmented or siloed data architecture. Every industry will have to achieve this unified governance in the context of its own unique systems of record, data pipelines, and requirements for security and compliance.

• Industry data ecosystems and sharing across platforms will provide a new foundation for AI-led growth. In every industry, technology leaders see promise in technology-agnostic data sharing across an industry ecosystem, in support of AI models and core operations that will drive more accurate, relevant, and profitable outcomes. Technology teams at insurers and retailers, for example, aim to ingest partner data to support real-time pricing and product offer decisions in online marketplaces, while manufacturers see data sharing as an important capability for continuous supply chain optimization. Sixty-four percent of survey respondents say the ability to share live data across platforms is “very important,” while an additional 31% say it is “somewhat important.” Furthermore, 84% believe a managed central marketplace for data sets, machine learning models, and notebooks is very or somewhat important.

• Preserving data and AI flexibility across clouds resonates with all verticals. Sixty-three percent of respondents across verticals believe that the ability to leverage multiple cloud providers is at least somewhat important, while 70% feel the same about open-source standards and technology. This is consistent with the finding that 56% of respondents see a single system to manage structured and unstructured data across business intelligence and AI as “very important,” while an additional 40% see this as “somewhat important.” Executives are prioritizing access to all of the organization’s data, of any type and from any source, securely and without compromise.

• Industry-specific requirements will drive the prioritization and pace by which generative AI use cases are adopted. Supply chain optimization is the highest-value generative AI use case for survey respondents in manufacturing, while it is real-time data analysis and insights for the public sector, personalization and customer experience for M&E, and quality control for telecommunications. Generative AI adoption will not be one-size-fits-all; each industry is taking its own strategy and approach. But in every case, value creation will depend on access to data and AI permeating the enterprise’s ecosystem and AI being embedded into its products and services.

Maximizing value and scaling the impact of AI across people, processes, and technology is a common goal across industries. But industry differences merit close attention for their implications on how intelligence is infused into the data and AI platforms. Whether it be for the retail associate driving omnichannel sales, the health-care practitioner pursuing real-world evidence, the actuary analyzing risk and uncertainty, the factory worker diagnosing equipment, or the telecom field agent assessing network health, the language and scenarios AI will support vary significantly when democratized to the front lines of every industry.

Download the report.

This content was produced by Insights, the custom content arm of MIT Technology Review. It was not written by MIT Technology Review’s editorial staff.

Navigating a shifting customer-engagement landscape with generative AI

One can’t step into the same river twice. This simple representation of change as the only constant was taught by the Greek philosopher Heraclitus more than 2000 years ago. Today, it rings truer than ever with the advent of generative AI. The emergence of generative AI is having a profound effect on today’s enterprises—business leaders face a rapidly changing technology that they need to grasp to meet evolving consumer expectations.

“Across all industries, customers are at the core, and tapping into their latent needs is one of the most important elements to sustain and grow a business,” says Akhilesh Ayer, executive vice president and global head of AI, analytics, data, and research practice at WNS Triange, a unit of WNS Global Services, a leading business process management company. “Generative AI is a new way for companies to realize this need.”

A strategic imperative

Generative AI’s ability to harness customer data in a highly sophisticated manner means enterprises are accelerating plans to invest in and leverage the technology’s capabilities. In a study titled “The Future of Enterprise Data & AI,” Corinium Intelligence and WNS Triange surveyed 100 global C-suite leaders and decision-makers specializing in AI, analytics, and data. Seventy-six percent of the respondents said that their organizations are already using or planning to use generative AI.

According to McKinsey, while generative AI will affect most business functions, “four of them will likely account for 75% of the total annual value it can deliver.” Among these are marketing and sales and customer operations. Yet, despite the technology’s benefits, many leaders are unsure about the right approach to take and mindful of the risks associated with large investments.

Mapping out a generative AI pathway

One of the first challenges organizations need to overcome is senior leadership alignment. “You need the necessary strategy; you need the ability to have the necessary buy-in of people,” says Ayer. “You need to make sure that you’ve got the right use case and business case for each one of them.” In other words, a clearly defined roadmap and precise business objectives are as crucial as understanding whether a process is amenable to the use of generative AI.

The implementation of a generative AI strategy can take time. According to Ayer, business leaders should maintain a realistic perspective on the duration required for formulating a strategy, conduct necessary training across various teams and functions, and identify the areas of value addition. And for any generative AI deployment to work seamlessly, the right data ecosystems must be in place.

Ayer cites WNS Triange’s collaboration with an insurer to create a claims process by leveraging generative AI. Thanks to the new technology, the insurer can immediately assess the severity of a vehicle’s damage from an accident and make a claims recommendation based on the unstructured data provided by the client. “Because this can be immediately assessed by a surveyor and they can reach a recommendation quickly, this instantly improves the insurer’s ability to satisfy their policyholders and reduce the claims processing time,” Ayer explains.

All that, however, would not be possible without data on past claims history, repair costs, transaction data, and other necessary data sets to extract clear value from generative AI analysis. “Be very clear about data sufficiency. Don’t jump into a program where eventually you realize you don’t have the necessary data,” Ayer says.

The benefits of third-party experience

Enterprises are increasingly aware that they must embrace generative AI, but knowing where to begin is another thing. “You start off wanting to make sure you don’t repeat mistakes other people have made,” says Ayer. An external provider can help organizations avoid those mistakes and leverage best practices and frameworks for testing and defining explainability and benchmarks for return on investment (ROI).

Using pre-built solutions by external partners can expedite time to market and increase a generative AI program’s value. These solutions can harness pre-built industry-specific generative AI platforms to accelerate deployment. “Generative AI programs can be extremely complicated,” Ayer points out. “There are a lot of infrastructure requirements, touch points with customers, and internal regulations. Organizations will also have to consider using pre-built solutions to accelerate speed to value. Third-party service providers bring the expertise of having an integrated approach to all these elements.”

Ayer offers the example of WNS Triange helping a travel intermediary use generative AI to deal with customer inquiries about airline rescheduling, cancellations, and other itinerary complications. “Our solution is immediately able to go into a thousand policy documents, pick out the policy parameters relevant to the query… and then come back quickly not only with the response but with a nice, summarized, human-like response,” he says.

In another example, Ayer shares that his company helped a global retailer create generative AI–driven designs for personalized gift cards. “The customer experience goes up tremendously,” he says.

Hurdles in the generative AI journey

As with any emerging technology, however, there are organizational, technical, and implementation barriers to overcome when adopting generative AI.

Organizational:  One of the major hurdles businesses can face is people. “There is often immediate resistance to the adoption of generative AI because it affects the way people work on a daily basis,” says Ayer.

As a result, securing internal buy-in from all teams and being mindful of a skills gap is a must. Additionally, the ability to create a business case for investment—and getting buy-in from the C-suite—will help expedite the adoption of generative AI tools.

Technical: The second set of obstacles relates to large language models (LLMs) and mechanisms to safeguard against hallucinations and bias and ensure data quality. “Companies need to figure out if generative AI can solve the whole problem or if they still need human input to validate the outputs from LLM models,” Ayer explains. At the same time, organizations must ask whether the generative AI models being used have been appropriately trained within the customer context or with the enterprise’s own data and insights. If not, there is a high chance that the response will be incorrect. Another related challenge is bias: If the underlying data has certain biases, the modeling of the LLM could be unfair. “There have to be mechanisms to address that,” says Ayer. Other issues, such as data quality, output authenticity, and explainability, also must be addressed.

Implementation: The final set of challenges relates to actual implementation. The cost of implementation can be significant, especially if companies cannot orchestrate a viable solution, says Ayer. In addition, the right infrastructure and people must be in place to avoid resource constraints. And users must be convinced that the output will be relevant and of high quality, so as to gain their acceptance for the program’s implementation.

Lastly, privacy and ethical issues must be addressed. The Corinium Intelligence and WNS Triange survey showed that almost 72% of respondents were concerned about ethical AI decision-making.

The focus of future investment

The entire ecosystem of generative AI is moving quickly. Enterprises must be agile and adapt quickly to change to ensure customer expectations are met and maintain a competitive edge. While it is almost impossible to anticipate what’s next with such a new and fast-developing technology, Ayer says that organizations that want to harness the potential of generative AI are likely to increase investment in three areas:

  • Data modernization, data management, data quality, and governance: To ensure underlying data is correct and can be leveraged.
  • Talent and workforce: To meet demand, training, apprenticeships, and injection of fresh talent or leveraging market-ready talent from service providers will be required.
  • Data privacy solutions and mechanisms: To ensure privacy is maintained, C-suite leaders must also keep pace with relevant laws and regulations across relevant jurisdictions.

However, it shouldn’t be a case of throwing everything at the wall and seeing what sticks. Ayer advises that organizations examine ROI from the effectiveness of services or products provided to customers. Business leaders must clearly demonstrate and measure a marked improvement in customer satisfaction levels using generative AI–based interventions.

“Along with a defined generative AI strategy, companies need to understand how to apply and build use cases, how to execute them at scale and speed to market, and how to measure their success,” says Ayer. Leveraging generative AI for customer engagement is typically a multi-pronged approach, and a successful partnership can help with every stage.

This content was produced by Insights, the custom content arm of MIT Technology Review. It was not written by MIT Technology Review’s editorial staff.

Developing climate solutions with green software

After years of committing to sustainable practices in his personal life from recycling to using cloth-based diapers, Asim Hussain, currently the director of green software and ecosystems at Intel, began to ask questions about the practices in his work: software development.

Developers often asked if their software was secure enough, fast enough, or cost-effective enough but, Hussain says, they rarely considered the environmental consequences of their applications. Hussain would go on to work at Intel and become the executive director and chairperson of the Green Software Foundation, a non-profit aiming to create an ecosystem of people, tooling, and best practices around sustainable software development.

“What we need to do as software developers and software engineers is we need to make sure that it is emitting the least amount of carbon for the same amount of value and user functionality that we’re getting out of it,” says Hussain.

The three pillars of green software are energy efficiency, hardware efficiency, and carbon awareness. Making more efficient use of hardware and energy consumption when developing applications can go a long way toward reducing emissions, Hussain says. And carbon-aware computing involves divestment from fossil fuels in favor of renewable energy sources to improve efficiency without compromising performance.

Often, when something is dubbed “green,” there is an assumption that the product, application, or practice functions worse than its less environmentally friendly version. With software, however, the opposite is true.

“Being green in the software space means being more efficient, which translates almost always to being faster,” says Hussain. “When you factor in the hardware efficiency component, oftentimes it translates to building software that is more resilient, more fault-tolerant. Oftentimes it also translates then into being cheaper.”

Instituting green software necessitates not just a shift in practices and tooling but also a culture change within an enterprise. While regulations and ESG targets help to create an imperative, says Hussain, a shift in mindset can enable some of the greatest strides forward.

“If there’s anything we really need to do is to drive that behavior change, we need to drive behavior change so people actually invest their time on making software more energy efficient, more hardware efficient, or more carbon aware.”

This episode of Business Lab is produced in partnership with Intel.

Full Transcript

Laurel Ruma: From MIT Technology Review, I’m Laurel Ruma and this is Business Lab, the show that helps business leaders make sense of new technologies coming out of the lab and into the marketplace.

Our topic is green software, from apps to devices to the cloud. Computing runs the world around us. However, there is a better way to do it with a focus on sustainability.

Two words for you: sustainable code.

My guest is Asim Hussain, who is the director of the Office of Green Software and Ecosystems at Intel, as well as the chairperson of the Green Software Foundation.

This podcast is produced in partnership with Intel.

Welcome, Asim.

Asim Hussain: Hi Laurel. Thank you very much for having me.

Laurel: Well, glad you’re here. So for a bit of background, you’ve been working in software development and sustainability advocacy from startups to global enterprises for the last two decades. What drew you into sustainability as a focus and what are you working on now?

Asim: I’ve personally been involved and interested in the sustainability space for quite a while on a very personal level. Then around the birth of my first son, about five years ago now, I started asking myself this one question, which was how come I was willing to do all these things I was doing for sustainability to recycle, we were using cloth-based nappies, all sorts of these different things. Yet I could not remember in my entire career, my entire career, I could not remember one single moment where in any technical discussion, in any architectural meeting, in any discussion about how we’re going to build this piece of software. I mean, people oftentimes raise points around is this secure enough? Is this fast enough? Does this cost too much? But at no point I’d ever heard anybody ask the question, is this emitting too much carbon? Is this piece of software, is this solution that we’re talking about right now, how does that solution, what kind of environmental impacts does that have? I’ve never, ever, ever heard anybody raise that question.

So I really started to ask that question myself. I found other people who are like me. Five years ago, there weren’t many of us, but were all asking the same questions. I joined and then I started to become a co-organizer of a community called ClimateAction.Tech. Then the community just grew. A lot of people were starting to ask themselves these questions and some answers were coming along. At the time, I used to work at Microsoft and I pitched and formed something called the green cloud advocacy team, where we talked about how to actually build applications in a greener way on the cloud.

We formed something called the Green Software Foundation, which is a consortium of now 60 member organizations, which I am a chairperson of. Over a year ago I joined Intel because Intel has been heavily investing in sustainable software space. If you think about what Intel does, pretty much everything that Intel produces, developers use it and developers write software and write code on Intel’s products. So it makes sense for Intel to have a strong green software strategy. That’s kind of why I was brought in and I’ve since then been working on Intel’s green software strategy internally.

Laurel: So a little bit more about that. How can organizations make their software greener? Then maybe we should take a step back and define what green software actually is.

Asim: Well, I think we have to define what green software actually is first. The way the conversation’s landed in recent years and the Green Software Foundation has been a large part of this is we’ve coalesced around this idea of carbon efficiency, which is if you are building a piece of software … Everything we do emits carbon, everything we do emits carbon, this tool we’re using right now to record this session is emitting carbon right now. What we need to do as software developers and software engineers is we need to make sure that it is emitting the least amount of carbon for the same amount of value and user functionality that we’re getting out of it. That’s what we call carbon efficiency.

What we say is there’s three pillars underneath, there’s only really three ways to make your software green. The first is to make it more energy efficient, to use less energy. Most electricity is still created through the burning of fossil fuels. So just using less electricity is going to emit fewer carbon emissions into the atmosphere. So the first is energy efficiency. The second is hardware efficiency because all software runs on hardware and depends on the, if you’re talking about a mobile phone, typically people are forced to move on from mobile phones because the software just doesn’t run on their older models. In the cloud it tends to be more around a conversation around utilization by making more use of the servers that you already have in the cloud, making just more efficient use of the hardware. The third one is a very interesting space. It’s a very new space. It’s called carbon awareness or carbon-aware computing. That is you are going to be using electricity anyway. Can you make your software? Can you architect your software in such a way?

So it does more when the electricity is clean and does less when the electricity is dirty. So can you architect an application? So for instance, it does more when there’s more renewable energy on the grid right now, and it does less when more coal or gas is getting burnt. There’s some very interesting projects in this space that have been happening, very high-profile projects and carbon-aware computing is an area where there’s a lot of interest because it’s a stepping stone. It might not get you your 50, 60, 70% carbon reductions, but it will get you your 1, 2, 3, and 4% carbon reductions and it’ll get you that with very minimal investments. There’s a lot of interest in carbon-aware computing. But those are basically the three areas, what we call the three pillars of green software, energy efficiency, hardware efficiency, and carbon awareness.

Laurel: So another reason we’re talking about all of this is that technology can contribute to the environmental issues that it is trying to actually help. So for example, a lot of energy is needed to train AI models. Also, blockchain was key in the development of energy-efficient microgrids, but it’s also behind the development of cryptocurrency platforms, some of which consume more energy than that of a small country. So how can advanced technologies like AI, machine learning, and blockchain contribute positively to the development of green software?

Asim: That’s an interesting question because sometimes the focus oftentimes is how do we actually make that technology greener? But I don’t believe that is necessarily the whole story. The story is the broader story. How can we use that technology to make software greener? I think there’s many ways you can probably tackle that question. One thing that’s been interesting for me since my journey as a software developer joining Intel is me realizing how little I knew about hardware. There is so much, I describe it as the gap between software and silicon. The gap is quite large right now. If you’re building software these days, you have very little understanding of the silicon that’s running that software. Through a greater understanding of exactly how your software is exactly getting executed by the silicon to implement the functionality, that’s where we are seeing that there’s a lot of great opportunities to reduce emissions and to make that software more energy efficient, more hardware efficient.

I think that’s where places like AI can really help out. Developer productivity has been the buzzword in this space for a very long time. Developers are extremely expensive. Getting to market fast and beating your competition is the name of the game these days. So it’s always been about how do we implement the functionality we need as fast as possible, make sure it’s secure, get it out the door. But oftentimes the only way you can do that is to increase the gap between the software and silicon and just make it a little bit more inefficient. I think AI can really help there. You can build AI solutions that can, there’s copilot solutions which can help as you’re developing code could actually suggest to you. If you were to write your code in a slightly different way, it could be more efficient. So that’s one way AI can help out.

Another way that I’m seeing AI utilized in this space as well is when you deploy … Silicon and the products that we produce can actually, they come out of the box configured in a certain way, but they can actually be tuned to actually execute that particular piece of software much more efficiently. So if you have a data center running just one type of software, you can actually tune the hardware so that software is run more efficiently on that hardware. We’re seeing AI solutions come on the market these days, which can then automatically just figure out what type of application are you, how do you run, how do you work? We have a solution called Granulate, which does part of this as well. It can then figure out how do you tune the underlying hardware in such a way so it executes that software more efficiently. So I think that’s kind of a couple of ways that this technology could actually be used to make software itself greener.

Laurel: To bridge that gap between software and silicon, you must be able to measure the progress and meet targets. So what parameters do you use to measure the energy efficiency of software? Could you talk us through the tenets of actually measuring?

Asim: So measuring is an extremely challenging problem. When we first launched the Green Software Foundation three years ago, I remember asking all the members, what is your biggest pain point? They all came back, almost all came back with measuring. Measuring is very, very challenging. It’s so nuanced, there’s so many different levels to it. For instance, at Intel, we have technology in our chips to actually measure the energy of the whole chip. Those counters on the chip which measure it. Unfortunately, that only gives you the energy of the entire chip itself. So it does give you a measurement, but then if you are a developer, there’s maybe 10 processes running on that chip and only one of them is yours. You need to know how much energy is your process consuming because that’s what you can optimize for. That’s what you can see. Currently, the best way to measure at that level is using models, models which are either generated again through AI or through other processes where you can effectively just run lots large amounts of data and generate statistical models.

Oftentimes a model that’s used is one that uses CPU [central processing unit] utilization, so how busy a CPU is and translate that into energy. So you can see my process is consuming 10% of the CPU. There are models out there that can convert that into energy, but again, all models are wrong, some models are useful. So there’s always so much nuance to this whole space as well, because how have you tweaked your computer? What else is running on your computer? It can also affect how those numbers are measured. So, unfortunately, this is a very, very challenging area.

But this is really the really big area that a lot of people are trying to resolve right now. We are not at the perfect solution, but we are way, way better than we were three, four or five years ago. It’s actually a very exciting time for measurement in this space.

Laurel: Well, and I guess part of it is that green software seems to be developed with greater scrutiny and higher quality controls to ensure that the product actually meets these standards to reduce emissions. Measurement is part of that, right? So what are some of the rewards beyond emissions reduction or meeting green goals of developing software? You kind of touched on that earlier with the carbon efficiency as well as hardware efficiency.

Asim: Yeah, so this is something I used to think about a lot because the term green has a lot associated with it. I mean, oftentimes when people historically have used the word green, you can have the main product or the green version of the product. There’s an idea in your mind that the green version is somehow less than, it’s somehow not as good. But actually in the software space it’s so interesting because the exact opposite. Being green in the software space means being more efficient, which translates almost always to being faster. When you factor in the hardware efficiency component, oftentimes it translates to building software that is more resilient, more fault-tolerant. Oftentimes it also translates then into being cheaper. So actually green has a lot of positive associations with it already.

Laurel: So in that vein, how can external standards help provide guidance for building software and solutions? I mean, obviously, there’s a need to create something like the Green Software Foundation, and with the focus that most enterprises have now on environmental, social, and governance goals or ESG, companies are now looking more and more to build those ideas into their everyday workflow. So how do regulations help and not necessarily hinder this kind of progress?

Asim: So standards are very, very important in this space. Standards, I mean, one of the things, again, when we look to the ecosystem about three, four years ago, the biggest problem the enterprises had, I mean a lot of them were very interested in green software, but the biggest problem they had was what do they trust? What can I trust? Whose advice should I take? That’s where standards come in. That’s where standards are most important. Standards are, at least the way we develop standards inside the Green Software Foundation, they’re done via consensus. There are like 60 member organizations. So when you see a standard that’s been created by that many people and that many people have been involved with it, it really builds up that trust. So now you know what to do. Those standards give you that compass direction to tell you which direction to go in and that you can trust.

There’s several standards that we’ve been focusing on in the Green Software Foundation, one’s called the SEI, which is a software carbon intensity specification. Again, to prove it as an ISO standard, you have to reach consensus through 196 countries. So then you get even more trust into a standard so you can use it. So standards really help to build up that trust, which organizations can use to help guide them in the directions to take. There’s a couple of other standards that are really coming up in the foundation that I think are quite interesting. One is called Real-Time Cloud. One of the challenges right now is, and again always comes back to measurement, it always always comes back to measurement. Right now measurement is very discreet, it happens oftentimes just a few times a year. Oftentimes when you get measurement data, it is very delayed. So one of the specs that’s been worked on right now is called Real-Time Cloud.

It’s trying to ask the question, is it possible? Is it possible to get data that is real-time? Oftentimes when you want to react and change behaviors, you need real-time data. If you want data so that when somebody does something, they know instantly the impact of that action so they can make adjustments instantly. If they’re having to wait three months, that behavior change might not happen. Real-time [data] is oftentimes at log aheads with regulations because oftentimes you have to get your data audited and auditing data that’s real-time is very, very challenging. So one of the questions we’re trying to ask is, is it possible to have data which is real-time, which then over the course of a year, you can imagine it just aggregates up over the course of a year. Can that aggregation then provide enough trust so that an auditor can then say, actually, we now trust this information and we will allow that to be used in regulatory reporting.

That’s something that we’re very excited about because you really need real-time data to drive behavior change. If there’s anything we really need to do is to drive that behavior change, we need to drive behavior change so people actually invest their time on making software more energy efficient, more hardware efficient, or more carbon aware. So that’s some of the ways where standards are really helping in this space.

Laurel: I think it’s really helpful to talk about standards and how they are so ingrained with software development in general because there are so many misconceptions about sustainability. So what are some of the other misconceptions that people kind of get stuck on, maybe that even calling it green, right? Are there philosophies or strategies that you can caution against or you try to advocate for?

Asim: So as a couple of things I talk about, so one of the things I talk about is it does take everybody, I mean, I remember very early on when I was talking in this space, oftentimes a conversation went, oh, don’t bother talking to that person or don’t talk to this sector of developers, don’t talk to that type of developers. Only talk to these people, these people who have the most influence to make the kind of changes to make software greener. But it really takes a cultural change. This is what’s very important, really takes a cultural change inside an organization. It takes everybody. You can’t really talk to one slice of the developer ecosystem. You need to talk to everybody. Every single developer or engineer inside an organization really needs to take this on board. So that’s one of the things I say is that you have to speak to every single person. You cannot just speak to one set of people and exclude another set of people.

Another challenge that I often see is that people, when they talk about this space, one of the misconceptions they talk about is they rank where effort should be spent in terms of the carbon slice of the pie that it is responsible for and I’ll talk about this in general. But really how you should be focusing is you should be focusing not on the slice of the pie, but on the ability to decarbonize that slice of the pie. That’s why green software is so interesting and that’s why it’s such a great place to spend effort and time. It might not be, I mean it is, depending on which academic paper you look at, it can be between 2 to 4% of global emissions. So some people might say, well, that’s not really worth spending the time in.

But my argument is actually the ability for us to decarbonize that 2 to 4% is far easier than our ability to decarbonize other sectors like airlines or concrete or these other sectors. We know what we need to do oftentimes in the software space, we know the choices. There doesn’t need to be new technology made, there just needs to be decisions made to prioritize this work. That’s something I think is very, very important. We should rank everything in terms of our ability to decarbonize the ease of decarbonization and then work on the topmost item first down, rather than just looking at things in just terms of tons of carbon, which I think leads to wrong decision making.

Laurel: Well, I think you’re laying out a really good argument because green initiatives, they can be daunting, especially for large enterprises looking to meet those decarbonization thresholds within the next decade. For those companies that are making the investment into this, how should they begin? Where are the fundamental things just to be aware of when you’re starting this journey?

Asim: So the first step is, I would say training. What we’re describing here in terms of, especially in terms of the green software space, it’s a very new movement. It’s a very new field of computing. So a lot of the terms that I talk about are just not well understood and a lot of the reasons for those terms are not well understood as well. So the number one thing I always say is you need to focus on training. There’s loads of training out there. The Green Software Foundation’s got some training, learn.GreenSoftware.Foundation, it’s just two hours, it’s free. We send that over to anybody who’s starting in this space just to understand the language, the terminology, just to get everybody on the same page. That is usually a very good start. Now in terms of how do you motivate inside, I think about this a lot.

If you’re the lead of an organization and you want to make a change, how do you actually make that change? I’m a big, big believer in trusting your team, trusting your people. If you give engineers a problem, they will find a solution to that problem. But what they oftentimes need is permission, a thumbs up from leadership that this is a priority. So that’s why it’s very important for organizations to be very public about their commitments, make public commitments. Same way Intel has made public commitments. Be very vocal as a leader inside your organization and be very clear that this is a priority for you, that you will listen to people and to teams who bring you solutions in this space.

You will find that people within your organization are already thinking about this space, already have ideas, already probably have decks ready to present to you. Just create an environment where they feel capable of presenting it to you. I guarantee you, your solutions are already within your organization and already within the minds of your employees.

Laurel: Well, that is all very inspiring and interesting and so exciting. So when you think about the next three to five years in green software development and adoption, what are you looking forward to the most? What excites you?

Asim: I think I’m very excited right now, to be honest with you. I look back, I look back five years ago the very, very early days, first looked at this, and I still remember if there was one article, one mentioning green software, we would all lose our heads. We’d get so excited about it, we’d share it, we’d pour over it. Now I’m inundated with information. This movement has grown significantly. There are so many organizations that are deeply interested in this space. There’s so much research, so much academic research.

I have so many articles coming my way every single week. I do not have time to read them. So that gives me just a lot of hope for the future. That really excites me. I might just be because I’m at this kind of cutting edge of this space, so I see a lot of this stuff before anybody else, but I see a huge amount of interest and I see also a huge amount of activity as well. I see a lot of people working on solutions, not just talking about problems, but working on solutions to those problems. That honestly just excites me. I don’t know where we’re going to end up in five years time, but if this is our growth so far, I think we’re going to end up in a very good place.

Laurel: Oh, that’s excellent. Awesome. Thank you so much for joining us today on the Business Lab.

Asim: Thank you very much for having me.

Laurel: That was Asim Hussain, the director of the Office of Green Software and Ecosystems at Intel, who I spoke with from Cambridge, Massachusetts, the home of MIT and MIT Technology Review.

That’s it for this episode of Business Lab. I’m your host, Laurel Ruma. I’m the director of Insights, the custom publishing division of MIT Technology Review. We were founded in 1899 at the Massachusetts Institute of Technology, and you can also find us in print on the web and at events each year around the world. For more information about us and the show, please check out our website at technologyreview.com.

This show is available wherever you get your podcasts. If you enjoyed this episode, we hope you’ll take a moment to rate and review us. Business Lab is a production of MIT Technology Review. This episode was produced by Giro Studios. Thanks for listening.

This content was produced by Insights, the custom content arm of MIT Technology Review. It was not written by MIT Technology Review’s editorial staff.

Mapping the micro and macro of biology with spatial omics and AI

37 trillion. That is the number or cells that form a human being. How they all work together to sustain life is possibly the biggest unsolved puzzle in biology. A group of up-and-coming technologies for spatially resolved multi omics, here collectively called “spatial omics,” may provide researchers with the solution.

Over the last 20 years, the omics revolution has enabled us to understand cell and tissue biology at ever increasing resolutions. Bulk sequencing techniques that emerged in the mid 2000s allowed the study of mixed populations of cells. A decade later, single-cell omics methods became commercially available, revolutionizing our understanding of cell physiology and pathology. These methods, however, required dissociating cells from their tissue of origin, making it impossible to study their spatial organization in tissue.

Spatial omics refers to the ability to measure the activity of biomolecules (RNA, DNA, proteins, and other omics) in situ—directly from tissue samples. This is important because many biological processes are controlled by highly localized interactions between cells that take place in spatially heterogeneous tissue environments. Spatial omics allows previously unobservable cellular organization and biological events to be viewed in unprecedented detail.

A few years ago, these technologies were just prototypes in a handful of labs around the world. They worked only on frozen tissue and they required impractically large amounts of precious tissue biopsies. But as these challenges have been overcome and the technologies commercialized by life science technology providers, these tools have become available to the wider scientific community. Spatial omics technologies are now improving at a rapid pace, increasing the number of biomolecules that can be profiled from hundreds to tens of thousands, while increasing resolution to single-cell and even subcellular scales.

Complementary advances in data and AI will expand the impact of spatial omics on life sciences and health care—while also raising new questions. How are we going to generate the large datasets that are necessary to make clinically relevant discoveries? What will data scientists see in spatial omics data through the lens of AI?

Discovery requires large-scale spatial omics datasets

Several areas of life science are already benefiting from discoveries made possible by spatial omics, with the biggest impacts in cancer and neurodegenerative disease research. However, spatial omics technologies are very new, and experiments are challenging and costly to execute. Most present studies are performed by single institutions and include only a few dozen patients. Complex cell interactions are highly patient-specific, and they cannot be fully understood from these small cohorts. Researchers need the data to enable hypothesis generation and discovery. 

This requires a shift in mentality toward collaborative projects, which can generate large-scale reference datasets both on healthy organs and human diseases. Initiatives such as The Cancer Genome Atlas (TCGA) have transformed our understanding of cancer. Similar large-scale spatial omics efforts are needed to systematically interrogate the role of spatial organization in healthy and diseased tissues; they will generate large datasets to fuel many discovery programs. In addition, collaborative initiatives steer further improvement of spatial omics technologies, generate data standards and infrastructures for data repositories, and drive the development and adoption of computational tools and algorithms.

At Owkin we are pioneering the generation of such datasets. In June 2023, we launched an initiative to create the world largest spatial omics dataset in cancer, with a vision to include data from 7,000 patients with seven difficult-to-treat cancers. The project, known as MOSAIC (Multi-Omics Spatial Atlas in Cancer), won’t stop at the data generation, but will mine the data to learn disease biology and identify new molecular targets against which to design new drugs.

Owkin is well placed to drive this kind of initiative. We can tap a vast network of collaborating hospitals across the globe: to create the MOSAIC dataset, we are working with five world-class cancer research hospitals. And we have deep experience in AI: In the last five years, we have published 54 research papers generating AI methodological innovation and building predictive models in several disease areas, including many types of cancer.

AI’s transformative role in discovering new biology

Spatial omics was recognized as method of the year 2020 by Nature Methods, and it was named one of the top 10 emerging technologies by the World Economic Forum in 2023—alongside generative AI.

With these two technologies developing in tandem, the opportunities for AI-driven biological discoveries from spatial omics are numerous. Looking at the fast-evolving landscape of spatial omics AI methods, we see two broad categories of new methods breaking through.

In the first category are AI methods that aim to improve the usability of spatial omics and enable richer downstream analyses for researchers. Such methods are specially designed to deal with the high dimensionality and the signal-to-noise ratio that are specific to spatial omics. Some are used to remove technical artifacts and batch effects from the data. Other methods, collectively known as “super-resolution methods,” use AI to increase the resolution of spatial omics assays to near single-cell levels. Another group of approaches looks to integrate dissociated single-cell omics with spatial omics. Collectively, these AI methods are bridging the gap with future spatial omics technologies.

In the second category, AI methods aim at discovering new biology from spatial omics. By exploiting the localization information of spatial omics, they shed light on how groups of cells organize and communicate with unprecedented resolution.  Such methods are sharpening our understanding of how cells interact to form complex tissues.

At Owkin, we are developing methods to identify new therapeutic targets and patient subpopulations using spatial omics. We have pioneered methods allowing researchers to understand how cancer patient outcomes are linked to tumor heterogeneity, directly from tumor biopsy images. Building on this expertise and the MOSAIC consortium, we are developing the next generation of AI methods, which will link patient-level outcomes with an understanding of disease heterogeneity at the molecular level.

Looking ahead

Spatial biology has the potential to radically change our understanding of biology. It will change how we see a biomarker, going from the mere presence of a particular molecule in a sample to patterns of cells expressing a certain molecule in a tissue. Promising research on spatial biomarkers has been published for several diseases, including Alzheimer’s disease and ovarian cancer. Spatial omics has already been used in research associated with clinical trials to monitor tumor progression in patients.

Five years from now, spatial technologies will be capable of mapping every human protein, RNA, and metabolite at subcellular resolution. The computing infrastructure to store and analyze spatial omics data will be in place, as will the necessary standards for data and metadata and the analytical algorithms. The tumor microenvironment and cellular composition of difficult-to-treat cancers will be mapped through collaborative efforts such as MOSAIC.

Spatial omics datasets from patient biopsies will quickly become an essential part of pharmaceutical R&D, and through the lens of AI methods, they will be used to inform the design of new, more efficacious drugs and to drive faster and better-designed clinical trials to bring those drugs to patients. In the clinic, spatial omics data will routinely be collected from patients, and doctors will use purpose-built AI models to extract clinically relevant information about a patient’s tumor and what drugs it will best respond to.

Today we are witnessing the convergence of three forces: spatial omics technologies becoming increasingly high-throughput and high-resolution, large-scale datasets from patient biopsies being generated, and AI models becoming ever more sophisticated. Together, they will allow researchers to dissect the complex biology of health and diseases, enabling ever more sophisticated therapeutic interventions. 

Davide Mantiero, PhD, Joseph Lehár, PhD, and Darius Meadon also contributed to this piece.

This content was produced by Owkin. It was not written by MIT Technology Review’s editorial staff.

Capitalizing on machine learning with collaborative, structured enterprise tooling teams

Advances in machine learning (ML) and AI are emerging on a near-daily basis—meaning that industry, academia, government, and society writ large are evolving their understanding of the associated risks and capabilities in real time. As enterprises seek to capitalize on the potential of AI, it’s critical that they develop, maintain, and advance state-of-the-art ML practices and processes that will offer both strong governance and the flexibility to change as the demands of technology requirements, capabilities, and business imperatives change.

That’s why it’s critical to have strong ML operations (MLOps) tooling, practices, and teams—those that build and deploy a set of software development practices that keep ML models running effectively and with agility. Capital One’s core ML engineering teams demonstrate firsthand the benefits collaborative, well-managed, and adaptable MLOps teams can bring to enterprises in the rapidly evolving AI/ML space. Below are key insights and lessons learned during Capital One’s ongoing technology and AI journey.

Standardized, reusable components are critical

Most MLOps teams have people with extensive software development skills who love to build things. But the continuous build of new AI/ML tools must also be balanced with risk efficiency, governance, and risk mitigation.

Many engineers today are experimenting with new generative AI capabilities. It’s exciting to think about the possibilities that something like code generation can unlock for efficiency and standardization, but auto-generated code also requires sophisticated risk management and governance processes before it can be accepted into any production environment. Furthermore, a one-size-fits-all approach to things like generating code won’t work for most companies, which have industry, business, and customer-specific circumstances to account for.

As enterprise platform teams continue to explore the evolution of ML tools and techniques while prioritizing reusable tools and components, they can look to build upon open-source capabilities. One example is Scikit-Learn, a Python library containing numerous supervised and unsupervised learning algorithms that has a strong user community behind it and which can be used as a foundation to further customize for specific and reusable enterprise needs.

Cross-team communication is vital

Most large enterprises have data scientists and engineers working on projects through different parts of the company. This means it can also be difficult to know where new technologies and tools are built, resulting in arbitrary uniqueness.

This underscores the importance of creating a collaborative team culture where communication about the big picture, strategic goals, and initiatives is prioritized—including the ability to find out where tools are being built and evolved. What does this look like in practice?

Ensure your team knows what tools and processes it owns and contributes to. Make it clear how their work supports the broader company’s mission. Demonstrate how your team can feel empowered not to build something from scratch. Incentivize reuse and standardization. It takes time and effort to create a culture of “innersourcing” innovation and build communications mechanisms for clarity and context, but it’s well worth it to ensure long-term value creation, innovation, and efficiency.

Tools must map to business outcomes

Enterprise MLOps teams have a broader role than building tools for data scientists and engineers: they need to ensure those tools both mitigate risk and enable more streamlined, nimble technology capabilities for their business partners. Before setting off on building new AI/ML capabilities, engineers and their partners should ask themselves a few core questions. Does this tool actually help solve a core problem for the business? Will business partners be able to use it? Will it work with existing tools and processes? How quickly can we deliver it, and is there something similar that already exists that we should build upon first?

Having centralized enterprise MLOps and engineering teams ask these questions can free up the business to solve customer problems, and to consider how technology can continue to support the evolution of new solutions and experiences.

Don’t simply hire unicorns, build them

There’s no question that delivering for the needs of business partners in the modern enterprise takes significant amounts of MLOps expertise. It requires both software engineering and ML engineering experience, and—especially as AI/ML capabilities evolve—people with deeply specialized skill sets, such as those with deep graphics processing (GPU) expertise.

Instead of hiring a “unicorn” individual, companies should focus on building a unicorn team with the best of both worlds. This means having deep subject matter experts in science, engineering, statistics, product management, DevOps, and other disciplines. These are all complementary skill sets that add up to a more powerful collective. Together, individuals who can work effectively as a team, show a curiosity for learning, and an ability to empathize with the problems you’re solving are just as important as their unique domain skills.

Develop a product mindset to produce better tools

Last but not least, it’s important to take a product-backed mindset when building new AI and ML tools for internal customers and business partners. It requires not just thinking about what you build as just a task or project to be checked off the list, but understanding the customer you’re building for and taking a holistic approach that works back from their needs.

Often, the products MLOps teams build—whether it’s a new feature library or an explainability tool—look different than what traditional product managers deliver, but the process for creating great products should be the same. Focusing on the customer needs and pain points helps everyone deliver better products; it’s a muscle that many data science and engineering experts have to build, but ultimately helps us all create better tooling and deliver more value for the customer.

The bottom line is that today, the most effective MLOps strategies are not just about technical capabilities, but also involve intentional and thoughtful culture, collaboration, and communication strategies. In large enterprises, it’s important to be cognizant that no one operates in a vacuum. As hard as it may be to see in the day-to-day, everything within the enterprise is ultimately connected, and the capabilities that AI/ML tooling and engineering teams bring to bear have important implications for the entire organization.

This content was produced by Capital One. It was not written by MIT Technology Review’s editorial staff.

Sustainability starts with the data center

When asked why he targeted banks, notorious criminal Willie Sutton reportedly answered, “Because that’s where the money is.” Similarly, when thoughtful organizations target sustainability, they look to their data centers—because that’s where the carbon emissions are.

The International Energy Agency (IEA) attributes about 1.5% of total global electricity use to data centers and data transmission networks. This figure is much higher, however, in countries with booming data storage sectors: in Ireland, 18% of electricity consumption was attributable to data centers in 2022, and in Denmark, it is projected to reach 15% by 2030. And while there have been encouraging shifts toward green-energy sources and increased deployment of energy-efficient hardware and software, organizations need to accelerate their data center sustainability efforts to meet ambitious net-zero targets.

For data center operators, options for boosting sustainability include shifting energy sources, upgrading physical infrastructure and hardware, improving and automating workflows, and updating the software that manages data center storage. Hitachi Vantara estimates that emissions attributable to data storage infrastructure can be reduced as much as 96% by using a combination of these approaches.

Critics might counter that, though data center decarbonization is a worthy social goal, it also imposes expenses that a company focused on its bottom line can ill afford. This, however, is a shortsighted view.

Data center decarbonization initiatives can provide an impetus that enables organizations to modernize, optimize, and automate their data centers. This leads directly to improved performance of mission-critical applications, as well as a smaller, denser, more efficient data center footprint—which then creates savings via reduced energy costs. And modern data storage and management solutions, beyond supporting sustainability, also create a unified platform for innovation and new business models through advanced data analytics, machine learning, and AI.

Dave Pearson, research vice president at IDC, says, “Decarbonization and the more efficient energy utilization of the data center are supported by the same technologies that support data center modernization. Modernization has sustainability goals, but obviously it provides all kinds of business benefits, including enabling data analytics and better business processes.”

Download the full report.

This content was produced by Insights, the custom content arm of MIT Technology Review. It was not written by MIT Technology Review’s editorial staff.

Augmenting the realities of work

Imagine an integrated workplace with 3D visualizations that augment presentations, interactive and accelerated onboarding, and controlled training simulations. This is the future of immersive technology that global head of Immersive Technology Research at JPMorgan Chase, Blair MacIntyre is working to build. Augmented reality (AR) and virtual reality (VR) technologies can blend physical and digital dimensions together and infuse new innovations and efficiencies into business and customer experiences.

“These technologies can offer newer ways of collaborating over distance both synchronously and asynchronously than we can get with the traditional work technologies that we use right now,” says MacIntyre. “It’s these new ways to collaborate, ways of using the environment and space in new and interesting ways that will hopefully offer new value and change the way we work.”

Many enterprises are integrating VR into business practices like video conference calls. But having some participants in a virtual world and some sidelined creates imbalances in the employee experience. MacIntyre’s team is looking for ways to use AR/VR technologies that can be additive, like 3D data visualizations that enhance financial forecasting within a bank, not ones that overhaul entire experiences.

Although the potential of AR/VR is quickly evolving, it’s unlikely that customers’ interactions or workplace environments will be entirely moved to the virtual world anytime soon. Rather, MacIntyre’s immersive technology research looks to infuse efficiencies into existing practices.

“It’s thinking about how the technologies integrate and how we can add value where there is value and not trying to replace everything we do with these technologies,” MacIntyre says.

AI can help remove some of the tedium from immersive technologies that have made them impractical for widespread enterprise use in the past. Using VR technology in the workplace may prohibit taking notes and having access to traditional input devices and files. AI tools can take and transcribe notes and fill in any other gaps to help remove that friction and eliminate redundancies.

Connected Internet of things (IoT) devices are also key to enabling AR/VR technologies. To create a valuable immersive experience, MacIntyre says, it’s imperative to know as much about the surrounding world of the user as well as their needs, habits, and preferences.

“If we can figure out more ways of enabling people to work together in a distributed way, we can start enabling more people to participate meaningfully in a wider variety of jobs,” says MacIntyre.

This episode of Business Lab is produced in association with JPMorgan Chase.

Full transcript

Laurel: From MIT Technology Review, I’m Laurel Ruma, and this is Business Lab, the show that helps business leaders make sense of new technologies coming out of the lab and into the marketplace.

Our topic today is emerging technologies, specifically, immersive technologies like augmented and virtual reality. Keeping up with technology trends may be a challenge for most enterprises, but it’s a critical way to think about future possibilities from product to customer service to employee experience. Augmented and virtual realities aren’t necessarily new, but when it comes to applying them beyond gaming, it’s a brave new world.

Two words for you: emerging realities.

My guest is Blair MacIntyre, who is the global head of Immersive Technology Research at JPMorgan Chase.

This podcast is produced in association with JPMorgan Chase.

Welcome, Blair.

Blair MacIntyre: Thank you. It’s great to be here.

Laurel: Well, let’s do a little bit of context setting. Your career has been focused on researching and exploring immersive technology, including software and design tools, privacy and ethics, and game and experience design. So what brought you to JPMorgan Chase, and could you describe your current role?

Blair: So before joining the firm, I had spent the last 23 years as a professor at Georgia Tech and Northeastern University. During that time, as you say, I explored a lot of ways that we can both create things with these technologies, immersive technologies and also, what they might be useful for and what the impacts on people in society and how we experience life are. But as these technologies have become more real, moved out of the lab, starting to see real products from real companies, we have this opportunity to actually see how they might be useful in practice and to have, for me, an impact on how these technologies will be deployed and used that goes beyond the traditional impact that professors might have. So beyond writing papers, beyond teaching students. That’s what brought me to the firm, and so my current role is, really, to explore that, to understand all the different ways that immersive technology could impact the firm and its customers. Right? So we think about not just customer-facing and not just products, but also employees and their experience as well.

Laurel: That’s really interesting. So why does JPMorgan Chase have a dedicated immersive technology focus in its global technology applied research division, and what are the primary goals of your team’s research within finance and large enterprises as a whole?

Blair: That’s a great question. So JPMorgan Chase has a fairly wide variety of research going on within the company. There’s large efforts in AI/ML, in quantum computing, blockchain. So they’re interested in looking at all of the range of new technologies and how they might impact the firm and our customers, and immersive technologies represent one of those technologies that could over time have a relatively large impact, I think, especially on the employee experience and how we interact with our customers. So they really want to have a group of people focusing on, really, looking both in the near and long term, and thinking about how we can leverage the technology now and how we might be able to leverage it down the road, and not just how we can, but what we should not do. Right? So we’re interested in understanding of these applications that are being proposed or people are imagining could be used. Which ones actually have value to the company, and which ones may not actually have value in practice?

Laurel: So when people think of immersive technologies like augmented reality and virtual reality, AR and VR, many think of headsets or smartphone apps for gaming and retail shopping experiences. Could you give an overview of the state of immersive technology today and what use cases you find to be the most innovative and interesting in your research?

Blair: So, as you say, I think many people think about smartphones, and we’ve seen, at least in movies and TV shows, head mounts of various kinds. The market, I would divide it right now into the two parts, the handheld phone and tablet experience. So you can do augmented reality now, and that really translates to we take the camera feed, and we can overlay computer graphics on it to do things like see what something you might want to buy looks like in your living room or do, in an enterprise situation, remote maintenance assistance where I can take my phone, point it at a piece of technology, and a remote expert could draw on it or help me do something with it.

There’s the phone-based things, and we carry these things in our pockets all the time, and they’re relatively cheap. So there’s a lot of opportunities when it’s appropriate to use those, but the big downside of those devices is that you have to hold them in your hands, so if you wanted to try to put information all around you, you would have to hold the device up and look around, which is uncomfortable and awkward. So that is where the head mount displays come in.

So either virtual reality displays which, right now, many of us think about computer games and education as use cases in the consumer world or augmented reality displays. These sorts of displays now let us do the same kind of things we might do with our phones, but we can do it without our hands having to hold something so we can be doing whatever work it was we wanted to do, right? Repairing the equipment, taking notes, working with things in the world around us, and we can have information spread all around us, which I think is the big advantage of head mounts.

So many of the things people imagine when they think about augmented reality in particular involve this serendipitous access to information. I’m walking into a conference room, and I see sort of my notes and information about the people I’m meeting there and the materials from our last meeting, whatever it is, or I’m walking down the street, and I see advertising or other kinds of, say, tourism information, but those things only work if the device is out of mind. If I can put it on, and then go about my life, I’m not going to walk into a conference room, and hold up a phone, and look at everybody through it.

So that, I think, is the big difference. You could implement the same sorts of applications on both the handheld devices and the head-worn devices, but the two different form factors are going to make very different applications appropriate for those two sorts of technologies.

On the virtual reality side, we’re at the point now where the displays we can buy are light enough and comfortable enough that we could wear them for half an hour, a couple hours without discomfort. So a lot of the applications that people imagine there, I think the most popular things that people have done research on and that I see having a near-term impact in the enterprise are immersive training applications where you can get into a situation rather than, say, watching a video or a little click-through presentation as part of your annual training. You could really be in an experience and hopefully learn more from it. So I think those sorts of experiences where we’re totally immersed and focused is where virtual reality comes in.

The big thing that I think is most exciting about head-worn displays in particular where we can wear them while we’re doing work as opposed to just having these ephemeral experiences with a phone is the opportunity to do things together, to collaborate. So I might want to look at a map on a table and see a bunch of data floating above the map, but it would be better if you and our other colleagues were around the table with me, and we can all see the same things, or if we want to take a training experience, I could be in there getting my training experience, but maybe someone else is joining me and being able to both offer feedback or guidance and so on.

Essentially, when I think about these technologies, I think about the parallels to how we do work regularly, right? We generally collaborate with people. We might grab a colleague and have them look at our laptop to show them something. I might send someone something on my phone, and then we can talk about it. So much of what we do involves interactions with other people and with the data that we are doing our job with that anything we do with these immersive technologies is really going to have to mimic that and give us the ability to do our real work in these immersive spaces with the people that we normally work with.

Laurel: Well, speaking of working with people, how can the scale of an institution like JPMorgan Chase help propel this research forward in immersive technology, and what opportunities does it provide that are otherwise limited in a traditional university or startup research environment?

Blair: I think it comes down to a few different things. On one hand, we have the access to people who are really doing the things that we want to build technologies to help with. Right? So if I wanted to look how I could use immersive visualization of data to help people in human resources do planning or help people who are doing financial modeling look at the data in new and interesting ways, now I could actually do the research in conjunction with the real people who do that work. Right? So I’ve already and I’ve been at the firm for a little over a year, and many conversations we’ve had were either we’ve had an idea or somebody has come to us with an idea. Through the course of the conversations, relatively quickly, we hone in on things that are much more sophisticated, much more powerful than what we might have thought of at a university where we didn’t have that sort of direct access to people doing the work.

On the other hand, if we actually build something, we can actually test it with the same people, which is an amazing opportunity. Right? When I go to a conference, we’re going to put 20 people who actually represent the real users of those systems. So, for me, that’s where I think the big opportunity of doing research in an enterprise is, is building solutions for the real people of that enterprise and being able to test it with those people.

Laurel: Recent years have actually changed what customers and employees expect from enterprises as well, like omnichannel retail experiences. So immersive technologies can be used to bridge gaps between physical and virtual environments as you were saying earlier. What are the different opportunities that AR and VR can offer enterprises, and how can these technologies be used to improve employee and customer experience?

Blair: So I alluded back to some of that in previous answers. I think the biggest opportunities have to do with how employees within the organization can do new things together, can interact, and also how companies can interact with customers. Now, we’re not going to move all of our interactions with our customers into the virtual world, or the metaverse, or whatever you want to call it nowadays anytime soon. Right? But I think there are opportunities for customers who are interested in those technologies, and comfortable with them, and excited by them to get new kinds of experiences and new ways of interacting with our firm or other firms than you could get with webpages and in-person meetings.

The other big opportunity I think is as we move to a more hybrid work environment and a distributed work environment, so a company like JPMorgan Chase is huge and spread around the world. We have over 300,000 employees now in most countries around the world. There might be groups of people, but they’re connected together through video right now. These technologies, I think, can offer new ways of collaborating over distance both synchronously and asynchronously than we can get with the traditional work technologies that we use right now. So it’s those new ways to collaborate, ways of using the environment and space in new and interesting ways that is going to, hopefully, offer new value and change the way we work.

Laurel: Yeah, and staying on that topic, we can’t really have a discussion about technology without talking about AI which is another evolving, increasingly popular technology. So that’s being used by many enterprises to reduce redundancies and automate repetitive tasks. In this way, how can immersive technology provide value to people in their everyday work with the help of AI?

Blair: So I think the big opportunity that AI brings to immersive technologies is helping ease a lot of the tedium and burden that may have prevented these technologies from being practical in the past, and this could happen in a variety of ways. When I’m in a virtual reality experience, I don’t have access to a keyboard, I don’t have access to traditional input devices, I don’t have necessarily the same sorts of access to my files, and so on. With a lot of the new AI technologies that are coming around, I can start relying on the computer to take notes. I can have new ways of pulling up information that I otherwise wouldn’t have access to. So, I think AI reducing the friction of using these technologies is a huge opportunity, and the research community is actively looking at that because friction has been one of the big problems with these technologies up till now.

Laurel: So, other than AI, what are other emerging technologies that can aid in immersive technology research and development?

Blair: So, aside from AI, if we step back and look at all of the emerging technologies as a whole and how they complement each other, I think we can see new opportunities. So, in our research, we work closely with people doing computer vision and other sort of sensing research to understand the world. We work closely with people looking at internet of things and connected devices because at a 10,000-foot level, all of these technologies are based on the idea of understanding, sensing the world, understanding what people are doing in it, understanding what people’s needs might be, and then somehow providing information to them or actuating things in the world, displaying stuff on walls or displays.

From that viewpoint, immersive technologies are primarily one way of displaying things in a new and interesting way and getting input from people, knowing what people want to do, allowing them to interact with data. But in order to do that, they need to know as much about the world around the user as possible, the structure of it, but also, who’s there, what we are doing, and so on. So all of these other technologies, especially the Internet of things (IoT) and other forms and ways of sensing what’s happening in the world are very complimentary and together can create new sorts of experiences that neither could do alone.

Laurel: So what are some of the challenges, but also, possible opportunities in your research that contrast the future potential of AR and VR to where the technology is today?

Blair: So I think one of the big limitations of technology today is that most of the experiences are very siloed and disconnected from everything else we do. During the pandemic, many of us experimented with how we could have conferences online in various ways, right? A lot of companies, small companies and larger companies, started looking at how you could create immersive meetings and big group experiences using virtual reality technology, but all of those experiences that people created were these closed systems that you couldn’t bring things into. So one of the things we’re really interested in is how we stop thinking about creating new kinds of experiences and new ways of doing things, and instead think about how do we add these technologies to our existing work practices to enhance them in some way.

So, for example. Right now, we do video meetings. It would be more interesting for some people to be able to join those meetings, say, in VR. Companies have experimented with that, but most of the experiments that people are doing assume that everyone is going to move into virtual reality, or we’re going to bring, say, the people in as a little video wall on the side of a big virtual reality room, making them second class citizens.

I’m really interested and my team is interested in how we can start incorporating technologies like this while keeping everyone a first-class participant in these meetings. As one example, a lot of the systems that large enterprises build, and we’re no different, are web-based right now. So if, let’s say, I have a system to do financial forecasting, you could imagine there’s a bunch of those at a bank, and it’s a web-based system, I’m really interested in how do we add the ability for people to go into a virtual reality or augmented reality experience, say, a 3D visualization of some kind of data at the moment they want to do it, do the work that they want to do, invite colleagues in to discuss things, and then go back to the work as it was always done on a desktop web browser.

So that idea of thinking of these technologies as a capability, a feature instead of a new whole application and way of doing things permeates all the work we’re doing. When I look down the road at where this can go, I see in, say, let’s say, two to five years, I see people with displays maybe sitting on their desk. They have their tablet and their phone, and they might also have another display or two sitting there. They’re doing their work, and at different times, they might be in a video chat, they might pick up a head mount and put it on to do different things, but it’s all integrated. I’m really interested in how we connect these together and reduce friction. Right? If it takes you four or five minutes to move your work into a VR experience, nobody is going to do it because it just is too problematic. So it’s that. It’s thinking about how the technologies integrate and how we can add value where there is value and not trying to replace everything we do with these technologies.

Laurel: So to stay on that future focus, how do you foresee the immersive technology landscape entirely evolving over the next decade, and how will your research enable those changes?

Blair: So, at some level, it’s really hard to answer that question. Right? So if I think back 10 years to where immersive technologies were, it would have been inconceivable for us to imagine the videos that are coming out. So, at some level, I can say, “Well, I have no idea where we’re going to be in 10 years.” On the other hand, it’s pretty safe to imagine the kinds of technologies that we’re experimenting with now just getting better, and more comfortable, and more easy to integrate into work. So I think the landscape is going to evolve in the near term to be more amenable to work.

Especially for augmented reality, the threshold that these devices would have to get to such that a lot of people would be willing to wear them all the time while they’re walking down the street, playing sports, doing whatever, that’s a very high bar because it has to be small, it has to be light, it has to be cheap, it has to have a battery that lasts all day, etcetera, etcetera. On the other hand, in the enterprise, in any business situation, it’s easy to imagine the scenario I described. It’s sitting on my desk, I pick it up, I put it on, I take it off.

In the medium term after that, I think we will see more consumer applications as people start solving more of the problems that are preventing people from wearing these devices for longer periods of time. Right? It’s not just size, and battery power, and comfort, it’s also things like optics. Right? A lot of people — not a lot, but say, let’s say 10%, 15% of people might experience headaches, or nausea, or other kinds of discomfort when they wear a VR display as they’re currently built, and a lot of that has to do with the fact that the optics that you’re looking at when you’re putting this display are built in a way that makes it hard to comfortably focus at objects at different distances away from you without getting into the nitty-gritty details. For many of us, that’s fine. We can deal with the slight problems. But for some people, it’s problematic.

So as we figure out how to solve problems like that, more people can wear them, and more people can use them. I think that’s a really critical issue for not just consumers, but for the enterprise because if we think about a future where more of our business applications and the kind of way we work are done with technologies like this, these technologies have to be accessible to everybody. Right? If that 10% or 15% of people get headaches and feel nauseous wearing this device, you’ve now disenfranchised a pretty significant portion of your workforce, but I think those can be solved, and so we need to be thinking about how we can enable everybody to use them.

On the other hand, technologies like this can enfranchise more people, where right now, working remotely, working in a distributed sense is hard. For many kinds of work, it’s difficult to do remotely. If we can figure out more ways of enabling people to work together in a distributed way, we can start enabling more people to participate meaningfully in a wider variety of jobs.

Laurel: Blair, that was fantastic. It’s so interesting. I really appreciate your perspective and sharing it here with us on the Business Lab.

Blair: It was great to be here. I enjoyed talking to you.

Laurel: That was Blair MacIntyre, the global head of Immersive Technology Research at JPMorgan Chase, who I spoke with from Cambridge, Massachusetts, the home of MIT and MIT Technology Review.

That’s it for this episode of Business Lab. I’m your host, Laurel Ruma. I’m the global director of Insights, the custom publishing division of MIT Technology Review. We were founded in 1899 at the Massachusetts Institute of Technology, and you can find us in print, on the web, and at events each year around the world. For more information about us and the show, please check out our website at technologyreview.com.

This show is available wherever you get your podcasts. If you enjoyed this episode, we hope you’ll take a moment to rate and review us. Business Lab is a production of MIT Technology Review. This episode was produced by Giro Studios. Thanks for listening.

This content was produced by Insights, the custom content arm of MIT Technology Review. It was not written by MIT Technology Review’s editorial staff.


This podcast is for informational purposes only and it is not intended as legal, tax, financial, investment, accounting or regulatory advice. Opinions expressed herein are the personal views of the individual(s) and do not represent the views of JPMorgan Chase & Co. The accuracy of any statements, linked resources, reported findings or quotations are not the responsibility of JPMorgan Chase & Co.

Procurement in the age of AI

Procurement professionals face challenges more daunting than ever. Recent years’ supply chain disruptions and rising costs, deeply familiar to consumers, have had an outsize impact on business buying. At the same time, procurement teams are under increasing pressure to supply their businesses while also contributing to business growth and profitability.

Deloitte’s 2023 Global Chief Procurement Officer Survey reveals that procurement teams are now being called upon to address a broader range of enterprise priorities. These range from driving operational efficiency (74% of respondents) and enhancing corporate social responsibility (72%) to improving margins via cost reduction (71%).

To meet these rising expectations, many procurement teams are turning to advanced analytics, AI, and machine learning (ML) to transform the way they make smart business buying decisions and create value for the organization.

New procurement capabilities unlocked by AI

AI and ML tools have long helped procurement teams automate mundane and manual procurement processes, allowing them to focus on more strategic initiatives. But recent advances in natural language processing (NLP), pattern recognition, cognitive analytics, and large language models (LLMs) are “opening up opportunities to make procurement more efficient and effective,” says Julie Scully, director of software development at Amazon Business.

The good news is procurement teams are already well-positioned to capitalize on these technological advances. Their access to rich data sources, ranging from contracts to invoices, enables AI/ML solutions that can illuminate the insights contained within this data. Acting on these insights unlocks new capabilities that can enhance decision-making and improve spending patterns across the organization.

Predicting supply chain disruptions. In an era of constant supply chain disruptions, procurement teams are often faced with inconsistent item availability, which can negatively impact employee and customer experience. Indeed, the Deloitte 2023 Global Chief Procurement Officer survey finds that only 25% of firms are able to identify supply disruptions promptly “to a large extent.”

AI tools can help address this issue by recognizing patterns that indicate an emerging supply shortage and automatically recommending two or three product alternatives to business buyers, thereby preventing supply disruptions. These predictive capabilities also empower procurement teams to establish buying policies that proactively account for items that are more likely to go out of stock.

Answering pressing questions quickly. Sifting through data to understand the cause of a supply chain disruption, product defect, or other risk is time-consuming for a procurement professional. LLM-powered chatbots can streamline these processes by understanding complex queries about orders and “putting together a nuanced answer,” says Scully. “AI can query a wide variety of sources to fully answer a question quickly and in a way that feels natural and understandable.” In addition to providing fast and accurate answers to pressing questions, AI promises to reduce the need to explain procurement issues eventually. Instead, it will proactively analyze orders, buying patterns, and the current situation to provide instant support.

Offering customized recommendations. As business buyers increasingly demand personalized experiences, procurement officers seek ways to customize their interactions with business procurement systems. Scully provides the example of an employee tasked with hosting a holiday party for 150 employees who needs help deciding what to order. An AI-based procurement tool posed that scenario, she says, could generate a proposed shopping cart, sifting through “millions and billions of data points to recommend and suggest items that the employee may not have even thought of.”

Better yet, she adds, “as we get into really large language models, AI/ML can help answer questions or help buy items you didn’t even know you needed by understanding your particular situation in a much more detailed way.”

Influencing compliance spend. Procurement professionals aim to balance employees’ freedom to purchase the items they need with minimal intervention. However, self-sufficiency should not come at the cost of proper spend management, productivity, and policy compliance. “There’s always a healthy tension between how a company ensures they have the right controls and oversight but also enables a federated spend model,” says Scully. Fortunately, she says, “AI can offer huge value” in alerting procurement teams to any “outliers” before any damage is done.

AI can also help ensure compliance by enforcing spending policies and expectations so that employees “can still confidently buy the right items,” says Scully. This capability can minimize the risk of overspending and also help with companies’ contractual obligations, such as fulfilling a spending commitment to a particular supplier. In the future, an AI-powered anomaly detection trigger might even be used to examine large datasets to identify non-compliant purchases.

Increasing spending visibility. AI and analytics tools can provide greater transparency into overall procurement spending by automatically analyzing data and unlocking timely analysis. These data-driven insights provide procurement officers a comprehensive view of where they’re allocating budget and areas where they might be able to cut costs.

But greater transparency into procurement spend can also empower organizations to respond to emerging business priorities, such as adopting more socially responsible purchasing practices. “Companies want to prioritize locally owned businesses or businesses that prioritize a lower carbon footprint,” says Scully. With greater visibility into their procurement patterns, organizations can direct business buyers to climate-friendly products or suppliers that help meet  their environmental, social, and governance goals.

Driving procurement productivity. Monitoring supplier performance, ensuring spend compliance, and identifying supply chain disruptions—these are all time-consuming activities that distract procurement professionals from more business-critical objectives. “If the procurement team is bogged down in day-to-day processes, they can’t be thinking about their overall strategic goals for the company, if they’re able to deliver them, and where they might want to provide optimizations,” says Scully. By automating labor-intensive processes such as spend analysis, product selection, and tracking down orders, advanced procurement tools can free procurement teams to focus on value-added activities.

Best practices for AI-powered procurement

For all the advantages of advanced analytics and AI/ML solutions, procurement teams must take steps to ensure the best use of these innovative tools. AI models are only as relevant as the training data they ingest. For this reason, Scully says, organizations need “to be aware that a model may sometimes have blind spots or not immediately recognize if the business is beginning a change in strategic focus.” As an organization’s priorities evolve, the model training data must keep pace to reflect new business goals and circumstances.

To get the most from its advanced technology tools, procurement teams should ensure that they support the company’s overall procurement goals and business strategy. These goals may range from working with a more diverse supplier base to purchasing more sustainable goods. Whatever the desired end, the procurement function must link its use of new AI-powered tools to achieving its business goals and regularly evaluate the results.

The new procurement capabilities unlocked by advanced analytics and AI/ML can help businesses rethink how procurement gets done. As generative AI and related technologies advance, sophisticated procurement use cases are likely to multiply, offering substantial financial and operational gains to procurement teams.

This content was produced by Insights, the custom content arm of MIT Technology Review. It was not written by MIT Technology Review’s editorial staff.

Learn how Amazon Business is leveraging AI/ML to offer procurement professionals more efficient processes, a greater understanding of smart business buying habits and, ultimately, reduced prices.

Finding value in generative AI for financial services

With tools such as ChatGPT, DALLE-2, and CodeStarter, generative AI has captured the public imagination in 2023. Unlike past technologies that have come and gone—think metaverse—this latest one looks set to stay. OpenAI’s chatbot, ChatGPT, is perhaps the best-known generative AI tool. It reached 100 million monthly active users in just two months after launch, surpassing even TikTok and Instagram in adoption speed, becoming the fastest-growing consumer application in history.

According to a McKinsey report, generative AI could add $2.6 trillion to $4.4 trillion annually in value to the global economy. The banking industry was highlighted as among sectors that could see the biggest impact (as a percentage of their revenues) from generative AI. The technology “could deliver value equal to an additional $200 billion to $340 billion annually if the use cases were fully implemented,” says the report. 

For businesses from every sector, the current challenge is to separate the hype that accompanies any new technology from the real and lasting value it may bring. This is a pressing issue for firms in financial services. The industry’s already extensive—and growing—use of digital tools makes it particularly likely to be affected by technology advances. This MIT Technology Review Insights report examines the early impact of generative AI within the financial sector, where it is starting to be applied, and the barriers that need to be overcome in the long run for its successful deployment. 

The main findings of this report are as follows:

  • Corporate deployment of generative AI in financial services is still largely nascent. The most active use cases revolve around cutting costs by freeing employees from low-value, repetitive work. Companies have begun deploying generative AI tools to automate time-consuming, tedious jobs, which previously required humans to assess unstructured information.
  • There is extensive experimentation on potentially more disruptive tools, but signs of commercial deployment remain rare. Academics and banks are examining how generative AI could help in impactful areas including asset selection, improved simulations, and better understanding of asset correlation and tail risk—the probability that the asset performs far below or far above its average past performance. So far, however, a range of practical and regulatory challenges are impeding their commercial use.
  • Legacy technology and talent shortages may slow adoption of generative AI tools, but only temporarily. Many financial services companies, especially large banks and insurers, still have substantial, aging information technology and data structures, potentially unfit for the use of modern applications. In recent years, however, the problem has eased with widespread digitalization and may continue to do so. As is the case with any new technology, talent with expertise specifically in generative AI is in short supply across the economy. For now, financial services companies appear to be training staff rather than bidding to recruit from a sparse specialist pool. That said, the difficulty in finding AI talent is already starting to ebb, a process that would mirror those seen with the rise of cloud and other new technologies.
  • More difficult to overcome may be weaknesses in the technology itself and regulatory hurdles to its rollout for certain tasks. General, off-the-shelf tools are unlikely to adequately perform complex, specific tasks, such as portfolio analysis and selection. Companies will need to train their own models, a process that will require substantial time and investment. Once such software is complete, its output may be problematic. The risks of bias and lack of accountability in AI are well known. Finding ways to validate complex output from generative AI has yet to see success. Authorities acknowledge that they need to study the implications of generative AI more, and historically they have rarely approved tools before rollout.

Download the full report.

This content was produced by Insights, the custom content arm of MIT Technology Review. It was not written by MIT Technology Review’s editorial staff.