Sign In or Create an Account.

By continuing, you agree to the Terms of Service and acknowledge our Privacy Policy

Technology

What Does OpenAI’s New Breakthrough Mean for Energy Consumption?

Why the new “reasoning” models might gobble up more electricity — at least in the short term

A robot with a smokestack coming out of its head.
Heatmap Illustration/Getty Images

What happens when artificial intelligence takes some time to think?

The newest set of models from OpenAI, o1-mini and o1-preview, exhibit more “reasoning” than existing large language models and associated interfaces, which spit out answers to prompts almost instantaneously.

Instead, the new model will sometimes “think” for as long as a minute or two. “Through training, they learn to refine their thinking process, try different strategies, and recognize their mistakes,” OpenAI announced in a blog post last week. The company said these models perform better than their existing ones on some tasks, especially related to math and science. “This is a significant advancement and represents a new level of AI capability,” the company said.

But is it also a significant advancement in energy usage?

In the short run at least, almost certainly, as spending more time “thinking” and generating more text will require more computing power. As Erik Johannes Husom, a researcher at SINTEF Digital, a Norwegian research organization, told me, “It looks like we’re going to get another acceleration of generative AI’s carbon footprint.”

Discussion of energy use and large language models has been dominated by the gargantuan requirements for “training,” essentially running a massive set of equations through a corpus of text from the internet. This requires hardware on the scale of tens of thousands of graphical processing units and an estimated 50 gigawatt-hours of electricity to run.

Training GPT-4 cost “more than” $100 million OpenAI chief executive Sam Altman has said; the next generation models will likely cost around $1 billion, according to Anthropic chief executive Dario Amodei, a figure that might balloon to $100 billion for further generation models, according to Oracle founder Larry Ellison.

While a huge portion of these costs are hardware, the energy consumption is considerable as well. (Meta reported that when training its Llama 3 models, power would sometimes fluctuate by “tens of megawatts,” enough to power thousands of homes). It’s no wonder that OpenAI’s chief executive Sam Altman has put hundreds of millions of dollars into a fusion company.

But the models are not simply trained, they're used out in the world, generating outputs (think of what ChatGPT spits back at you). This process tends to be comparable to other common activities like streaming Netflix or using a lightbulb. This can be done with different hardware and the process is more distributed and less energy intensive.

As large language models are being developed, most computational power — and therefore most electricity — is used on training, Charlie Snell, a PhD student at University of California at Berkeley who studies artificial intelligence, told me. “For a long time training was the dominant term in computing because people weren’t using models much.” But as these models become more popular, that balance could shift.

“There will be a tipping point depending on the user load, when the total energy consumed by the inference requests is larger than the training,” said Jovan Stojkovic, a graduate student at the University of Illinois who has written about optimizing inference in large language models.

And these new reasoning models could bring that tipping point forward because of how computationally intensive they are.

“The more output a model produces, the more computations it has performed. So, long chain-of-thoughts leads to more energy consumption,” Husom of SINTEF Digital told me.

OpenAI staffers have been downright enthusiastic about the possibilities of having more time to think, seeing it as another breakthrough in artificial intelligence that could lead to subsequent breakthroughs on a range of scientific and mathematical problems. “o1 thinks for seconds, but we aim for future versions to think for hours, days, even weeks. Inference costs will be higher, but what cost would you pay for a new cancer drug? For breakthrough batteries? For a proof of the Riemann Hypothesis? AI can be more than chatbots,” OpenAI researcher Noam Brown tweeted.

But those “hours, days, even weeks” will mean more computation and “there is no doubt that the increased performance requires a lot of computation,” Husom said, along with more carbon emissions.

But Snell told me that might not be the end of the story. It’s possible that over the long term, the overall computing demands for constructing and operating large language models will remain fixed or possibly even decline.

While “the default is that as capabilities increase, demand will increase and there will be more inference,” Snell told me, “maybe we can squeeze reasoning capability into a small model ... Maybe we spend more on inference but it’s a much smaller model.”

OpenAI hints at this possibility, describing their o1-mini as “a smaller model optimized for STEM reasoning,” in contrast to other, larger models that “are pre-trained on vast datasets” and “have broad world knowledge,” which can make them “expensive and slow for real-world applications.” OpenAI is suggesting that a model can know less but think more and deliver comparable or better results to larger models — which might mean more efficient and less energy hungry large language models.

In short, thinking might use less brain power than remembering, even if you think for a very long time.

Blue

You’re out of free articles.

Subscribe today to experience Heatmap’s expert analysis 
of climate change, clean energy, and sustainability.
To continue reading
Create a free account or sign in to unlock more free articles.
or
Please enter an email address
By continuing, you agree to the Terms of Service and acknowledge our Privacy Policy
Energy

Exclusive: Japan’s Tiny Nuclear Reactors Are Headed to Texas

The fourth-generation gas-cooled reactor company ZettaJoule is setting up shop at an unnamed university.

A Texas sign at a ZettaJoule facility.
Heatmap Illustration/Getty Images, ZettaJoule

The appeal of next-generation nuclear technology is simple. Unlike the vast majority of existing reactors that use water, so-called fourth-generation units use coolants such as molten salt, liquid metal, or gases that can withstand intense heat such as helium. That allows the machines to reach and maintain the high temperatures necessary to decarbonize industrial processes, which currently only fossil fuels are able to reach.

But the execution requirements of these advanced reactors are complex, making skepticism easy to understand. While the U.S., Germany, and other countries experimented with fourth-generation reactors in earlier decades, there is only one commercial unit in operation today. That’s in China, arguably the leader in advanced nuclear, which hooked up a demonstration model of a high-temperature gas-cooled reactor to its grid two years ago, and just approved building another project in September.

Keep reading...Show less
Blue
Spotlight

The 5 Fights to Watch in 2026

Spoiler: A lot of them are about data centers.

Data centers and clean energy.
Heatmap Illustration/Getty Images

It’s now clear that 2026 will be big for American energy, but it’s going to be incredibly tense.

Over the past 365 days, we at The Fight have closely monitored numerous conflicts over siting and permitting for renewable energy and battery storage projects. As we’ve done so, the data center boom has come into full view, igniting a tinderbox of resentment over land use, local governance and, well, lots more. The future of the U.S. economy and the energy grid may well ride on the outcomes of the very same city council and board of commissioners meetings I’ve been reporting on every day. It’s a scary yet exciting prospect.

Keep reading...Show less
Yellow
Hotspots

A Texas Data Center Dispute Turns Tawdry

Plus a resolution for Vineyard Wind and more of the week’s big renewables fights.

The United States.
Heatmap Illustration/Getty Images

1. Hopkins County, Texas – A Dallas-area data center fight pitting developer Vistra against Texas attorney general Ken Paxton has exploded into a full-blown political controversy as the power company now argues the project’s developer had an improper romance with a city official for the host community.

  • For those who weren’t around for the first go, here’s the low-down: The Dallas ex-urb of Sulphur Springs is welcoming a data center project proposed by a relatively new firm, MSB Global. But the land – a former coal plant site – is held by Vistra, which acquired the property in a deal intended for remediating the site. After the city approved the project, Vistra refused to allow construction on the land, so Sulphur Springs sued, and in its bid to win the case, the city received support from Texas attorney general Ken Paxton, whose office then opened an antitrust investigation into the power company’s land holdings.
  • Since we first reported this news, the lawsuit has escalated. Vistra’s attorneys have requested Sulphur Springs’ attorney be removed from the court proceedings because, according to screenshots of lengthy social media posts submitted to the court, the city itself has confirmed that the attorney dated a senior executive for MSB Global as recently as the winter of 2024.
  • In a letter dated December 10, posted online by activists fighting the data center, Vistra’s attorneys now argue the relationship is what led to the data center coming to the city in the first place, and that the attorney cannot argue on behalf of the city because they’ll be a fact witness who may need to provide testimony in the case: “These allegations make awareness of negotiations surrounding the deed and the City’s subsequent conduct post-transaction, including any purported ‘reliance’ on Vistra Parties’ actions and omissions, relevant.”
  • I have not heard back from MSB Global or Sulphur Springs about this case, but if I do, you’ll be hearing about it.

2. La Plata County, Colorado – This county has just voted to extend its moratorium on battery energy storage facilities over fire fears.

Keep reading...Show less
Yellow