You’re out of free articles.
Log in
To continue reading, log in to your account.
Create a Free Account
To unlock more free articles, please create a free account.
Sign In or Create an Account.
By continuing, you agree to the Terms of Service and acknowledge our Privacy Policy
Welcome to Heatmap
Thank you for registering with Heatmap. Climate change is one of the greatest challenges of our lives, a force reshaping our economy, our politics, and our culture. We hope to be your trusted, friendly, and insightful guide to that transformation. Please enjoy your free articles. You can check your profile here .
subscribe to get Unlimited access
Offer for a Heatmap News Unlimited Access subscription; please note that your subscription will renew automatically unless you cancel prior to renewal. Cancellation takes effect at the end of your current billing period. We will let you know in advance of any price changes. Taxes may apply. Offer terms are subject to change.
Subscribe to get unlimited Access
Hey, you are out of free articles but you are only a few clicks away from full access. Subscribe below and take advantage of our introductory offer.
subscribe to get Unlimited access
Offer for a Heatmap News Unlimited Access subscription; please note that your subscription will renew automatically unless you cancel prior to renewal. Cancellation takes effect at the end of your current billing period. We will let you know in advance of any price changes. Taxes may apply. Offer terms are subject to change.
Create Your Account
Please Enter Your Password
Forgot your password?
Please enter the email address you use for your account so we can send you a link to reset your password:
A conversation with Resources for the Future’s David Wear on the fires in the Carolinas and how the political environment could affect the future of forecasting.
The Wikipedia article for “wildfire” has 22 photographs, including those of incidents in Arizona, Utah, Washington, and California. But there is not a single picture of a fire in the American Southeast, despite researchers warning that the lower righthand quadrant of the country will face a “perfect storm” of fire conditions over the next 50 years.
In what is perhaps a grim premonition of what is to come, several major fires are burning across the Southeast now — including the nearly 600-acre Melrose Fire in Polk County, North Carolina, a little over 80 miles to the west of Charlotte, and the more than 2,000-acre Carolina Forest fire in Horry County, South Carolina. The region is also battling hundreds of smaller brush fires, the smoke from which David Wear — the land use, forestry, and agriculture program director at Resources for the Future — could see out his Raleigh-area window.
Wear is also the co-author of a study by RFF and the U.S. Forest Service that came out in late 2024 and singled out the Southeast as facing a “particularly worrisome” rise in wildfire risk over the next half-century. I spoke with him this week to learn more about why the Carolinas are burning and what the future of fire looks like for the region. Our conversation has been edited and condensed for clarity.
When discussing fires in the American West, we often talk about how historic suppression efforts are responsible for the megafires we see today. What was the historic fire regime like in the Southeast? What’s going on to make it a hot spot for wildfires?
First, there are the similarities. Both Western and Southeastern forests, especially pine forests, are fire-adapted systems; they need regular fires to maintain health. Anything that takes those forests out of balance is a problem, and fire suppression is an issue in the East and the West, and especially in the Southeast. But forests in the Southeast are the most heavily managed forests in the country — perhaps in the world. In many cases, they’re regularly burned; the South does more prescribed burning than the rest of the country combined. It’s a very, very common practice in this part of the world.
So we shouldn’t be surprised that there is fire in Southeastern forests. There have been big, episodic fires in the South, though they’re not as common. There was the fire in 2016 in East Tennessee, from the Smokies into Gatlinburg, with a number of fatalities and lots of structures damaged or destroyed. There have been big fire years in east and west Texas. And there have been big fire seasons in Florida, though it’s been a while.
How is population growth in the Southeast adding to the strain?
We’re accustomed to talking about the wildland-urban interface in the West, but it’s also a big issue in the Southeast. Some of our urban growth centers in the Southeast include the Raleigh-Durham area, where I live, and Atlanta, Nashville, and Florida. These are generally flat landscapes, as well as very heavily forested landscapes. As the population grows out of the city centers, they go into pine and mixed-pine hardwood forests that are fire-adapted ecosystems. Then you have interspersed communities with forest vegetation, and that’s a big issue.
I also read in your report that much of that land is privately owned, which makes management tricky.
Private ownership is about 89% of forests in the South. [Editor’s note: By comparison, only about a third of forests in the West are privately owned.] Even where you have public ownership, a lot of that is by the Department of Defense and concentrated in a couple of different areas in the Ozarks and southern Appalachians. Much of the landscape in the coastal plain and Piedmont — which is most of the South — is predominantly private ownership.
There’s a distinction to be made between commercial owners, like timber investment management companies or real estate investment trusts, who actively manage landscapes. With timber harvesting, there are a lot of risk mitigation activities and a lot of prescribed burning. But then you have over a million non-industrial private landowners with small holdings. If you’re trying to coordinate any kind of wildfire mitigation scheme using fuel treatments and the like, it requires some work.
Horry County, South Carolina, and Polk County, North Carolina, were not part of your paper’s list of counties vulnerable to wildfire. I’m curious if you think what we’re seeing now says something about the limits of the study and the data you had available, or if you have another takeaway about what’s going on.
Importantly, our study looked at long-term averages. Throughout the South, there is a fire regime, and in any given year, it is possible to have wildfires of consequence. I would point out that we were especially concerned this year because Hurricane Helene laid down an awful lot of trees and created a fuel load.
We’re also entering one of the two fire seasons in the South. Wildfire is most predominant in the spring and in the fall; it’s at those times when temperatures begin to rise but humidity remains low, and there are extended dry periods that allow the fuels to dry out. You have warm temperatures and wind in the spring, setting the stage for wildfire. Typically, that window will begin to close at the end of April because it’s pretty darn humid in the South at that point, and it’s much less likely that fuels will get dry enough to carry a fire.
The same thing happens in the fall: Temperatures may remain high, and if we don’t have a lot of precipitation and humidity — usually in October and into November — then you have the conditions right for fire. But as the climate shifts, we see the length of those seasons growing to the point where the fall is approaching the spring. Wildfires in January and February indicate that these two seasons are growing toward one another and providing a much longer season. Our paper showed that, when you account for climate change across all of those global climate models and representative concentration pathways, the windows for more wildfire activity and more intense wildfire activity are expanding.
Your paper cited wildfire risks across the Sun Belt. Today, the National Weather Service is warning of “potentially historic” fire conditions in central Texas. Can local emergency managers use your modeling to prepare for such situations?
Things like the year-to-year fire projections and the day-to-day forecasts best serve local emergency managers. Wildfire in the South is determined by the drying of fuels and temperature and humidity conditions, which vary daily. If we look over the last week, Saturday was beautiful in the Carolinas. It was sunny, in the 70s, dry, and a little windy. That was the day [hundreds of] fires started across the Southeast. And the next day, there were very few new fires. Mid-week projections of wildfire potential in the Southeast show that it’s really low, with the exception of Texas. It changes day to day, driven by fine-grain weather forecasts, and that gives emergency managers some insight into where they might want to pre-position crews or do pre-suppression activities.
What we’re doing with the modeling is asking, What is this going to look like in 50 years? The takeaway is that wildfire activity is going to remain strong and perhaps grow in the West, but the big structural change is really strong growth and active fire in the Southeast, where you have wildfire and wildlands proximal to millions of people and more vulnerable communities. It’s a fire regime that’s going to affect more people.
I also wanted to ask about the USDA Forest Service’s contributions to your paper. Do you think research like this could still happen today, given the Trump administration’s efforts to eliminate anything climate-related from the federal agenda?
I came to Resources for the Future six years ago after a long career with the Forest Service, so it’s hard for me to remain a dispassionate scientist here. I think we need to see how the dust settles. It’s hard to imagine a future where the agency and federal government do not have a high level of concern regarding fire — and I don’t think you can do any kind of effective planning, or thinking about the future, or targeting of activities without understanding how climate is likely to impact these disturbance regimes.
I don’t have the crystal ball that many people are seeking right now. We’ll have to wait to see. But our research demonstrates the vital role of understanding climate dynamics, and it shows how critical weather forecasts are for people with boots on the ground who are trying to stay ahead of disaster.
Editor’s note: This story has been updated to reflect that about a third of land in the West is privately owned, not publicly owned.
Log in
To continue reading, log in to your account.
Create a Free Account
To unlock more free articles, please create a free account.
A climate tech company powered by natural gas has always been an odd concept. Now as it moves into developing data centers, it insists it’s remaining true to its roots.
Crusoe Energy has always been a confusing company, whose convoluted green energy credentials raise some eyebrows. It started as a natural gas-powered Bitcoin miner, then became a climate tech unicorn thanks to the fact that its crypto operations utilized waste gas that would have otherwise been flared into the atmosphere. It’s received significant backing from major clean tech investors such as G2 Venture Partners and Lowercarbon Capital. And it touts sustainability as one of its main selling points, describing itself as “on a mission to align the future of computing with the future of the climate,” in part by “harnessing large-scale clean energy.”
But these days, the late-stage startup valued at $2.8 billion makes the majority of its revenue as a modular data center manufacturer and cloud services provider, and is exploring myriad energy solutions — from natural gas to stranded solar and wind assets — beyond its original focus. Earlier this week, it announced that it would acquire more than 4 gigawatts of new natural gas capacity to power its data center buildout. It’s also heavily involved in the Trump-endorsed $500 billion AI push known as the Stargate Project. The company’s Elon Musk-loving CEO Chase Lochmiller told The Information that his team is “pouring concrete at three in the morning” to build out its Stargate Project data centers at “ludicrous speed.”
Some will understandably take a glance at this rising data center behemoth and wonder if climate tech is really an accurate description of what Crusoe actually does these days. As the steady drumbeat of announcements and press surrounding Crusoe’s partnerships and power deals has built up, I certainly wondered whether the company had pivoted to simply churning out data centers as quickly as possible. But investors — and the company itself — told me that’s far from true.
Clay Dumas, a partner at Lowercarbon Capital, which invested in the company’s $128 million Series B and $350 million Series C rounds, told me that Crusoe remains as mission-focused as ever. “When it comes to power, Crusoe is the most aggressive innovator in the AI infrastructure space,” Dumas said via text message. “There is no better team to integrate new energy sources for compute workloads so we don’t turn the whole world into one giant fracking operation.”
Ben Kortlang, a partner at G2 Venture Partners, which led the company’s Series C round, agreed, telling me that Crusoe is best positioned to build out data centers in a way that doesn’t “plant the seeds for 50 or 100 years of environmental damage.”
Yet it’s hard to pin down exactly what the energy mix will end up looking like for the high-profile data centers in Crusoe’s pipeline, including the complex it’s currently building for OpenAI, which is part of the Stargate project in Abilene, Texas. The company announced on Tuesday that it had started construction on the second phase of the facility, which expands the total scope from around 200 megawatts of power across two facilities to include a total of eight buildings over 4 million square feet, using 1.2 gigawatts of power. Crusoe’s spokesperson, Andrew Schmitt, declined to comment on whether this additional capacity would serve Stargate.
What Schmitt did confirm via email is that while the project has a 1.2 gigawatt grid interconnection — enough to meet the entirety of its power needs — Crusoe will also rely on natural gas as “backup energy,” as well as behind-the-meter energy solutions such as solar and battery storage to “create a highly optimized and efficient power plan for the full site.”
The company also won’t speculate on how much energy will come from each particular source. To some degree, the exact grid energy mix and what additional energy resources will get built is unknowable, though Schmitt told me that Crusoe chose Abilene for the area’s abundant wind resources. There’s often too much of it for the grid to handle, meaning the excess energy is curtailed or sold at a negative price. But if a large load — say, a Crusoe data center — were added to the grid, less renewable energy would go to waste, thereby increasing the profitability of renewables projects and incentivizing more buildout overall.
This strategy, Schmitt told me, “reflects [Crusoe’s] guiding principle of bringing load to stranded and under-utilized energy” rather than bringing energy sources to the data center load itself, as the industry has traditionally done. G2, the venture capital firm, is all in on this premise. “By putting a big load center right there in a fantastic renewable resource environment, the thing that will naturally get built is renewables,” Kortlang told me. “Crusoe doesn’t need to mandate that, or control that, or be the one building the renewables. They’re creating the demand.”
But this approach is only net-positive for the climate if it increases the share of renewables in the mix overall, i.e. if new, large loads are leading to more solar and wind buildout than new natural gas buildout. And while a renewables-heavy buildout seems to be what Crusoe and its investors are assuming will happen, Crusoe can’t actually control what gets put on the grid or the economic or political factors that drive those decisions.
It appears to be inevitable that gas will play some role, even if it’s providing power directly to the data center itself and not to the grid overall. According to Business Insider, public filings with the Texas Commission on Environmental Quality show that so far, Crusoe plans to operate on-site natural gas turbines at the Abilene facility totaling 360 megawatts of power. That represents 30% of the data center’s total 1.2 gigawatts of announced capacity.
Although powering data centers with new solar or wind is usually the cheapest option — especially in places like Abilene — building natural gas can be quicker and more reliable, assuming you’re able to acquire the severely backlogged turbines. That’s something Kortlang readily acknowledged to me. “We will see a lot of buildout of natural gas over the last half of this decade, because it’s the easiest thing to controllably build that gets you large amounts of baseload power quickly,” he said.
Kortlang didn’t seem fazed by Crusoe’s announcement this Monday that it’s pursuing a joint venture with the investment firm Engine No. 1, giving the company access to a whopping 4.5 gigawatts of natural gas power. To put that in perspective, there’s only about 25 gigawatts of existing data center capacity in the U.S. today. Schmitt told me this latest announcement is unrelated to the Stargate Project.
Engine No. 1 has secured seven GE Vernova natural gas turbines through a partnership with Chevron announced in January. As Chevron puts it, this joint development will create “scalable, reliable power solutions for United States-based data centers running on U.S. natural gas.” But critically, as Crusoe emphasized, “plans for these data centers include the use of post-combustion carbon capture systems,” which are designed to capture the CO2 from power plants after the fossil fuels are burned, but before they’re released to the atmosphere.
Presumably, these plans will also incorporate either some way to utilize the CO2 in industry or to permanently sequester it underground, though the company hasn’t mentioned anything to this effect. This technology hasn’t been a part of the company’s strategy in the past, though Kortlang told me that Crusoe has been evaluating the viability of carbon capture and storage for as long as G2 has been involved.
Gas-fired power plants paired with carbon capture have never really caught on, simply because they’re pretty much bound to cost more than not building carbon capture. When I asked Kortlang if this meant Crusoe was banking on its data center customers being willing to pay more for greener power, he told me that was “to be determined.” Who exactly was going to design and build the carbon capture technology — Crusoe, Chevron, or another to-be-named project partner — was also “to be determined.” But there’s not actually all that much time to figure it out. In Chevron’s announcement, the company said it was planning to deliver power by the end of 2027.
So, is Crusoe still a climate tech company? The answer seems to be yes — or at least it’s definitely still trying to be.
No other developer has been as diligent about utilizing stranded assets to power data centers. And with its expansion into carbon capture, it certainly seems Crusoe is leaning into an all-of-the-above approach to data center decarbonization. As Dumas told me, “before too long” we’ll also see Crusoe powering its operations with “geothermal, bioenergy, and after that fusion technologies that keep them out ahead of the pack.”
But Crusoe’s business model — and its clean tech bonafides in general — have always relied upon ultimately unprovable counterfactuals. First it was: If this waste gas weren’t powering Bitcoin mining, it would be vented into the atmosphere. That seemed fairly certain, since flaring is common practice in many areas. Now the company is pitching a somewhat fuzzier hypothetical: If this Crusoe data center, powered by some combination of natural gas and stranded renewables, were instead built by another company, it would inevitably be dirtier. Whether or not Crusoe is a boon for the climate ultimately depends upon the degree to which that unquantifiable claim ends up being true.
On Energy Transfer’s legal win, battery storage, and the Cybertruck
Current conditions: Red flag warnings are in place for much of Florida • Spain is bracing for extreme rainfall from Storm Martinho, the fourth named storm in less than two weeks • Today marks the vernal equinox, or the first day of spring.
A jury has ordered Greenpeace to pay more than $660 million in damages to one of the country’s largest fossil fuel infrastructure companies after finding the environmental group liable for defamation, conspiracy, and physical damages at the Dakota Access Pipeline. Greenpeace participated in large protests, some violent and disruptive, at the pipeline in 2016, though it has maintained that its involvement was insignificant and came at the request of the local Standing Rock Sioux Tribe. The project eventually went ahead and is operational today, but Texas-based Energy Transfer sued the environmental organization, accusing it of inciting the uprising and encouraging violence. “We should all be concerned about the future of the First Amendment, and lawsuits like this aimed at destroying our rights to peaceful protest and free speech,” said Deepa Padmanabha, senior legal counsel for Greenpeace USA. The group said it plans to appeal.
The Department of Energy yesterday approved a permit for the Calcasieu Pass 2 liquified natural gas terminal in Louisiana, allowing the facility to export to countries without a free trade agreement. The project hasn’t yet been constructed and is still waiting for final approvals from the independent Federal Energy Regulatory Commission, but the DOE’s green light means it faces one less hurdle.
CP2 was awaiting DOE’s go-ahead when the Biden administration announced its now notorious pause on approvals for new LNG export facilities. The project’s opponents argue it’s a “carbon bomb.” Analysis from the National Resources Defense Council suggested the greenhouse gases from the project would be equivalent to putting more than 1.85 million additional gas-fueled automobiles on the road, while the Sierra Club found it would amount to about 190 million tons of carbon dioxide equivalent annually.
President Trump met with 15 to 20 major oil and gas executives from the American Petroleum Institute at the White House yesterday. This was the president’s first meeting with fossil fuel bosses since his second term began in January. Interior Secretary Doug Burgum and Energy Secretary Chris Wright were also in the room. Everyone is staying pretty quiet about what exactly was said, but according to Burgum and Wright, the conversation focused heavily on permitting reform and bolstering the grid. Reuters reported that “executives had been expected to express concerns over Trump’s tariffs and stress the industry view that higher oil prices are needed to help meet Trump’s promise to grow domestic production.” Burgum, however, stressed that oil prices didn’t come up in the chat. “Price is set by supply and demand,” he said. “There was nothing we could say in that room that could change that one iota, and so it wasn’t really a topic of discussion.” The price of U.S. crude has dropped 13% since Trump returned to office, according to CNBC, on a combination of recession fears triggered by Trump’s tariffs and rising oil output from OPEC countries.
The U.S. installed 1,250 megawatts of residential battery storage last year, the highest amount ever and nearly 60% more than in 2023, according to a new report from the American Clean Power Association and Wood Mackenzie. Overall, battery storage installations across all sectors hit a new record in 2024 at 12.3 gigawatts of new capacity. Storage is expected to continue to grow next year, but uncertainties around tariffs and tax incentives could slow things down.
China is delaying approval for construction of BYD’s Mexico plant because authorities worry the electric carmaker’s technology could leak into the United States, according to the Financial Times. “The commerce ministry’s biggest concern is Mexico’s proximity to the U.S.,” sources told the FT. As Heatmap’s Robinson Meyer writes, BYD continues to set the global standard for EV innovation, and “American and European carmakers are still struggling to catch up.” This week the company unveiled its new “Super e-Platform,” a new standard electronic base for its vehicles that it says will allow incredibly fast charging — enabling its vehicles to add as much as 249 miles of range in just five minutes.
Tesla has recalled 46,096 Cybertrucks over an exterior trim panel that can fall off and become a road hazard. This is the eighth recall for the truck since it went on sale at the end of 2023.
This fusion startup is ahead of schedule.
Thea Energy, one of the newer entrants into the red-hot fusion energy space, raised $20 million last year as investors took a bet on the physics behind the company’s novel approach to creating magnetic fields. Today, in a paper being submitted for peer review, Thea announced that its theoretical science actually works in the real world. The company’s CEO, Brian Berzin, told me that Thea achieved this milestone “quicker and for less capital than we thought,” something that’s rare in an industry long-mocked for perpetually being 30 years away.
Thea is building a stellarator fusion reactor, which typically looks like a twisted version of the more common donut-shaped tokamak. But as Berzin explained to me, Thea’s stellarator is designed to be simpler to manufacture than the industry standard. “We don’t like high tech stuff,” Berzin told me — a statement that sounds equally anathema to industry norms as the idea of a fusion project running ahead of schedule. “We like stuff that can be stamped and forged and have simple manufacturing processes.”
The company thinks it can achieve simplicity via its artificial intelligence software, which controls the reactor’s magnetic field keeping the unruly plasma at the heart of the fusion reaction confined and stabilized. Unlike typical stellarators, which rely on the ultra-precise manufacturing and installment of dozens of huge, twisted magnets, Thea’s design uses exactly 450 smaller, simpler planar magnets, arranged in the more familiar donut-shaped configuration. These magnets are still able to generate a helical magnetic field — thought to keep the plasma better stabilized than a tokamak — because each magnet is individually controlled via the company’s software, just like “the array of pixels in your computer screen,” Berzin told me.
“We’re able to utilize the control system that we built and very specifically modulate and control each magnet slightly differently,” Berzin explained, allowing Thea to “make those really complicated, really precise magnetic fields that you need for a stellarator, but with simple hardware.”
This should make manufacturing a whole lot easier and cheaper, Berzin told me. If one of Thea’s magnets is mounted somewhat imperfectly, or wear and tear of the power plant slightly shifts its location or degrades its performance over time, Thea’s AI system can automatically compensate. “It then can just tune that magnet slightly differently — it turns that magnet down, it turns the one next to it up, and the magnetic field stays perfect,” Berzin explained. As he told me, a system that relies on hardware precision is generally much more expensive than a system that depends on well-designed software. The idea is that Thea’s magnets can thus be mass manufactured in a way that’s conducive to “a business versus a science project.”
In 2023, Thea published a technical report proving out the physics behind its so-called “planar coil stellarator,” which allowed the company to raise its $20 million Series A last year, led by the climate tech firm Prelude Ventures. To validate the hardware behind its initial concept, Thea built a 3x3 array of magnets, representative of one section of its overall “donut” shaped reactor. This array was then integrated with Thea’s software and brought online towards the end of last year.
The results that Thea announced today were obtained during testing last month, and prove that the company can create and precisely control the complex magnetic field shapes necessary for fusion power. These results will allow the company to raise a Series B in the “next couple of years,” Berzin said. During this time, Thea will be working to scale up manufacturing such that it can progress from making one or two magnets per week to making multiple per day at its New Jersey-based facility.
The company’s engineers are also planning to stress test their AI software, such that it can adapt to a range of issues that could arise after decades of fusion power plant operation. “So we’re going to start breaking hardware in this device over the next month or two,” Berzin told me. “We’re purposely going to mismount a magnet by a centimeter, put it back in and not tell the control system what we did. And then we’re going to purposely short out some of the magnetic coils.” If the system can create a strong, stable magnetic field anyway, this will serve as further proof of concept for Thea’s software-oriented approach to a simplified reactor design.
The company is still years away from producing actual fusion power though. Like many others in the space, Thea hopes to bring fusion electrons to the grid sometime in the 2030s. Maybe this simple hardware, advanced software approach is what will finally do the trick.