The world can't keep on with what it's doing and expect to reach its goals when it comes to climate change. Radical innovations are needed at this point, writes Scott Nyquist. Photo via Getty Images

Almost 3 years ago, McKinsey published a report arguing that limiting global temperature rises to 1.5 degrees Celsius above pre-industrial levels was “technically achievable,” but that the “math is daunting.” Indeed, when the 1.5°C figure was agreed to at the 2015 Paris climate conference, the assumption was that emissions would peak before 2025, and then fall 43 percent by 2030.

Given that 2022 saw the highest emissions ever—36.8 gigatons—the math is now more daunting still: cuts would need to be greater, and faster, than envisioned in Paris. Perhaps that is why the Intergovernmental Panel on Climate Change (IPCC) noted March 20 (with “high confidence”) that it was “likely that warming will exceed 1.5°C during the 21st century.”

I agree with that gloomy assessment. Given the rate of progress so far, 1.5°C looks all but impossible. That puts me in the company of people like Bill Gates; the Economist; the Australian Academy of Science, and apparently many IPCC scientists. McKinsey has estimated that even if all countries deliver on their net zero commitments, temperatures will likely be 1.7°C higher in 2100.

In October, the UN Environment Program argued that there was “no credible pathway to 1.5°C in place” and called for “an urgent system-wide transformation” to change the trajectory. Among the changes it considers necessary: carbon taxes, land use reform, dietary changes in which individuals “consume food for environmental sustainability and carbon reduction,” investment of $4 trillion to $6 trillion a year; applying current technology to all new buildings; no new fossil fuel infrastructure. And so on.

Let’s assume that the UNEP is right. What are the chances of all this happening in the next few years? Or, indeed, any of it? President Obama’s former science adviser, Daniel Schrag, put it this way: “ Who believes that we can halve global emissions by 2030?... It’s so far from reality that it’s kind of absurd.”

Having a goal is useful, concentrating minds and organizing effort. And I think that has been the case with 1.5°C, or recent commitments to get to net zero. Targets create a sense of urgency that has led to real progress on decarbonization.

The 2020 McKinsey report set out how to get on the 1.5°C pathway, and was careful to note that this was not a description of probability or reality but “a picture of a world that could be.” Three years later, that “world that could be” looks even more remote.

Consider the United States, the world’s second-largest emitter. In 2021, 79 percent of primary energy demand (see chart) was met by fossil fuels, about the same as a decade before. Globally, the figures are similar, with renewables accounting for just 12.5 percent of consumption and low-emissions nuclear another 4 percent. Those numbers would have to basically reverse in the next decade or so to get on track. I don’t see how that can happen.

No alt text provided for this image

Credit: Energy Information Administration

But even if 1.5°C is improbable in the short term, that doesn’t mean that missing the target won’t have consequences. And it certainly doesn’t mean giving up on addressing climate change. And in fact, there are some positive trends. Many companies are developing comprehensive plans for achieving net-zero emissions and are making those plans part of their long-term strategy. Moreover, while global emissions grew 0.9 percent in 2022, that was much less than GDP growth (3.2 percent). It’s worth noting, too, that much of the increase came from switching from gas to coal in response to the Russian invasion of Ukraine; that is the kind of supply shock that can be reversed. The point is that growth and emissions no longer move in lockstep; rather the opposite. That is critical because poorer countries are never going to take serious climate action if they believe it threatens their future prosperity.

Another implication is that limiting emissions means addressing the use of fossil fuels. As noted, even with the substantial rise in the use of renewables, coal, gas, and oil are still the core of the global energy system. They cannot be wished away. Perhaps it is time to think differently—that is, making fossil fuels more emissions efficient, by using carbon capture or other technologies; cutting methane emissions; and electrifying oil and gas operations. This is not popular among many climate advocates, who would prefer to see fossil fuels “stay in the ground.” That just isn’t happening. The much likelier scenario is that they are gradually displaced. McKinsey projects peak oil demand later this decade, for example, and for gas, maybe sometime in the late 2030s. Even after the peak, though, oil and gas will still be important for decades.

Second, in the longer term, it may be possible to get back onto 1.5°C if, in addition to reducing emissions, we actually remove them from the atmosphere, in the form of “negative emissions,” such as direct air capture and bioenergy with carbon capture and storage in power and heavy industry. The IPCC itself assumed negative emissions would play a major role in reaching the 1.5°C target; in fact, because of cost and deployment problems, it’s been tiny.

Finally, as I have argued before, it’s hard to see how we limit warming even to 2°C without more nuclear power, which can provide low-emissions energy 24/7, and is the largest single source of such power right now.

None of these things is particularly popular; none get the publicity of things like a cool new electric truck or an offshore wind farm (of which two are operating now in the United States, generating enough power for about 20,000 homes; another 40 are in development). And we cannot assume fast development of offshore wind. NIMBY concerns have already derailed some high-profile projects, and are also emerging in regard to land-based wind farms.

Carbon capture, negative emissions, and nuclear will have to face NIMBY, too. But they all have the potential to move the needle on emissions. Think of the potential if fast-growing India and China, for example, were to develop an assembly line of small nuclear reactors. Of course, the economics have to make sense—something that is true for all climate-change technologies.

And as the UN points out, there needs to be progress on other issues, such as food, buildings, and finance. I don’t think we can assume that such progress will happen on a massive scale in the next few years; the actual record since Paris demonstrates the opposite. That is troubling: the IPCC notes that the risks of abrupt and damaging impacts, such as flooding and crop yields, rise “with every increment of global warming.” But it is the reality.

There is one way to get us to 1.5°C, although not in the Paris timeframe: a radical acceleration of innovation. The approaches being scaled now, such as wind, solar, and batteries, are the same ideas that were being discussed 30 years ago. We are benefiting from long-term, incremental improvements, not disruptive innovation. To move the ball down the field quickly, though, we need to complete a Hail Mary pass.

It’s a long shot. But we’re entering an era of accelerated innovation, driven by advanced computing, artificial intelligence, and machine learning that could narrow the odds. For example, could carbon nanotubes displace demand for high-emissions steel? Might it be possible to store carbon deep in the ocean? Could geo-engineering bend the curve?

I believe that, on the whole, the world is serious about climate change. I am certain that the energy transition is happening. But I don’t think we are anywhere near to being on track to hit the 1.5°C target. And I don’t see how doing more of the same will get us there.

------

Scott Nyquist is a senior advisor at McKinsey & Company and vice chairman, Houston Energy Transition Initiative of the Greater Houston Partnership. The views expressed herein are Nyquist's own and not those of McKinsey & Company or of the Greater Houston Partnership. This article originally ran on LinkedIn.

Ad Placement 300x100
Ad Placement 300x600

CultureMap Emails are Awesome

Houston quantum simulator research reveals clues for solar energy conversion

energy flow

Rice University scientists have used a programmable quantum simulator to mimic how energy moves through a vibrating molecule.

The research, which was published in Nature Communications last month, lets the researchers watch and control the flow of energy in real time and sheds light on processes like photosynthesis and solar energy conversion, according to a news release from the university.

The team, led by Rice assistant professor of physics and astronomy Guido Pagano, modeled a two-site molecule with one part supplying energy (the donor) and the other receiving it (the acceptor).

Unlike in previous experiments, the Rice researchers were able to smoothly tune the system to model multiple types of vibrations and manipulate the energy states in a controlled setting. This allowed the team to explore different types of energy transfer within the same platform.

“By adjusting the interactions between the donor and acceptor, coupling to two types of vibrations and the character of those vibrations, we could see how each factor influenced the flow of energy,” Pagano said in the release.

The research showed that more vibrations sped up energy transfer and opened new paths for energy to move, sometimes making transfer more efficient even with energy loss. Additionally, when vibrations differed, efficient transfer happened over a wider range of donor–acceptor energy differences.

“The results show that vibrations and their environment are not simply background noise but can actively steer energy flow in unexpected ways,” Pagano added.

The team believes the findings could help with the design of organic solar cells, molecular wires and other devices that depend on efficient energy or charge transfer. They could also have an environmental impact by improving energy harvesting to reduce energy losses in electronics.

“These are the kinds of phenomena that physical chemists have theorized exist but could not easily isolate experimentally, especially in a programmable manner, until now,” Visal So, a Rice doctoral student and first author of the study, added in the release.

The study was supported by The Welch Foundation,the Office of Naval Research, the National Science Foundation CAREER Award, the Army Research Office and the Department of Energy.

The EPA is easing pollution rules — here’s how it’s affecting Texas

In the news

The first year of President Trump’s second term has seen an aggressive rollback of federal environmental protections, which advocacy groups fear will bring more pollution, higher health risks, and less information and power for Texas communities, especially in heavily industrial and urban areas.

Within Trump’s first 100 days in office, his new Environmental Protection Agency administrator, Lee Zeldin, announced a sweeping slate of 31 deregulatory actions. The list, which Zeldin called the agency’s “greatest day of deregulation,” targeted everything from soot standards and power plant pollution rules to the Endangerment Finding, the legal and scientific foundation that obligates the EPA to regulate climate-changing pollution under the Clean Air Act.

Since then, the agency froze research grants, shrank its workforce, and removed some references to climate change and environmental justice from its website — moves that environmental advocates say send a clear signal: the EPA’s new direction will come at the expense of public health.

Cyrus Reed, conservation director of the Lone Star Chapter of the Sierra Club, said Texas is one of the states that feels EPA policy changes directly because the state has shown little interest in stepping up its environmental enforcement as the federal government scales back.

“If we were a state that was open to doing our own regulations there’d be less impact from these rollbacks,” Reed said. “But we’re not.”

“Now we have an EPA that isn’t interested in enforcing its own rules,” he added.

Richard Richter, a spokesperson at the state’s environmental agency, Texas Commission on Environmental Quality, said in a statement that the agency takes protecting public health and natural resources seriously and acts consistently and quickly to enforce federal and state environmental laws when they’re violated.

Methane rules put on pause

A major EPA move centers on methane, a potent greenhouse gas that traps heat far more efficiently than carbon dioxide over the short term. It accounts for roughly 16% of global greenhouse gas emissions and is a major driver of climate change. In the U.S., the largest source of methane emissions is the energy sector, especially in Texas, the nation’s top oil and gas producer.

In 2024, the Biden administration finalized long-anticipated rules requiring oil and gas operators to sharply reduce methane emissions from wells, pipelines, and storage facilities. The rule, developed with industry input, targeted leaks, equipment failures, and routine flaring, the burning off of excess natural gas at the wellhead.

Under the rule, operators would have been required to monitor emissions, inspect sites with gas-imaging cameras for leaks, and phase out routine flaring. States are required to come up with a plan to implement the rule, but Texas has yet to do so. Under Trump’s EPA, that deadline has been extended until January 2027 — an 18-month postponement.

Texas doesn’t have a rule to capture escaping methane emissions from energy infrastructure. Richter, the TCEQ spokesperson, said the agency continues to work toward developing the state plan.

Adrian Shelley, Texas director of the watchdog group Public Citizen, said the rule represented a rare moment of alignment between environmentalists and major oil and gas producers.

“I think the fossil fuel industry generally understood that this was the direction the planet and their industry was moving,” he said. Shelley said uniform EPA rules provided regulatory certainty for changes operators saw as inevitable.

Reed, the Sierra Club conservation director, said the delay of methane rules means Texas still has no plan to reduce emissions, while neighboring New Mexico already has imposed its own state methane emission rules that require the industry to detect and repair methane leaks and ban routine venting and flaring.

These regulations have cut methane emissions in the New Mexico portion of the Permian Basin — the oil-rich area that covers West Texas and southeast New Mexico — to half that of Texas, according to a recent data analysis by the Environmental Defense Fund. That’s despite New Mexico doubling production since 2020.

A retreat from soot standards

Fine particulate matter or PM 2.5, one of six pollutants regulated under the Clean Air Act, has been called by researchers the deadliest form of air pollution.

In 2024, the EPA under President Biden strengthened air rules for particulate matter by lowering the annual limit from 12 to 9 micrograms per cubic meter. It was the first update since 2012 and one of the most ambitious pieces of Biden’s environmental agenda, driven by mounting evidence that particulate pollution is linked to premature death, heart disease, asthma, and other respiratory illnesses.

After the rule was issued, 24 Republican-led states, including Kentucky and West Virginia, sued to revert to the weaker standard. Texas filed a separate suit asking to block the rule’s recent expansion.

State agencies are responsible for enforcing the federal standards. The TCEQ is charged with creating a list of counties that exceed the federal standard and submitting those recommendations to Gov. Greg Abbott, who then finalizes the designations and submits them to the EPA.

Under the 9 microgram standard, parts of Texas, including Dallas, Harris (which includes Houston), Tarrant (Fort Worth), and Bowie (Texarkana) counties, were in the process of being designated nonattainment areas — which, when finalized, would trigger a legal requirement for the state to develop a plan to clean up the air.

That process stalled after Trump returned to office. Gov. Greg Abbott submitted his designations to EPA last February, but EPA has not yet acted on his designations, according to Richter, the TCEQ spokesperson.

In a court filing last year, the Trump EPA asked a federal appeals court to vacate the stricter standard, bypassing the traditional notice and comment administrative process.

For now, the rule technically remains in effect, but environmental advocates say the EPA’s retreat undermines enforcement of the rule and signals to polluters that it may be short-lived.

Shelley, with Public Citizen, believes the PM2.5 rule would have delivered the greatest health benefit of any EPA regulation affecting Texas, particularly through reductions in diesel pollution from trucks.

“I still hold out hope that it will come back,” he said.

Unraveling the climate framework

Beyond individual pollutants, the Trump EPA has moved to dismantle the federal architecture for addressing climate change.

Among the proposals is eliminating the Greenhouse Gas Reporting Program, which requires power plants, refineries, and oil and gas suppliers to report annual emissions. The proposal has drawn opposition from both environmental groups and industry, which relies on the data for planning and compliance.

Colin Leyden, Texas state director and energy lead at the nonprofit Environmental Defense Fund, said eliminating the program could hurt Texas industry. If methane emissions are no longer reported, then buyers and investors of natural gas, for example, won’t have an official way to measure how much methane pollution is associated with that gas, according to Leyden. That makes it harder to judge how “clean” or “climate-friendly” the product is, which international buyers are increasingly demanding.

“This isn’t just bad for the planet,” he said. “It makes the Texas industry less competitive.”

The administration also proposed last year rescinding the Endangerment Finding, issued in 2009, which obligates the EPA to regulate climate pollution. Most recently, the EPA said it will stop calculating how much money is saved in health care costs as a result of air pollution regulations that curb particulate matter 2.5 and ozone, a component of smog. Both can cause respiratory and health problems.

Leyden said tallying up the dollar value of lives saved when evaluating pollution rules is a foundational principle of the EPA since its creation.

“That really erodes the basic idea that (the EPA) protects health and safety and the environment,” he said.

___

This story was originally published by The Texas Tribune and distributed through a partnership with The Associated Press.

New report predicts major data center boom in Texas by 2028

data analysis

Data centers are proving to be a massive economic force in Texas.

For instance, a new report from clean energy company Bloom Energy predicts Texas will see a 142 percent increase in its market share for data centers from 2025 to 2028. That would be the highest increase of any state.

Bloom Energy expects Texas to exceed 40 gigawatts of data-center capacity by 2028, representing a nearly 30 percent share of the U.S. market. A typical AI data center consumes 1 to 2 gigawatts of energy.

“Data center and AI factory developers can’t afford delays,” Natalie Sunderland, Bloom Energy’s chief marketing officer, said in the report. “Our analysis and survey results show that they’re moving into power‑advantaged regions where capacity can be secured faster — and increasingly designing campuses to operate independently of the grid.”

“The surge in AI demand creates a clear opportunity for states that can adapt to support large-scale AI deployments at speed,” Sunderland adds.

Further evidence of the data center explosion in Texas comes from ConstructConnect, a provider of data and software for contractors and manufacturers. ConstructConnect reported that in the 12-month span through November 2025, data-center construction starts in Texas accounted for $11 billion in spending. At $12.5 billion, only Louisiana surpassed the Texas total.

Capital expenses for U.S. data centers were expected to surpass $425 billion last year, according to ratings agency S&P Global.

ConstructConnect also reports that Texas is among five states collectively grabbing 80 percent of potential data center construction starts. Currently, Texas hosts around 400 data centers, with close to 60 of them in the Houston market.

A large pool of data-center construction spending in Texas is flowing from Google, which announced in November that it would earmark $40 billion for new AI data centers in the state.

“Texas leads in AI and tech innovation,” Gov. Greg Abbott proclaimed when the Google investment was unveiled.

Other studies and reports lay out just how much data centers are influencing economic growth in the Lone Star State:

  • A study by Texas Royalty Brokers indicates Texas leads the U.S. with 17 clusters of AI data centers. The study measured the density of AI data centers by counting the number of graphics processing units (GPUs) installed in those clusters. GPUs are specialized chips built to run AI models and perform complex calculations.
  • Citing data from construction consulting company FMI, The Wall Street Journal reported that spending on construction of data centers is expected to rise 23 percent in 2026 compared with last year. Much of that construction spending will happen in Texas. In the 12 months through November 2025, the average data center cost $597 million, according to ConstructConnect.
  • Data published in 2025 by commercial real estate services company Cushman & Wakefield shows three Texas markets — Austin, Dallas and San Antonio — boast the lowest construction costs for data centers among the 19 U.S. markets that were analyzed. The mid-range of costs in that trio of markets is roughly $10.65 million per megawatt. Houston isn’t included in the data.

Although Houston isn’t cited in the Cushman & Wakefield data, it nonetheless is playing a major role in the data-center boom. Houston-area energy giants Chevron and ExxonMobil are chasing opportunities to supply natural gas as a power source for data centers, for example.

“As Houston rapidly evolves into a hub for AI, cloud computing, and data infrastructure, the city is experiencing a surge in data-center investments driven by its unique position at the intersection of energy, technology, and innovation,” says the Greater Houston Partnership.