A new study puts Texas at No. 2 among the states when it comes to manufacturing. Photo via Getty Images

Texas ranks among the country’s biggest hubs for manufacturing, according to a new study.

The study, conducted by Chinese manufacturing components supplier YIJIN Hardware, puts Texas at No. 2 among the states when it comes to manufacturing-hub status. California holds the top spot.

YIJIN crunched data from the U.S. Census Bureau, International Trade Administration, and National Association of Manufacturers to analyze manufacturing activity in each state. The study weighed factors such as number of manufacturing establishments, number of manufacturing employees, total value of manufacturing output, total manufacturing exports and manufacturing’s share of a state’s gross domestic product.

Here are Texas’ figures for those categories:

  • 19,526 manufacturing establishments
  • 847,470 manufacturing employees
  • Total manufacturing output of $292.6 billion
  • Total manufacturing exports of $291.9 billion
  • 11.3 percent share of state GDP

According to Texas Economic Development & Tourism, the state’s largest manufacturing sectors include automotive, tech, petroleum, chemicals, and food and beverage.

“The Lone Star State is truly a manufacturing powerhouse,” the state agency says.

In an October speech, Texas Gov. Greg Abbott praised the state’s robust manufacturing industry.

“We are proud that Texas is home to a booming manufacturing sector,” he said. “Thanks to our strong manufacturing sector, ‘Made in Texas’ has never been a bigger brand.”

Houston is a cornerstone of Texas’ manufacturing industry. The region produces more than $75 billion worth of goods each year, according to the Greater Houston Partnership. That makes Houston the second-ranked U.S. metro area for manufacturing GDP. The more than 7,000 manufacturing establishments in the area employ over 223,000 people.

“As one of the most important industrial bases in the world, Houston has access to many global markets thanks to its central location within the U.S. and the Americas,” the partnership says.

TMEIC will move its headquarters to Houston next year and open a new manufacturing facility in the region later this year. Photo via tmeic.com

Japanese energy tech manufacturer to relocate US HQ to Houston, open new facility

moving in

A Japanese company has announced its moving its United States headquarters to Houston and is gearing up top open its new Houston-area factory as well.

TMEIC Corporation Americas, previously headquartered in Roanoke, Virginia, will officially be located in Houston, effect March of 2025. Additionally, the company will open a state-of-the-art 144,000-square-foot facility in Brookshire, which will be dedicated to manufacturing utility-scale PV inverters. The expansion is expected to create 300 local jobs.

The TMEIC group specializes in photovoltaic inverters and energy storage systems, and has over 50 GW of renewable energy systems installed worldwide as of July 2024.

"We are excited to make these investments for an expanded presence in the Houston area with the relocation of our headquarters and the opening of our new manufacturing facility,” Manmeet S. Bhatia, president and CEO of TMEIC Corporation Americas, says in a news release. ”These investments and expansions will potentially create up to 300 jobs in the local community,"

The relocation to the Houston as the energy capital of the world is part of TMEIC’s strategic goals for growth in “renewable energy technology, domestic based manufacturing, and bolstering its global sustainability efforts,” according to a news release.

The Brookshire facility will be complete by October of 2024, and will be close to TMEIC’s existing uninterruptible power supply and medium voltage drive manufacturing plant in Katy. When operational, it will have the capacity to produce 9 gigawatts annually.

“This strategic expansion underscores TMEIC's dedication to the renewable energy industry, advancing clean energy technology, maintaining strong client relationships, and competing on a global basis while proudly manufacturing in the United States,” Bhatia adds.

Hear from guest columnist Onega Ulanova on AI and quality management systems in manufacturing. Photo via Getty Images

Expert: How AI is disrupting manufacturing and the future of quality management systems

guest column

The concept of quality management is so intrinsic to modern manufacturing — and yet so little understood by the general public — and has literally revolutionized our world over the past hundred years.

Yet, in the present day, quality management and the related systems that guide its implementation are far from static. They are continuously-evolving, shifting to ever-changing global conditions and new means of application unleashed by technological innovation.

Now, more than ever, they are essential for addressing and eliminating not only traditional sources of waste in business, such as lost time and money, but also the physical and pollutant waste that threatens the world we all inhabit.

But what are quality management systems, or QMS, exactly? Who created them, and how have they evolved over time? Perhaps most pressingly, where can they be of greatest help in the present world, and when can they be implemented by businesses in need of change and improvement?

In this article, we will explore the history of QMS, explain their essential role in today’s manufacturing practices, and examine how these systems will take us into the future of productivity.

Quality Management Systems: A Definition

In the United States and globally, the gold standard of quality management standards and practices is the American Society for Quality. This preeminent organization, with over 4,000 members in 130 countries, was established in 1946 and has guided practices and implementation of quality management systems worldwide.

The Society defines a quality management system as “a formalized system that documents processes, procedures, and responsibilities for achieving quality policies and objectives,” and further states that “a QMS helps coordinate and direct an organization’s activities to meet customer and regulatory requirements and improve its effectiveness and efficiency on a continuous basis.”

From this definition, it can be understood that a good quality management system’s purpose is to establish the conditions for consistent and ever-increasing improvement through the use of standardized business culture practices.

Which QMS Standards are Most Widely Used?

The results of quality management’s remarkable growth since the 1940s has led to the rise of a number of widely-used standards, which can serve as the basis for companies and organizations to design and implement their own practices. Most of these modern quality management standards are globally recognized, and are specifically tailored to ensure that a company’s newly-developed practices include essential elements that can increase the likelihood of success.

The most widely-known entity which has designed such guidance is the International Organization for Standardization (ISO), a global organization which develops and publishes technical standards. Since the 1980s, the ISO has provided the 9000 series of standards (the most famous of which is 9001:2015) which outline how organizations can satisfy the checklists of quality management requirements and create their own best practices.

In 2020, over 1.2 million organizations worldwide were officially certified by the ISO for their quality management implementation practices.

However, it should be understood that the ISO 9000 standards are merely guidelines for the design and implementation of a quality management system; they are not systems in and of themselves.

Furthermore, the ISO is far from the only relevant player in this field. Many industry-specific standards, such as the American Petroleum Institute’s API Q1 standard, have been developed to target the highly specialized needs of particular business practices of oil and gas industry. These industry-specific standards are generally aligned with the ISO 9000 standards, and serve as complimentary additional guidance, rather than a replacement. It is entirely possible, and in many cases desirable, for a company to receive both ISO certification and certification from an industry-specific standards body, as doing so can help ensure the company’s newly-developed QMS procedures are consistent with both broad and specialized best practices.

A History of Quality Management

The concept of quality management is intrinsically tied to the development of industrial production. Previous to the industrial revolution, the concept of ‘quality’ was inherently linked to the skill and effort of craftspeople, or in other words, individual laborers trained in specialized fields who, either individually or in small groups, produced goods for use in society.

Whether they were weaving baskets or building castles, these craftspeople were primarily defined by a skill that centered them in a specific production methodology, and it was the mastery of this skill which determined the quality. Guilds of craftspeople would sign their works, placing a personal or group seal on the resulting product and thereby accepting accountability for its quality.

Such signatures and marks are found dating back at least 4,500 years to the construction of Egypt’s Great Pyramid of Giza, and came into widespread practice in medieval Europe with the rise of craft guilds.

In these early confederations of workers, a person’s mastery of a skill or craft could become a defining part of their identity and life, to the extent that many craftspeople of 13th Century Europe lived together in communal settings, while the Egyptian pyramid workers may have belonged to life-long ‘fraternities’ who returned, year after year, to fulfill their roles in ‘work gangs’.

However, in the Industrial Revolution, craft and guild organizations were supplanted by factories. Though ancient and medieval projects at times reached monumental scale, the rise of thousands of factories, each requiring human and machine contributions to generate masses of identical products, required a completely different scale of quality management.

The emphasis on mass production necessitated the use of workers who were no longer crafts masters, and thus resulted in a decrease in the quality of products. This in turn necessitated the rise of the product inspection system, which was steadily refined from the start of the Industrial Revolution in 1760 into the early 20th century.

However, inspection was merely a system of quality control, rather than quality management; in other words, simply discarding defective products did not in and of itself increase total product quality or reduce waste.

As influential American engineer Joseph M. Juran explained, in 1920s-era America, it was common to throw away substantial portions of produced inventory due to defects, and when Juran prompted inspectors at his employer’s company to do something, they refused, saying it was the responsibility of the production line to improve. Quality control, in and of itself, would not yield quality management.

As is often the case in human history, war was the driver of change. In World War II, the mobilization of millions of American workers into wartime roles coincided with the need to produce greater quantities of high-quality products than ever before.

To counteract the loss of skilled factory labor, the United States government implemented the Training Within Industry program, which utilized 10-hour courses to educate newly-recruited workers in how to conduct their work, evaluate their efficiency, and suggest improvements. Similar training programs for the trainers themselves were also developed. By the end of the war, more than 1.6 million workers had been certified under the Training Within Industry program.

Training Within Industry represented one of the first successful implementations of quality management systems, and its impact was widely felt after the end of the war. In the ashes of conflict, the United States and the other Allied Powers were tasked with helping to rebuild the economies of the other wartime combatants. Nowhere was this a more pressing matter than Japan, which had seen widespread economic devastation and had lost 40 percent of all its factories. Further complicating the situation was the reality that, then as now, Japan lacked sufficient natural resources to serve its economic scale.

And yet, within just 10 years of the war’s end, Japan’s economy war growing twice as fast per year than it had been before the fighting started. The driver of this miraculous turnaround was American-derived quality management practices, reinterpreted and implemented with Japanese ingenuity.

In modern business management, few concepts are as renowned, and oft-cited for success, as kaizen. This Japanese word, which simply means “improvement,” is the essential lesson and driver of Japan’s postwar economic success.

Numerous books written outside Japan have attempted to explain kaizen’s quality management principles, often by citing them as being ‘distinctly Japanese.’ Yet, the basis for kaizen is actually universal and applicable in any culture or context; it is, simply put, an emphasis on remaining quality-focused and open to evolution. The development of kaizen began in the post-war period when American statistician William Edwards Deming was brought to Japan as part of the US government’s rebuilding efforts.

A student of earlier quality management thought leaders, Deming instructed hundreds of Japanese engineers, executives, and scholars, urging them to place statistical analysis and human relationships at the center of their management practices. Deming used statistics to track the number and origin of product defects, as well to analyze the effectiveness of remedies. He also reinstated a key idea of the craftsperson creed: that the individual worker is not just a set of hands performing a task, but a person who can, with time, improve both the self and the whole of the company.

Deming was not alone in these efforts; the aforementioned Joseph M. Juran, who came to Japan as part of the rebuilding program several years later, also gave numerous lectures expounding similar principles.

Like Deming, Juran had previously tried to impart these approaches to American industry, but the lessons often fell on deaf ears. Japanese managers, however, took the lessons to heart and soon began crafting their own quality management systems.

Kaoru Ishikawa, who began by translating the works of Deming and Juran into Japanese, was one of the crucial players who helped to create the ideas now known as kaizen. He introduced a bottom-up approach where workers from every part of the product life cycle could initiate change, and popularized Deming’s concept of quality circles, where small groups of workers would meet regularly to analyze results and discuss improvements.

By 1975, Japanese product quality, which had once been regarded as poor, had transformed into world-class thanks to the teachings of Deming, Juran, and kaizen.

By the 1980s, American industry had lost market share and quality prestige to Japan. It was now time for US businesses to learn from Deming and Juran, both of whom at last found a receptive audience in their home country. Deming in particular achieved recognition for his role in the influential 1980 television documentary If Japan Can, Why Can’t We?, in which he emphasized the universal applicability of quality management.

So too did kaizen, which influenced a new generation of global thought leaders. Arising out of this rapid expansion of QMS were new systems in the 1970s and ‘80s, including the Six Sigma approach pioneered by Bill Smith and Motorola in 1987. Ishikawa, who saw his reputation and life transformed as his ideas spread worldwide, eventually summed up the explanation as the universality of human nature and its desire to improve. As Ishikawa said, “wherever they are, human beings are human beings”.

In no small part due to the influence of the thought leaders mentioned, quality management systems are today a cornerstone of global business practice. So influential are the innovators of these systems that they are often called ‘gurus.’ But what are the specific benefits of these systems, and how best can they be implemented?

How QMS Benefits Organizations, and the World

The oft-cited benefits of quality management systems are operational efficiency, employee retention, and reduction of waste. From all of these come improvements to the company’s bottom line and reputation. But far from being dry talking points, each benefit not only serves its obvious purpose, but also can dramatically help benefit the planet itself.

Operational efficiency is the measurement, analysis, and improvement of processes which occur within an organization, with the purpose of utilizing data and consideration to eliminate or mediate any areas where current practices are not effective.

Quality management systems can increase operational efficiency by utilizing employee analysis and feedback to quickly identify areas where improvements are possible, and then to guide their implementation.

In a joint study conducted in 2017 by Forbes and the American Society for Quality, 56 percent of companies stated that improving operational efficiency was a top concern; in the same survey, 59 percent of companies received direct benefit to operations by utilizing quality management system practices, making it the single largest area of improvement across all business types.

Because operational improvements inherently reduce both waste and cost, conducting business in a fully-optimized manner can simultaneously save unnecessary resource expenditure, decrease pollutants and discarded materials, and retain more money which the company can invest into further sustainable practices. Efficiency is itself a kind of ‘stealth sustainability’ that turns a profit-focused mindset into a generator of greater good. It is this very point that the

United States government’s Environmental Protection Agency (EPA) has emphasized in their guidance for Environmental Management Systems (EMS). These quality management system guidelines, tailored specifically to benefit operational efficiency in a business setting, are also designed to benefit the global environment by utilizing quality management practices.

Examples in the EPA’s studies in preparing these guidelines showcased areas where small companies could reduce environmental waste, while simultaneously reducing cost, in numerous areas. These added to substantial reductions and savings, such as a 15 percent waste water reduction which saved a small metal finishing company $15,000 per year.

Similarly, a 2020 study by McKinsey & Company identified ways that optimizing operations could dramatically aid a company’s sustainability with only small outlays of capital, thereby making environmental benefit a by-product of improved profitability.

Employee retention, and more broadly the satisfaction of employees, is another major consideration of QMS. Defined simply, retention is not only the maintenance of a stable workforce without turnover, but the improvement of that workforce with time as they gain skill, confidence, and ability for continued self and organizational improvement. We may be in the post-Industrial Revolution, but thanks to the ideas of QMS, some of the concept of the craftsperson has returned to modern thinking; the individual, once more, has great value.

Quality management systems aid employee retention by allowing the people of an organization to have a direct hand in its improvement. In a study published in 2023 by the journal Quality Innovation Prosperity, 40 percent of organizations which implemented ISO 9001 guidance for the creation of a QMS reported that the process yielded greater employee retention.

A crucial success factor for employee satisfaction is how empowered the employee feels to apply judgment. According to a 2014 study by the Harvard Business Review, companies which set clear guidelines, protect and celebrate employee proposals for quality improvement, and clearly communicate the organization’s quality message while allowing the employees to help shape and implement it, have by far the highest engagement and retention rates. The greatest successes come from cultures where peer-driven approaches increase employee engagement, thereby eliminating preventable employee mistakes. Yet the same study also pointed out that nearly half of all employees feel their company’s leadership lacks a clear emphasis on quality, and only 10 percent felt their company’s existing quality statements were truthful and viable.

Then as now, the need to establish a clear quality culture, to manage and nurture that culture, and to empower the participants is critical to earning the trust of the employee participants and thereby retaining workers who in time can become the invaluable craftspeople of today.

Finally, there is the reduction of waste. Waste can be defined in many ways: waste of time, waste of money, waste of resources. The unifying factor in all definitions is the loss of something valuable, and irretrievable. All inevitably also lead to the increase of another kind of waste: pollution and discarded detritus which steadily ruin our shared planet.

Reducing waste with quality management can take many forms, but ultimately, all center on the realization of strategies which use only what is truly needed. This can mean both operational efficiencies and employee quality, as noted above. The Harvard Business Review survey identified that in 2014, the average large company (having 26,000 employees or more) loses a staggering $350 million each year due to preventable employee errors, many of which could be reduced, mitigated, or eliminated entirely with better implementation of quality management.

This is waste on an almost unimaginable financial scale. Waste eliminated through practices which emphasize efficiency and sustainability, as noted in the McKinsey & Company study, can also yield tremendous savings. In one example, a company which purchased asphalt and previously prioritized only the per-ton price found that, when examining the logistical costs of transporting the asphalt from distant suppliers, they were actually paying more than if they purchased it locally. The quality management analysis they performed yielded them a cost savings, and eliminated 40 percent of the carbon emissions associated with the asphalt’s procurement. In this case, not only was wasteful spending eliminated, but literal waste (pollution) was prevented.

In taking these steps, companies can meaningfully improve their bottom lines, while at the same time doing something worthwhile and beneficial for the planet. That, in turn, helps burnish their reputations. A remarkable plurality of consumers, 88 percent of Americans surveyed in a 2017 study to be exact, said they would be more loyal to a company that supports social or environmental issues.

It is therefore clear that any steps a company can take which save money, improve worker satisfaction, and yield increased positivity in the marketplace are well worth pursuing.

What is the Future of QMS?

Until the 2000s, quality management systems were just that: systems of desirable practices, outlined by individuals and implemented individually. That was the age of the gurus: the visionaries who outlined the systems. But what that age lacked was a practical and easy means for companies, sometimes located far away from direct guidance by the gurus, to implement their teachings.

In the intervening years, technology has radically changed that dynamic. Today, QMS software fills the marketplace, allowing businesses small and large to design and guide their quality management plans. But even these software solutions have not yet solved the last great challenge: personalized assistance in putting standards into practice.

That is why the latest innovations, particularly in artificial intelligence, have the potential to upend the equation. Already, major companies have started to use artificial intelligence in connection with QMS datasets managed by software, utilizing the programs for statistical analysis, suggested improvements, and even prediction of potential faults before they occur.

These are immensely valuable opportunities, hence why huge players such as Honeywell are spending billions of dollars to bring innovative AI technology companies into their platforms to refine existing QMS systems.

But while AI has already begun to significantly affect the biggest players, small and mid-sized companies remain eager, but not yet able, to take full advantage. It is thus the next great revolution for a new evolution of QMS, one which will bring these emerging technologies to all companies, regardless of size or scale. The future of QMS, and therefore the future of efficiency in business, rests upon this shift from companies being the recipients of ‘guru knowledge,’ to themselves being the designers of their own quality-minded futures.

------

Onega Ulanova is the CEO of QMS2GO, a provider of quality management systems leveraging AI in manufacturing.

This article originally ran on InnovationMap.

Ad Placement 300x100
Ad Placement 300x600

CultureMap Emails are Awesome

HETI leader reflects on four years focused on the energy transition

The View From HETI

It’s hard to believe only four years have passed since I joined the Greater Houston Partnership to lead the Houston Energy Transition Initiative. We talk about COVID years feeling simultaneously like just yesterday and a lifetime ago, and my time at HETI feels quite similar. The energy and climate landscape has evolved dramatically over the last few years, but one constant has endured…Houston remains at the heart of the energy industry.

When I joined HETI at the start of 2022, I could not have imagined the impact we would have over the next four years. Our vision, laid out in 2020 by then Partnership Chair Bobby Tudor, was bold – leverage Houston’s energy leadership to accelerate global solutions for an energy abundant, low-carbon future. Our mission was clear – ensure the long-term economic competitiveness of the Houston region in a changing energy landscape. Our strategy, developed with strong support from McKinsey, was sound – build a coalition of industry, academic and community partners to position Houston as a leading hub for energy and climate innovation, commercialization and economic growth.

And Houston’s energy leaders, across business, academia and community, came together to deliver that strategy. Bobby Tudor’s timely vision created an anchor for Houston’s energy community that allowed us to broaden the energy and climate discussion to one focused on meeting growing energy demand and reducing emissions. That vision was a catalyst for changing the perception of Houston from oil & gas capital to an “all of the above” energy capital with a thriving energy and climate tech innovation ecosystem. Houston is where you can work collaboratively to build new energy value chains and ensure an affordable, secure, and lower emissions energy mix for our region, our country, and the world.

In this fast-moving energy environment, it can be easy to focus on what to do next without taking time to reflect on the work and celebrate the wins. We took the opportunity at the end of 2025 to reflect on five years of impact. As we developed that report, I thought about how fortunate I have been to lead this work for the last four years. HETI’s impact has been incredibly meaningful for our region, our members, and the new companies and founders building the next generation of energy technology and talent. As my time as executive director comes to an end, HETI’s work continues. The need remains clear: more energy, less emissions, and continued collaboration that brings together innovation and infrastructure, policies and markets, and hard tech together with human ingenuity.

As I prepare to hand the reigns to HETI’s new SVP and executive cirector, Sophia Cunningham, at the end of March, I have been reflecting on the impact this role has had on me. Four years ago, I could never have imagined the opportunities, experiences and relationships this role has enabled. I am truly grateful for the support and engagement of Houston’s business and community leaders, the visionary leadership of Bobby Tudor, Scott Nyquist, HETI Members, and the Greater Houston Partnership in creating this initiative at exactly the right moment in time. I am incredibly proud of the HETI and partnership team members who have delivered with purpose and passion, and I greatly appreciate Houston’s energy and climate leaders and champions who have supported my agenda, challenged my thinking, broadened my perspectives, and worked with HETI to demonstrate the power of partnership in developing, innovating and advancing the ideas and technologies needed to meet this challenge for our region and the world.

As excited as I am to see where this next chapter takes me, I am even more excited to see where the work of HETI goes from here. I am still optimistic that the challenge of providing the world with affordable, reliable, and resilient energy while simultaneously reducing emissions is one Houston is uniquely positioned to meet.

———

This article originally appeared on the Greater Houston Partnership's Houston Energy Transition Initiative blog. HETI exists to support Houston's future as an energy leader. For more information about the Houston Energy Transition Initiative, EnergyCapitalHTX's presenting sponsor, visit htxenergytransition.org.

Texas could topple Virginia as biggest data-center market by 2030, JLL report says

data analysis

Everything’s bigger in Texas, they say—and that phrase now applies to the state’s growing data-center presence.

A new report from commercial real estate services provider JLL says Texas could overtake Northern Virginia as the world’s largest data-center market by 2030. Northern Virginia is a longtime holder of that title.

What’s driving Texas’ increasingly larger role in the data-center market? The key factor is artificial intelligence.

Companies like Google and Microsoft need more energy-hungry data centers to power AI innovations. In a 2023 article, Forbes explained that AI models consume a lot of energy because of the massive amount of data used to train them, as well as the complexity of those models and the rising volume of tasks assigned to AI.

“The data-center sector has officially entered hyperdrive,” Andy Cvengros, executive managing director at JLL and co-leader of its U.S. data-center business, said in the report. “Record-low vacancy sustained over two consecutive years provides compelling evidence against bubble concerns, especially when nearly all our massive construction pipeline is already pre-committed by investment-grade tenants.”

Dallas-Fort Worth has long dominated the Texas data-center market. But in recent years, West Texas has emerged as a popular territory for building data-center campuses, thanks in large part to an abundance of land and energy. Nearly two-thirds of data-center construction underway now is happening in “frontier markets” like West Texas, Ohio, Tennessee and Wisconsin, the JLL report says.

Northern Virginia, the current data-center champ in the U.S., boasted a data-center market with 6,315 megawatts of capacity at the end of 2025, the report says. That compares with 2,423 megawatts in Dallas-Fort Worth, 1,700 megawatts in the Austin-San Antonio corridor, 200 megawatts in West Texas, and 164 megawatts in Houston.

Fervo taps into its hottest-ever geothermal reservoir

heat record

Things are heating up at Houston-based geothermal power company Fervo Energy.

Fervo recently drilled its hottest well so far at a new geothermal site in western Utah. Fewer than 11 days of drilling more than 11,000 feet deep at Project Blanford showed temperatures above 555 degrees Fahrenheit, which exceeds requirements for commercial viability. Fervo used proprietary AI-driven analytics for the test.

Hotter geothermal reservoirs produce more energy and improve what’s known as energy conversion efficiency, which is the ratio of useful energy output to total energy input.

“Fervo’s exploration strategy has always been underpinned by the seamless integration of cutting-edge data acquisition and advanced analytics,” Jack Norbeck, Fervo’s co-founder and chief technology officer, said in a news release. “This latest ultra-high temperature discovery highlights our team’s ability to detect and develop EGS sweet spots using AI-enhanced geophysical techniques.”

Fervo says an independent review confirms the site’s multigigawatt potential.

The company has increasingly tapped into hotter and hotter geothermal reservoirs, going from 365 degrees at Project Red to 400 degrees at Cape Station and now more than 555 degrees at Blanford.

The new site expands Fervo’s geologic footprint. The Blanford reservoir consists of sedimentary formations such as sandstones, claystones and carbonates, which can be drilled more easily and cost-effectively than more commonly targeted granite formations.

Fervo ranks among the top-funded startups in the Houston area. Since its founding in 2017, the company has raised about $1.5 billion. In January, Fervo filed for an IPO that would value the company at $2 billion to $3 billion, according to Axios Pro.