As generative AI asks for more power, data centers seek more reliable, cleaner energy solutions

The tech industry should optimize infrastructure, rethink chip design, and collaborate with electricity providers to help secure a more sustainable future for data centers

Kate Hardin

United States

Gillian Crossan

United States

AI-driven data center power consumption will continue to surge, but data centers are not—in fact—that big a part of global energy demand. Deloitte predicts data centers will only make up about 2% of global electricity consumption, or 536 terawatt-hours (TWh), in 2025. But as power-intensive generative AI (gen AI) training and inference continues to grow faster than other uses and applications, global data center electricity consumption could roughly double to 1,065 TWh by 2030 (figure 1).1 To power those data centers and reduce the environmental impact, many companies are looking to use a combination of innovative and energy-efficient data center technologies and more carbon-free energy sources.

Nonetheless, it’s an uphill task for power generation and grid infrastructure to keep pace with a surge in electricity demand from AI data centers. Electricity demand was already growing fast due to electrification—the switch from fossil-fueled to electric-powered equipment and systems in the transport, building, and industrial segments—and other factors. But gen AI is an additional, and perhaps, an unanticipated source of demand. Moreover, data centers often have special requirements as they need 24/7 power supply with high levels of redundancy and reliability, and they’re working to have it be carbon-free.

Estimating global data centers’ electricity consumption in 2030 and beyond is challenging, as there are many variables to consider. Our assessment suggests that continuous improvements in AI and data center processing efficiency could yield an energy consumption level of approximately 1,000 TWh by 2030. However, if those anticipated improvements do not materialize in the coming years, the energy consumption associated with data centers could likely rise above 1,300 TWh, directly impacting electricity providers and challenging climate-neutrality ambitions.2 Consequently, driving forward innovations in AI and optimizing data center efficiency over the next decade will be pivotal in shaping a sustainable energy landscape.

Some parts of the world are already facing issues in generating power and managing grid capacity in the face of growing electricity demand from AI data centers.3 Critical power to support data centers’ most important components—including graphics processing unit (GPU) and central processing unit (CPU) servers, storage systems, cooling, and networking switches—is expected to nearly double between 2023 and 2026 to reach 96 gigawatts (GW) globally by 2026; and AI operations alone could potentially consume over 40% of that power.4 Worldwide, AI data centers’ annual power consumption is expected to reach 90 terawatt-hours by 2026 (or roughly one-seventh of the predicted 681 TWh of all data centers globally), roughly a tenfold increase from 2022 levels.5 As such, gen AI investments are fueling demand for so much electricity that in the first quarter of 2024, global net additional power demand from AI data centers was roughly 2 GW, an increase of 25% from the fourth quarter of 2023 and more than three times the level from the first quarter of 2023.6 Meeting data center power demand can be challenging because data center facilities are often geographically concentrated (especially in the United States) and their need for 24/7 power can burden existing power infrastructure.7

Deloitte predicts that both the technology and electric power industries can and will jointly address these challenges and contain the energy impact of AI—more specifically, gen AI. Already, many big tech and cloud providers are investing in carbon-free energy sources and pushing for net-zero targets,8 demonstrating their commitment to sustainability.

Hyperscalers plan massive expansion of gen AI data centers to help support growing customer demand

The surge in electricity demand is largely due to hyperscalers’ plans to build out data center capacity, globally.9 As AI demand—specifically gen AI—is expected to grow, companies and countries are racing to build more data centers to meet that demand. Governments are also establishing sovereign AI capabilities to maintain tech leadership.10 The data center real estate build-out has reached record levels based on select major hyperscalers’ capital expenditure, which is trending at roughly US$200 billion as of 2024, and estimated to exceed US$220B by 2025.11

Moreover, Deloitte’s State of Generative AI in the Enterprise survey noted that enterprises have been mostly piloting and experimenting until now.12 But as they experiment with getting value from gen AI, respondents are seeing tangible results and so intend to quickly scale up beyond pilots and proofs of concept. If usage grows as the technology matures, hyperscalers’ and cloud providers’ capital expenditure will most likely remain high through 2025 and 2026.

Two broad areas drive most of the electricity consumption in a data center: computing power and server resources like server systems (roughly 40% data center power consumption) and cooling systems (consume 38% to 40% power). These two are the most energy-intensive components even in AI data centers, and they will continue to fuel data centers’ power consumption. Internal power conditioning systems consume another 8% to 10%, while network and communications equipment and storage systems use about 5% each, and lighting facilities usually use 1% to 2% of power (figure 2).13 With gen AI requiring massive amounts of power, data center providers—including the hyperscalers and data center operators—need to look at alternate energy sources, new forms of cooling, and more energy-efficient solutions when designing data centers. Several efforts are already underway.

Gen AI is contributing to increased electricity demand

Data centers’ energy consumption has been surging since 2023, thanks to exploding demand for AI.14 Deploying advanced AI systems requires vast numbers of chips and processing capacity, and training complex gen AI models can require thousands of GPUs.

Hyperscalers and large-scale data center operators that are supporting gen AI and high-performance computing environments require high-density infrastructure to support computing power. Historically, data centers relied mainly on CPUs, which ran at roughly 150 watts to 200 watts per chip.15 GPUs for AI ran at 400 watts until 2022, while 2023 state-of-the-art GPUs for gen AI run at 700 watts, and 2024 next-generation chips are expected to run at 1,200 watts.16 These chips (about eight of them) sit on blades placed inside of racks (10 blades per rack) in data centers, and are using more power and producing more heat per square meter of footprint than traditional data center designs from only a few years ago.17 As of early 2024, data centers typically supported rack power requirements of 20 kW or higher. But the average power density is anticipated to increase from 36 kW per server rack in 2023 to 50 kW per rack by 2027.18

Total AI computing capacity, measured in floating-point operations per second (FLOPS), has also been increasing exponentially since the advent of gen AI. It’s grown 50% to 60% quarter over quarter globally since the first quarter of 2023 and will likely grow at that pace through the first quarter of 2025.19 But data centers don’t only measure capacity in FLOPS, they also measure megawatt hours (MWh) and TWh.

Gen AI’s multibillion parameter LLMs and the multibillion watts they consume

Gen AI large language models (LLMs) are becoming more sophisticated, incorporating more parameters (variables that enable AI learning and prediction) over time. From the 100 to 200 billion parameter models that were released initially during 2021 to 2022, current advanced LLMs (as of mid-2024) have scaled up to nearly two trillion parameters, which can interpret and decode complex images.20 And there’s competition to release LLMs with 10 trillion parameters. More parameters add to data processing and computing power needs, as the AI model must be trained and deployed. This can further accelerate demand for gen AI processors and accelerators, and in turn, electricity consumption.

Moreover, training LLMs is energy intensive. Independent research of select LLMs that were trained on more than 175 billion parameters of data sets demonstrated that they consumed anywhere between 324 MWh and 1,287 MWh of electricity for each training run … and models are often retrained.21

On average, a gen AI–based prompt request consumes 10 to 100 times more electricity than a typical internet search query.22 Deloitte predicts that if only 5% of daily internet searches, globally, use gen AI– based prompt requests, it would require approximately 20,000 servers (with eight specialized GPU cores in each of the servers) that consume 6.5 kW on an average per server to fulfill the prompt requests, amounting to an average daily electricity consumption of 3.12 GWh and annual consumption of 1.14 TWh23—which is equivalent to annual electricity consumed by approximately 108,450 US households.24

Data center demand could present challenges and opportunities for power sector transition

The electric power sector was already planning for rising demand: Many in the industry predicted as much as a tripling of electricity consumption by 2050 in some countries.25 But that trajectory has recently accelerated in some areas due to burgeoning data center demand. Previous forecasts in many countries have projected rising power demand due to electrification as well as increasing data center consumption and overall economic growth. But recent sharp spikes in data center demand, which could be just the tip of the iceberg, reveal the growing magnitude of the challenge.26

Round-the-clock, carbon-free electricity that many tech companies seek can be hard to come by, especially in the short term.

This comes against the backdrop of a multi-decade power industry transformation, as electric companies build, upgrade, and decarbonize electric grid infrastructure, and digitalize systems and assets. In many areas, electric companies are also hardening assets against increasingly severe weather and climate events and protecting networks from rising cybersecurity threats.27 Electric power grids in some countries are struggling to keep up with demand, especially for low- or zero-carbon electricity. In the United States, data centers are anticipated to represent 6% (or 260 TWh) of total electricity consumption in 2026.28 The United Kingdom may witness a sixfold growth in electricity demand within a period of just 10 years, largely due to AI.29 In China, data centers—including the ones that power AI—will likely make up 6% of the country’s total electricity demand by 2026.30 Data centers could also add to China’s pollution problem, since the country’s power generation is dominated by coal, which accounted for 61% of its energy use and 79% of its carbon dioxide emissions in 2021.31

Some countries that are facing the rising demand for electricity from data centers are responding with regulations. For instance, in Ireland, existing data centers consume a fifth of the country’s total electricity consumption and this is only expected to grow further as AI-driven data centers spring up more; households are even lowering their power consumption.32 Temporarily, Ireland halted the construction of new data centers connected to the grid, but has since reversed that position.33 Like Ireland, even the city of Amsterdam halted new data center construction to support sustainable urban development.34 Singapore announced new sustainability standards for data centers that require operators to gradually increase the overall operating temperatures of their facilities to 26°C or higher. Higher operating temperatures reduce the demand for cooling and lower power consumption, but at the cost of shortening the lifespan of the chips.35

The urgency and geographic concentration of data center demand—and the requirement for 24/7 carbon-free energy—can further complicate the challenge for tech companies and electricity providers, in addition to new demand from electrification, manufacturing, and other sources. The largest data center market globally is in northern Virginia,36 and the local utility, Dominion Energy, expects power demand to grow by about 85% over the next 15 years, with data center demand quadrupling.37 The round-the-clock, carbon-free electricity that many tech companies seek can be hard to come by, especially in the short term. Electricity providers are exploring multiple avenues to help meet demand while maintaining reliability and affordability. In addition to new renewable energy and battery storage, many electricity providers have also announced plans to build natural gas–fired power plants, which are not carbon-free.38 This could potentially make it more challenging to meet utility, state, and even national decarbonization targets.39

Despite being poised to consume massive amounts of clean energy, AI could also potentially help hasten the clean energy transition: Some utilities are already using AI to enable electric grids to operate more cheaply, efficiently, and reliably through improved weather and load forecasting, enhanced grid management and renewable asset performance, faster storm recovery, better wildfire risk assessment, and more.40

Data center cooling is water-intensive

Next-generation CPUs and GPUs have higher thermal density properties than their predecessors. At the same time, some server vendors are packing more and more power-hungry chips into each rack in an endeavor to cater to the growing demand for high-performance computing and AI applications. But denser racks will demand more water, especially to cool the gen AI chips. AI data centers’ freshwater demand could be as much as 1.7 trillion gallons (at the higher end) by 2027.41 A hyperscale data center that intends to manage excess heat with air-based cooling and evaporated drinking water would require over 50 million gallons of water every year (or roughly what it takes to make 14,700 smartphones).42 This water cannot be returned to the aquifer, reservoir, or water supply where it came from.43

Air-based cooling alone uses up to 40% of a typical data center’s electricity consumption. Therefore, data centers are looking at alternatives to traditional air-based cooling methods, mainly into liquid cooling, as its higher thermal transfer properties could help cool high-density server racks and enable them to reduce power usage by as much as 90% when compared with air-based methods.44 As liquid cooling directly delivers cooling to server racks, it can support dense power racks on the order of 50 kW to 100 KW or more.45 Moreover, it may help eliminate the need for chillers, which were traditionally used for producing the cooling water.

However, despite liquid cooling technology’s promise to help save energy across the data center stack,46 it’s still in its early days and is yet to be widely adopted or integrated into AI data centers, globally.47 Moreover, water is a finite resource, and therefore its cost and availability will likely affect future decisions about its usage. 

The tech industry is moving toward more sustainable solutions and carbon-free sources

To help expedite the move toward using carbon-free sources to power AI data centers, tech industry majors continue to be aggressive in their pursuit of renewable energy by way of power purchase agreements, or long-term contracts with renewable energy providers.48 These deals have helped bankroll renewable energy projects by enabling them to secure financing. In some cases, technology companies are working with electricity providers and innovators to help test and scale promising energy technologies, including advanced geothermal, advanced wind and solar technologies, hydropower, and even underwater data centers.  

In some areas, local grid constraints and long interconnection times for new renewable and battery storage facilities are causing delays in connecting these resources to the electric grid.49 These delays, which can be as long as five years in the United States, are often due to high demand and insufficient transmission infrastructure. As a result, tech companies are increasingly pursuing onsite, sometimes off-grid, energy solutions.50 Additionally, they are investing in new technologies such as long-duration energy storage and small modular nuclear reactors to help address these challenges. In some cases, tech companies and utilities are planning to coordinate to bring innovative clean energy technologies to scale, which could eventually benefit other organizations and help decarbonize the electric grid faster.51 Many of these research and development programs, pilots and other clean energy investments may take years before reaping benefits, demonstrating return on investment, and becoming commercially viable.52 For example, small, modular nuclear reactors are still in early development stages, and may not be a near-term zero-carbon solution.53

The technology sector consistently dominates US corporate renewable procurement and accounted for more than 68% of the nearly 200 deals of associated contracted capacity tracked over the 12 months prior to Feb. 28, 2024.54 Similarly, hyperscalers and data center operators in India are increasingly using solar to power their data centers in the region.55 Without these purchase commitments, many renewable energy projects would not be built.56

As such, the tech industry’s role in bankrolling clean energy technologies to help bring them to scale will continue to be valuable. In some cases, they’re working directly with innovators and renewable energy producers, and in other cases, they’re partnering with utilities.57 Importantly, the way tech companies inject capital to help advance the clean energy transition is critical, as neither the innovators nor the power industry would typically have the level of financial resources that the tech industry possesses.

Bottom line

What should the broader tech industry, hyperscalers, data center operators, utilities, and regulators do to help meet gen AI demand sustainably? Several considerations for hyperscalers and the broader tech industry are more or less in line with what we presented in Deloitte Global’s 2021 prediction on cloud migration.58 Though demand drivers may have changed and the pace of change has accelerated, the industry is working to achieve balance between sustainability and expedited time to market, while keeping data centers rising energy demands under control and finding more sustainable ways to power AI—specifically gen AI.

Do we need to keep racing to build bigger and bigger foundational models (for example, more than a trillion parameter models), or are smaller models sufficient, while being more sustainable?

1. Make gen AI chips more energy-efficient: Already, a new generation of AI chips can perform AI training in 90 days, consuming 8.6 GWh. This would be less than one-tenth the energy that the previous generation of chips takes to do the same function on the same data.59 Chip companies should continue to work with the broader semiconductor ecosystem to help intensify focus on improving FLOPS/watt performance, such that, future chips can train AIs several times larger than the largest AI systems currently available while using less electricity.

2. Optimize gen AI uses and shift processing to edge devices: This includes assessing whether it’s energy-efficient to do training and inference in the data center, or on the edge device, and accordingly rebalance data center equipment needs. Edge not only can support applications where response times are critical, but also for those use cases where sensitive data is involved, or privacy needs are high. It also helps save network and server bandwidth, rerouting gen AI workloads to local and near-location or co-location devices, while only transmitting select AI workloads to data centers.60

3. Implement changes in gen AI algorithms and rightsize AI workloads: Do we need to keep racing to build bigger and bigger foundational models (for example, more than a trillion parameter models), or are smaller models sufficient, while being more sustainable? Already, startups are developing on-device multimodal AI models, which do not require energy-intensive computations in the cloud.61 Customers should fine-tune and adjust the size of their AI workloads and go for targeted gen AI models (including preexisting models and training only when needed) based on real business needs, which can minimize energy use. Additionally, depending on specific needs with AI inferencing (for example, doing inference in real time and when latency is critical), CPUs can be more advantageous and efficient.62

4. Form strategic partnerships to serve local and cluster-level AI data center needs: For several small to midsize organizations (including universities) that may find it hard to tap into gen AI data center capacity, those organizations should work with specialized data center operators and cloud service providers that focus on delivering high-performance computing solutions for smaller high-performance computing GPU-cluster co-locations.63 A corollary: Data centers can then actively track usage and availability for potential opportunities and demand pockets to help deliver near-term co-location services.

5. Collaborate with other stakeholders and sectors to make an overall positive environmental impact: The various ecosystem players—including hyperscalers, their customers, third-party data center operators and co-location service providers, electricity providers, the local regulators and municipalities, and the real estate firms—should have ongoing conversations around what’s feasible and viable for the business, environment, and society.64 That collaboration should encompass multiple aspects including determining potential strategic co-location needs (where a data center company rents computing and server resources to one or more companies), assessing cooling needs such as adequate temperatures in liquid cooling systems, identifying solutions to manage heat and wastewater, and figuring out recycling needs. For example, in Europe, data center operators are directing waste heat to warm swimming pools in the vicinity.65 Electricity providers should consider working more closely with the tech industry to understand how to meet data center energy demand, while identifying ways the tech companies could potentially help fund and scale new energy technologies, which is a vital step to bring more clean energy to the grid.

The holistic efforts of hyperscalers and electricity providers to help increase the use of carbon-free sources to power data centers—including the ones being built exclusively for gen AI—may bear fruit in the longer term.

by

Kate Hardin

United States

Gillian Crossan

United States

Endnotes

  1. Deloitte analysis based on publicly available information sources and conversations with industry specialists. We used base electricity consumption data from the US Energy Information Administration’s (EIA) International Energy Outlook 2023 data on total electricity usage across residential, commercial, industrial, and transportation end uses (a reference to US Energy Information Administration, “Table: Delivered energy consumption by end-use sector and fuel,” accessed Nov. 4, 2024) to arrive at estimates and prediction values for global data centers’ electricity consumption (TWh) between 2022 and 2030. Our estimates and projections for data centers’ percent electricity consumption of global total are based on our research of multiple publicly available sources including SemiAnalysis, EPRI, Goldman Sachs, Bloomberg, and Latitude Media, and further validated based on our conversations with subject matter specialists in the areas of technology, energy, and sustainability. Total energy consumption by end-use sector and fuel (as noted from the aforementioned table from EIA’s International Energy Outlook 2023 data), globally, is estimated and forecast at 26,787 TWh in 2025, 27,256 TWh in 2026, and 29,160 TWh in 2030— increasing from 25,585 TWh back in 2022.

    View in Article
  2. As noted in endnote 1 above, we arrived at 2022 to 2030 data, estimates, and predictions based on a combination of in-depth secondary research of multiple publicly available sources, and validated further from our discussions with subject matter specialists. Also, see Prof. Dr. Bernhard Lorentz, Dr. Johannes Trüby, and Geoff Tuff, "Powering artificial intelligence," Deloitte Global, November 2024." 

    View in Article
  3. One-fifth of Ireland’s electricity is consumed by data centers, and this is expected to grow, even as households are lowering their electricity use. To read further, see: Chris Baraniuk, “Electricity grids creak as AI demands soar,” BBC, May 21, 2024.

    View in Article
  4. Dylan Patel, Daniel Nishball, and Jeremie Eliahou Ontiveros, “AI data center energy dilemma: Race for AI data center space,” SemiAnalysis, March 13, 2024.

    View in Article
  5. Ibid.

    View in Article
  6. Data center BMO report, Communications Infrastructure, “1Q24 data center leasing: Records are made to be broken,” April 28, 2024; Moreover, due to strong demand from cloud providers and AI workloads, the data center primary market supply in the United States alone was up 26% year over year to 5.2 GW in 2023, and more are under construction. See further: CBRE, “North America data center trends H2 2023,” March 6, 2024.

    View in Article
  7. Lisa Martine Jenkins and Phoebe Skok, “Mapping the data center power demand problem, in three charts,” Latitude Media, May 31, 2024.

    View in Article
  8. Based on our analysis of multiple publicly available information and reports from what companies self-report, and further validated from third-party sources.

    View in Article
  9. For context, hyperscalers are large cloud service providers and data centers that offer huge amounts of computing and storage resources typically at enterprise scale. See: Synergy Research Group, “Hyperscale operators and colocation continue to drive huge changes in data center capacity trends,” Aug. 7, 2024.

    View in Article
  10. Yifan Yu, “AI’s looming climate cost: Energy demand surges amid data center race,” Nikkei Asia, June 12, 2024.

    View in Article
  11. Data center BMO report, Communications Infrastructure, “1Q24 data center leasing: Records are made to be broken,” April 28, 2024. Further, Deloitte analysis based on information from select tech companies’ publicly available sources such as earnings releases and Dell’Oro Group’s market research data on data center IT capital expenditure shows that if we consider the capital expenditure spending of other data center providers, including third-party operators and outsourced cloud service providers, data centers’ aggregate capital expenditure spending could be at least US$250 billion in 2025. See: Baron Fung, “Market research on data center IT capex,” Dell’oro Group, accessed Nov. 4, 2024.

    View in Article
  12. Nitin Mittal, Costi Perricos, Brenna Sniderman, Kate Schmidt, and David Jarvis, “Now decides next: Getting real about generative AI,” Deloitte’s State of Generative AI in the Enterprise quarter two report, Deloitte, April 2024.

    View in Article
  13. Deloitte analysis based on publicly available research reports including: Wania Khan, Davide De Chiara, Ah-Lian Kor, and Marta Chinnici, “Advanced data analytics modeling for evidence-based data center energy management,” Physica A 624, 2023; Kazi Main Uddin Ahmed, Math H. J. Bollen, and Manuel Alvarez, “A review of data centers energy consumption and reliability modeling,” in IEEE Access 9, 2021: pp. 152536–152563.

    View in Article
  14. Tom Dotan and Asa Fitch, “Why the AI industry’s thirst for new data centers can’t be satisfied,” The Wall Street Journal, April 24, 2024.

    View in Article
  15. Noam Brousard, “Examining the impact of chip power reduction on data center economics,” Semiconductor Engineering, March 12, 2024.

    View in Article
  16. Based on our analysis of multiple publicly available sources including: Michael Studer, “The energy challenge of powering AI chips,” Robeco, Nov. 6, 2023; Agam Shah, “Generative AI to account for 1.5% of world’s power consumption by 2029,” HPCwire, July 8, 2024.

    View in Article
  17. From our study and analysis of select gen AI data center chip solutions offered by major AI chip vendors, further corroborated with publicly available third-party sources including: Beth Kindig, “AI power consumption: Rapidly becoming mission-critical,” Forbes, June 20, 2024.

    View in Article
  18. Jones Lang LaSalle, “Data centers 2024 global outlook,” Jan. 31, 2024; Doug Eadline, “The gen AI data center squeeze is here,” HPCwire, Feb. 1, 2024; Per IDC, besides graphics processing unit, servers, data centers also need to grapple with a corresponding growth in storage capacity, which is likely to double between 2023 and 2027 to reach 21 zettabytes in 2027. See: John Rydning, “Worldwide Global StorageSphere forecast, 2023 to 2027: Despite decreased petabyte demand near term, the installed base of storage capacity continues to grow long term,” IDC Corporate, May 2023.

    View in Article
  19. Patel, Nishball, and Eliahou Ontiveros, “AI data center energy dilemma.”

    View in Article
  20. Sean Michael Kerner, “What are large language models?” TechTarget, May 2024; Yu, “AI’s looming climate cost.”

    View in Article
  21. Alex de Vries, “ The growing energy footprint of artificial intelligence,” Joule 7, no. 10 (2023): pp. 2191–2194.

    View in Article
  22. Eren Çam, Zoe Hungerford, Niklas Schoch, Francys Pinto Miranda, and Carlos David Yáñez de León, “Electricity 2024: Analysis and forecast to 2026 report,” International Energy Agency, accessed Nov. 4, 2024.

    View in Article
  23. Deloitte analysis based on publicly available reports and sources including: de Vries “The growing energy footprint of artificial intelligence,” pp. 2191–2194.

    View in Article
  24. Deloitte analysis based on data related to energy use and electricity consumption in homes in the United States. See: US Energy Information Administration, “Use of energy explained,” accessed Dec. 18, 2023.

    View in Article
  25. Darren Sweeney, “Utility execs prepare for ‘tripling’ of electricity demand by 2050,” S&P Global, April 19, 2023.

    View in Article
  26. Robert Walton, “US electricity load growth forecast jumps 81% led by data centers,” Utility Dive, Dec. 13, 2023.

    View in Article
  27. Aaron Larson, “How utilities are planning for extreme weather events and mitigating risks,” POWER, March 13, 2024.

    View in Article
  28. Çam, Hungerford, Schoch, Miranda, and de León, “Electricity 2024.”

    View in Article
  29. Baraniuk, “Electricity grids creak as AI demands soar.”

    View in Article
  30. Yu, “AI’s looming climate cost.”

    View in Article
  31. Data on China’s energy use and CO2 emissions sourced from International Energy Agency, accessed September 25, 2024. See: International Energy Agency, “China’s energy use,” accessed Nov. 4, 2024; International Energy Agency, “China’s CO2 emissions,” accessed Nov. 4, 2024.

    View in Article
  32. Baraniuk, “Electricity grids creak as AI demands soar.”

    View in Article
  33. Paul O’Donoghue, “Build it and they will hum: What next for Ireland and data centers?The Journal, Sept. 2, 2024.

    View in Article
  34. Hosting Journalist, “City of Amsterdam puts halt to new data center construction,” Dec. 21, 2023.

    View in Article
  35. With every 1C increase, operators could save 2% to 5% on the energy they use for cooling equipment. To read further, see: Inno Flores, “Singapore unveils green data center road map amid AI boom that strains energy resources,” Tech Times, May 30, 2024.

    View in Article
  36. Julie R. Peasley, “Ranked: Top 50 data center markets by power consumption,” Visual Capitalist, Jan. 10, 2024.

    View in Article
  37. Whitney Pipkin, “Energy demands for Northern Virginia data centers almost too big to compute,” Bay Journal, June 18, 2024.

    View in Article
  38. Zach Bright, “Southeast utilities have a ‘very big ask’: More gas,” E&E News, Jan. 22, 2024.

    View in Article
  39. Ibid.

    View in Article
  40. Robert Walton, “AI is enhancing electric grids, but surging energy use and security risks are key concerns,” Utility Dive, Oct. 23, 2023.

    View in Article
  41. Karen Hao, “AI is taking water from the desert,” The Atlantic, March 1, 2024.

    View in Article
  42. Deloitte analysis based on publicly available information sources including: Jennifer Billock, “Photos: How much water it takes to create 30 common items,” North Shore News, Jan. 19, 2023.

    View in Article
  43. Hao, “AI is taking water from the desert”; One case in point is China—where its data centers’ annual water consumption is expected to increase from around 1.3 billion cubic meter as of 2023 to over 3 billion cubic meter by 2030. To read further, see: Yu, “AI’s looming climate cost.” 

    View in Article
  44. Eadline, “The gen AI data center squeeze is here.”

    View in Article
  45. Diana Goovaerts, “Data center operators want to run chips at higher temps. Here’s why.” Fierce Network, June 11, 2024.

    View in Article
  46. Scott Wilson, “Is immersion cooling the answer to sustainable data centers?” Ramboll, Dec. 13, 2023.

    View in Article
  47. David Eisenband, “100+ kW per rack in data centers: The evolution and revolution of power density,” Ramboll, March 13, 2024; Direct-to-chip cooling (also known as cold plate liquid cooling or direct liquid cooling) cools down servers by distributing heat directly to server components, while, immersion cooling involves submerging servers and components in a liquid dielectric coolant that also helps prevent electric discharge.

    View in Article
  48. Based on Deloitte’s analysis of developments and announcements from select major cloud hyperscalers and tech companies— and information gathered from publicly available sources (time period: 2023 and first three quarters of 2024).

    View in Article
  49. Joseph Rand, Nick Manderlink, Will Gorman, Ryan Wiser, Joachim Seel, Julie Mulvaney Kemp, Seongeun Jeong, and Fritz Kahrl, “Queued up: 2024 edition,” Lawrence Berkeley National Laboratory, April 2024.

    View in Article
  50. Based on Deloitte’s analysis of developments and announcements from select major cloud hyperscalers and tech companies— and information gathered from publicly available information sources between the first quarter of 2023 and the third quarter of 2024.

    View in Article
  51. Julian Spector, “Duke Energy wants to help Big Tech buy the 24/7 clean energy it needs,” Canary Media, June 11, 2024.

    View in Article
  52. For example, it’s not easy to submerge and drop a 1,300-ton data center unit underwater, especially since it demands special equipment to withstand pressure and corrosion caused by seawater. Moreover, there are concerns related to its impact on marine life.

    View in Article
  53. David Schlissel and Dennis Wamsted, “Small modular reactors: Still too expensive, too slow, and too risky,” Institute for Energy Economics and Financial Analysis, May 2024.

    View in Article
  54. Deloitte's analysis of data and information gathered from multiple reports from S&P Global Market Intelligence, published during March and August 2024.

    View in Article
  55. Manish Kumar, “India’s data center boom opens up a fresh segment for green developers,” Saur Energy International, July 1, 2024.

    View in Article
  56. Naureen S. Malik and Bloomberg, “With AI forcing data centers to consume more energy, software that hunts for clean electricity across the globe gains currency,” Fortune, Feb. 25, 2024.

    View in Article
  57. Based on Deloitte’s analysis of developments and announcements from select major cloud hyperscalers, tech companies, and power and utility players—on publicly available information sources during 2023 and the first three quarters of 2024.

    View in Article
  58. Duncan Stewart, Nobuo Okubo, Patrick Jehu, and Michael Liu, “The cloud migration forecast: Cloudy with a chance of clouds,” Deloitte Insights, Dec. 7, 2020.

    View in Article
  59. Wylie Wong, “Nvidia launched next-generation Blackwell GPUs amid AI ‘arms race’,” Data Center Knowledge, March 19, 2024; For instance, Nvidia notes that it can train a very large AI model using 2,000 Grace Blackwell chips in 90 days, consuming 4 MW power. In comparison, it would take as much as 8,000 of the previous generation chips to do the same work within the same time, consuming 15 MW power.

    View in Article
  60. To read further, see section “Generative AI comes to the enterprise edge: ‘On prem AI’ is alive and well” in our 2025 TMT Predictions chapter on “Updates”; Additionally, see: Sabuzima Nayak, Ripon Patgiri, Lilapati Waikhom, and Arif Ahmed, “A review on edge analytics: Issues, challenges, opportunities, promises, future directions, and applications,” Digital Communications and Networks 10, no. 3 (2024): pp. 783–804.

    View in Article
  61. Yu, “AI’s looming climate cost.”

    View in Article
  62. Luke Cavanagh, “GPUs vs. CPUs in the context of AI and web hosting platforms,” Liquid Web, Aug. 20, 2024.

    View in Article
  63. Eadline, “The gen AI data center squeeze is here.”

    View in Article
  64. Goovaerts, “Data center operators want to run chips at higher temps. Here’s why.

    View in Article
  65. Baraniuk, “Electricity grids creak as AI demands soar.”

    View in Article

Acknowledgments

The authors would like to thank Dilip Krishna, Marlene Motyka, Jim Thomson, Adrienne Himmelberger, Thomas Schlaak, Freedom-Kai Phillips, Johannes Truby, Clement Cabot, Negina Rood, Ankit Dhameja, Suzanna Sanborn, and Akash Chatterji for their contributions to this article.

Cover image by: Jaime Austin; Getty Images, Adobe Stock

Copyright and legal information

This article contains general information and predictions only and Deloitte is not, by means of this article, rendering accounting, business, financial, investment, legal, tax, or other professional advice or services. This article is not a substitute for such professional advice or services, nor should it be used as a basis for any decision or action that may affect your business. Before making any decision or taking any action that may affect your business, you should consult a qualified professional advisor.

Deloitte shall not be responsible for any loss sustained by any person who relies on this article.

About Deloitte

Deloitte refers to one or more of Deloitte Touche Tohmatsu Limited, a UK private company limited by guarantee (“DTTL”), its network of member firms, and their related entities. DTTL and each of its member firms are legally separate and independent entities. DTTL (also referred to as “Deloitte Global”) does not provide services to clients. In the United States, Deloitte refers to one or more of the US member firms of DTTL, their related entities that operate using the “Deloitte” name in the United States and their respective affiliates. Certain services may not be available to attest clients under the rules and regulations of public accounting. Please see www.deloitte.com/about to learn more about our global network of member firms.