Your browser does not support JavaScript!

The Impact of Energy Consumption on AI Development and Data Centers

GOOVER DAILY REPORT 6/7/2024
goover

TABLE OF CONTENTS

  1. Introduction
  2. Energy Demands of Generative AI
  3. Impact on Data Center Infrastructure
  4. Environmental Impact
  5. Future Outlook and Mitigation Strategies
  6. Glossary
  7. Conclusion
  8. Source Documents

1. Introduction

  • This report explores the significant impact of energy consumption on the development of generative AI and the operational challenges it poses to data centers worldwide.

2. Energy Demands of Generative AI

  • 2-1. Introduction to the Energy Problem

  • Generative AI requires a massive amount of power, leading to significant strains on data center infrastructure. Increased focus on generative AI has exacerbated a scarcity of data center colocation supply due to regional power limitations. Power demands are being pushed to the forefront as AI continues to evolve, exacerbating the global electricity consumption issue.

  • 2-2. Increase in Power Needs for Data Centers

  • Data center operators are now facing power requirements ranging from 300 megawatts (MW) to 500MW due to the increased computational demands of generative AI. This has caused a shift in site selection criteria and daily management, as the infrastructure needs to be robust enough to support these power requirements. Predictions indicate that electricity demand from data centers could more than double by 2028 in regions like Ireland.

  • 2-3. Generative AI as a Major Contributor to Rising Electricity Consumption

  • Generative AI, particularly models like GPT-4 which can have up to 1.7 trillion parameters, is a critical factor in the rise of global electricity consumption. The development and maintenance of these AI models consume vast amounts of power, increasing the strain on already struggling power grids. For example, data centers supporting AI and cryptocurrency sectors could see their electricity consumption double by 2026.

  • 2-4. Challenges Posed by High-Performance Computing and AI Workloads

  • The computational power required for generative AI creates significant heat, necessitating sophisticated cooling solutions. Traditional air cooling systems are often inadequate, leading to a preference for liquid cooling in high-performance scenarios. This heat management challenge is compounded by the older infrastructure of many data centers, such as those in Europe where a significant portion of the grid is over 40 years old and requires modernization. Additionally, the increased data generation exacerbates these challenges, necessitating more efficient power and cooling infrastructure investments to prevent hardware damage and maintain performance.

3. Impact on Data Center Infrastructure

  • 3-1. Strain on Global Power Infrastructure

  • Power and storage requirements for data centers are growing exponentially, creating significant problems for the industry. Generative AI, which demands a tremendous amount of power, exacerbates the scarcity of data center colocation supply caused by regional power limitations. According to JLL’s global outlook report for 2024, increasing electricity consumption from high-performance computing (HPC) and AI workloads is putting pressure on data centers' energy infrastructure. Additionally, global data infrastructure, particularly in Europe, is aging, with one-third of Europe’s grid being over 40 years old, necessitating approximately $641 billion for modernization.

  • 3-2. Rising Demand and Site Selection for AI Data Centers

  • The demand for AI-specialized data centers, which require more densely clustered and performance-intensive infrastructure, is rising. Data center operators are now facing power requirements ranging from 300 megawatts (MW) to 500MW, shifting the criteria for site selection and daily management practices. The JLL report suggests that AI-specialized data centers have structural and operational differences from conventional facilities. The global electricity consumption for generative AI is expected to increase significantly, with regions like Ireland and Denmark anticipating substantial growth in data center electricity usage by 2030.

  • 3-3. Necessity of Advanced Cooling Systems

  • The heat generated by AI workloads necessitates the use of sophisticated cooling techniques. Efficient and advanced cooling systems are crucial to prevent hardware damage and maintain performance. Masood, chief AI architect at UST, emphasized that options like liquid cooling are often more efficient for high-performance chips, as traditional air cooling methods may fall short. Managing this heat efficiently is essential for the robustness of power infrastructure within data centers.

  • 3-4. Challenges of Cabling and Embedded Infrastructure

  • Cabling, often overlooked, remains a critical aspect of data center design. Embedded within the structure of data centers, cabling, along with power connections and cooling systems, can be incredibly expensive to replace, if not economically impossible, leading to significant issues. According to Lindqvist, a general manager at Onnec, the lack of attention to cabling affects the overall efficiency of data centers.

4. Environmental Impact

  • 4-1. Carbon Footprint of Generative AI

  • The development of generative AI brings significant environmental challenges due to its energy consumption. For instance, the carbon emissions resulting from training Google's BERT model in 2018 were approximately 280 metric tons, equivalent to the lifetime emissions of five cars. This problem has worsened with the growth of models; GPT-4, for example, has 1.7 trillion parameters. A 2022 study indicated that training a modern language model using renewable energy sources emits at least 25 metric tons of carbon equivalents, whereas using fossil fuels like coal and natural gas, as was the case for GPT-3, spiked the emissions to 500 metric tons, equating to a million miles driven by an average gasoline-powered car.

  • 4-2. Water Usage in Data Centers

  • Data centers, essential for hosting generative AI, have increasingly high water demands, primarily for cooling purposes. Microsoft reported a 34% year-over-year increase in water usage, using an additional 1.7 billion gallons in 2022 compared to 2021. Similarly, Google used 1.2 billion more gallons in 2022 than in the previous year, amounting to a 20% increase, with 93% of this water directed to cooling their data centers. Additionally, an interaction with ChatGPT involving between 10 to 50 queries uses approximately 500 ml of water. The EU has mandated annual disclosures on water and energy use starting September 2024, highlighting growing concerns about water usage in data centers.

  • 4-3. Global Greenhouse Gas Emissions from Data Centers

  • Data centers significantly contribute to global greenhouse gas emissions, accounting for 2.5% to 3.7% of global emissions, surpassing those from the aviation industry. The daily carbon footprint of GPT-3 is estimated at 8.4 tons of CO2 per year. With generative AI increasing demand for data center capacity, the total electricity consumption of data centers could surpass 1000 terawatt-hours by 2026, according to the International Energy Agency. This consumption is comparable to the electricity needs of Japan.

  • 4-4. Efforts for Renewable Energy Integration

  • Efforts to integrate renewable energy into data center operations are underway to counterbalance the environmental impact of generative AI. Using renewable energy can significantly reduce the carbon footprint of large language models, potentially by up to 98%. Data centers are attempting to reduce their reliance on fossil fuels, yet many still derive a majority of their energy from such sources. The transition to renewable energy and advancements in energy-efficient hardware like GPUs and TPUs, which can cut energy consumption by up to 50%, are critical to making generative AI more sustainable.

5. Future Outlook and Mitigation Strategies

  • 5-1. Importance of Energy-Efficient Hardware

  • The use of energy-efficient hardware, such as GPUs and TPUs, can significantly cut energy consumption. Compared to typical CPUs, these more advanced units can reduce energy consumption by up to 50%, thus lessening the environmental impact associated with AI development.

  • 5-2. Leveraging Renewable Energy and Carbon Offsetting Initiatives

  • Strategies to mitigate the environmental impact of AI include the use of renewable energy sources and carbon offsetting initiatives. Using renewable energy can reduce the carbon footprint of large language models by up to 98%. Carbon offsetting initiatives aim to compensate for greenhouse gas emissions related to the development and use of these models, helping to lower emissions even though they do not directly decrease environmental impact.

  • 5-3. Effective Model Design and Training Approaches

  • There are effective approaches to model design and training that can reduce the computing power needed, thus minimizing environmental impact. Techniques such as distillation, quantization, and model pruning can decrease the required computing power by as much as 90%, substantially lowering the negative environmental effects.

  • 5-4. The Role of Pre-trained Models in Energy Conservation

  • Reusing and sharing pre-trained models can significantly save energy. By employing pre-trained models for multiple tasks, up to 80% of the total energy used in training can be saved. This strategy not only reduces the environmental impact but also promotes collaboration within the AI community.

6. Glossary

  • 6-1. Generative AI [Technology]

  • Generative AI is a class of artificial intelligence that focuses on creating new content, such as text, images, or sound. It requires significant computational power, contributing to increased energy consumption and related environmental concerns.

  • 6-2. Data Center [Infrastructure]

  • Data centers are facilities used to house computer systems and associated components, such as storage systems. With the rise of generative AI, these centers are under increased pressure to meet growing power and cooling needs, affecting their operational and environmental efficiency.

  • 6-3. Cooling Systems [Technical term]

  • Cooling systems are essential in maintaining optimal temperatures within data centers to prevent hardware damage and ensure performance. Advanced techniques such as liquid cooling are necessary to handle the heat generated by high-performance computing required for AI workloads.

  • 6-4. Carbon Footprint [Environmental impact]

  • The carbon footprint refers to the total greenhouse gas emissions caused by an entity or activity. The energy-intensive processes required for training and maintaining generative AI models significantly contribute to the carbon emissions of data centers.

7. Conclusion

  • The development of generative AI has dramatically increased the demand for energy, posing significant challenges to data center infrastructure and creating environmental concerns. Immediate measures and strategies must be adopted to balance the technological advancement with sustainable practices to mitigate the ecological footprint.