International Energy Agency Archives - gettectonic.com
Generative AI Energy Consumption Rises

Generative AI Energy Consumption Rises

Generative AI Energy Consumption Rises, but Impact on ROI Unclear The energy costs associated with generative AI (GenAI) are often overlooked in enterprise financial planning. However, industry experts suggest that IT leaders should account for the power consumption that comes with adopting this technology. When building a business case for generative AI, some costs are evident, like large language model (LLM) fees and SaaS subscriptions. Other costs, such as preparing data, upgrading cloud infrastructure, and managing organizational changes, are less visible but significant. Generative AI Energy Consumption Rises One often overlooked cost is the energy consumption of generative AI. Training LLMs and responding to user requests—whether answering questions or generating images—demands considerable computing power. These tasks generate heat and necessitate sophisticated cooling systems in data centers, which, in turn, consume additional energy. Despite this, most enterprises have not focused on the energy requirements of GenAI. However, the issue is gaining more attention at a broader level. The International Energy Agency (IEA), for instance, has forecasted that electricity consumption from data centers, AI, and cryptocurrency could double by 2026. By that time, data centers’ electricity use could exceed 1,000 terawatt-hours, equivalent to Japan’s total electricity consumption. Goldman Sachs also flagged the growing energy demand, attributing it partly to AI. The firm projects that global data center electricity use could more than double by 2030, fueled by AI and other factors. ROI Implications of Energy Costs The extent to which rising energy consumption will affect GenAI’s return on investment (ROI) remains unclear. For now, the perceived benefits of GenAI seem to outweigh concerns about energy costs. Most businesses have not been directly impacted, as these costs tend to affect hyperscalers more. For instance, Google reported a 13% increase in greenhouse gas emissions in 2023, largely due to AI-related energy demands in its data centers. Scott Likens, PwC’s global chief AI engineering officer, noted that while energy consumption isn’t a barrier to adoption, it should still be factored into long-term strategies. “You don’t take it for granted. There’s a cost somewhere for the enterprise,” he said. Energy Costs: Hidden but Present Although energy expenses may not appear on an enterprise’s invoice, they are still present. Generative AI’s energy consumption is tied to both model training and inference—each time a user makes a query, the system expends energy to generate a response. While the energy used for individual queries is minor, the cumulative effect across millions of users can add up. How these costs are passed to customers is somewhat opaque. Licensing fees for enterprise versions of GenAI products likely include energy costs, spread across the user base. According to PwC’s Likens, the costs associated with training models are shared among many users, reducing the burden on individual enterprises. On the inference side, GenAI vendors charge for tokens, which correspond to computational power. Although increased token usage signals higher energy consumption, the financial impact on enterprises has so far been minimal, especially as token costs have decreased. This may be similar to buying an EV to save on gas but spending hundreds and losing hours at charging stations. Energy as an Indirect Concern While energy costs haven’t been top-of-mind for GenAI adopters, they could indirectly address the issue by focusing on other deployment challenges, such as reducing latency and improving cost efficiency. Newer models, such as OpenAI’s GPT-4o mini, are more economical and have helped organizations scale GenAI without prohibitive costs. Organizations may also use smaller, fine-tuned models to decrease latency and energy consumption. By adopting multimodel approaches, enterprises can choose models based on the complexity of a task, optimizing for both speed and energy efficiency. The Data Center Dilemma As enterprises consider GenAI’s energy demands, data centers face the challenge head-on, investing in more sophisticated cooling systems to handle the heat generated by AI workloads. According to the Dell’Oro Group, the data center physical infrastructure market grew in the second quarter of 2024, signaling the start of the “AI growth cycle” for infrastructure sales, particularly thermal management systems. Liquid cooling, more efficient than air cooling, is gaining traction as a way to manage the heat from high-performance computing. This method is expected to see rapid growth in the coming years as demand for AI workloads continues to increase. Nuclear Power and AI Energy Demands To meet AI’s growing energy demands, some hyperscalers are exploring nuclear energy for their data centers. AWS, Google, and Microsoft are among the companies exploring this option, with AWS acquiring a nuclear-powered data center campus earlier this year. Nuclear power could help these tech giants keep pace with AI’s energy requirements while also meeting sustainability goals. I don’t know. It seems like if you akin AI accessibility to more nuclear power plants you would lose a lot of fans. As GenAI continues to evolve, both energy costs and efficiency are likely to play a greater role in decision-making. PwC has already begun including carbon impact as part of its GenAI value framework, which assesses the full scope of generative AI deployments. “The cost of carbon is in there, so we shouldn’t ignore it,” Likens said. Generative AI Energy Consumption Rises Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Generative ai energy consumption

Growing Energy Consumption in Generative AI

Growing Energy Consumption in Generative AI, but ROI Impact Remains Unclear The rising energy costs associated with generative AI aren’t always central in enterprise financial considerations, yet experts suggest IT leaders should take note. Building a business case for generative AI involves both obvious and hidden expenses. Licensing fees for large language models (LLMs) and SaaS subscriptions are visible expenses, but less apparent costs include data preparation, cloud infrastructure upgrades, and managing organizational change. Growing Energy Consumption in Generative AI. One under-the-radar cost is the energy required by generative AI. Training LLMs demands vast computing power, and even routine AI tasks like answering user queries or generating images consume energy. These intensive processes require robust cooling systems in data centers, adding to energy use. While energy costs haven’t been a focus for GenAI adopters, growing awareness has prompted the International Energy Agency (IEA) to predict a doubling of data center electricity consumption by 2026, attributing much of the increase to AI. Goldman Sachs echoed these concerns, projecting data center power consumption to more than double by 2030. For now, generative AI’s anticipated benefits outweigh energy cost concerns for most enterprises, with hyperscalers like Google bearing the brunt of these costs. Google recently reported a 13% increase in greenhouse gas emissions, citing AI as a major contributor and suggesting that reducing emissions might become more challenging with AI’s continued growth. Growing Energy Consumption in Generative AI While not a barrier to adoption, energy costs play into generative AI’s long-term viability, noted Scott Likens, global AI engineering leader at PwC, emphasizing that “there’s energy being used — you don’t take it for granted.” Energy Costs and Enterprise Adoption Generative AI users might not see a line item for energy costs, yet these are embedded in fees. Ryan Gross of Caylent points out that the costs are mainly tied to model training and inferencing, with each model query, though individually minor, adding up over time. These expenses are often spread across the customer base, as companies pay for generative AI access through a licensing model. A PwC sustainability study showed that GenAI power costs, particularly from model training, are distributed among licensees. Token-based pricing for LLM usage also reflects inferencing costs, though these charges have decreased. Likens noted that the largest expenses still come from infrastructure and data management rather than energy. Potential Efficiency Gains Though energy isn’t a primary consideration, enterprises could reduce consumption indirectly through technological advancements. Newer, more cost-efficient models like OpenAI’s GPT-4o mini are 60% less expensive per token than prior versions, enabling organizations to deploy GenAI on a larger scale while keeping costs lower. Small, fine-tuned models can be used to address latency and lower energy consumption, part of a “multimodel” approach that can provide different accuracy and latency levels with varying energy demands. Agentic AI also offers opportunities for cost and energy savings. By breaking down tasks and routing them through specialized models, companies can minimize latency and reduce power usage. According to Likens, using agentic architecture could cut costs and consumption, particularly when tasks are routed to more efficient models. Rising Data Center Energy Needs While enterprises may feel shielded from direct energy costs, data centers bear the growing power demand. Cooling solutions are evolving, with liquid cooling systems becoming more prevalent for AI workloads. As data centers face the “AI growth cycle,” the demand for energy-efficient cooling solutions has fueled a resurgence in thermal management investment. Liquid cooling, being more efficient than air cooling, is gaining traction due to the power demands of AI and high-performance computing. IDTechEx projects that data center liquid cooling revenue could exceed $50 billion by 2035. Meanwhile, data centers are exploring nuclear power, with AWS, Google, and Microsoft among those considering nuclear energy as a sustainable solution to meet AI’s power demands. Future ROI Considerations While enterprises remain shielded from the full energy costs of generative AI, careful model selection and architectural choices could help curb consumption. PwC, for instance, factors in the “carbon impact” as part of its GenAI deployment strategy, recognizing that energy considerations are now a part of the generative AI value proposition. As organizations increasingly factor sustainability into their tech decisions, energy efficiency might soon play a larger role in generative AI ROI calculations. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Where Will AI Take Us?

Where Will AI Take Us?

Author Jeremy Wagstaff wrote a very thought provoking article on the future of AI, and how much of it we could predict based on the past. This insight expands on that article. Artificial Intelligence (AI) refers to the simulation of human intelligence in machines that are programmed to think and learn. These machines can perform tasks that typically require human intelligence, such as visual perception, speech recognition, decision-making, and language translation. Many people think of artificial intelligence in the vein of how they personally use it. Some people don’t even realize when they are using it. Artificial intelligence has long been a concept in human mythology and literature. Our imaginations have been grabbed by the thought of sentient machines constructed by humans, from Talos, the enormous bronze automaton (self-operating machine) that safeguarded the island of Crete in Greek mythology, to the spacecraft-controlling HAL in 2001: A Space Odyssey. Artificial Intelligence comes in a variety of flavors, if you will. Artificial intelligence can be categorized in several ways, including by capability and functionality: You likely weren’t even aware of all of the above categorizations of artificial intelligence. Most of us still would sub set into generative ai, a subset of narrow AI, predictive ai, and reactive ai. Reflect on the AI journey through the Three C’s – Computation, Cognition, and Communication – as the guiding pillars for understanding the transformative potential of AI. Gain insights into how these concepts converge to shape the future of technology. Beyond a definition, what really is artificial intelligence, who makes it, who uses it, what does it do and how. Artificial Intelligence Companies – A Sampling AI and Its Challenges Artificial intelligence (AI) presents a novel and significant challenge to the fundamental ideas underpinning the modern state, affecting governance, social and mental health, the balance between capitalism and individual protection, and international cooperation and commerce. Addressing this amorphous technology, which lacks a clear definition yet pervades increasing facets of life, is complex and daunting. It is essential to recognize what should not be done, drawing lessons from past mistakes that may not be reversible this time. In the 1920s, the concept of a street was fluid. People viewed city streets as public spaces open to anyone not endangering or obstructing others. However, conflicts between ‘joy riders’ and ‘jay walkers’ began to emerge, with judges often siding with pedestrians in lawsuits. Motorist associations and the car industry lobbied to prioritize vehicles, leading to the construction of vehicle-only thoroughfares. The dominance of cars prevailed for a century, but recent efforts have sought to reverse this trend with ‘complete streets,’ bicycle and pedestrian infrastructure, and traffic calming measures. Technology, such as electric micro-mobility and improved VR/AR for street design, plays a role in this transformation. The guy digging out a road bed for chariots and Roman armies likely considered none of this. Addressing new technology is not easy to do, and it’s taken changes to our planet’s climate, a pandemic, and the deaths of tens of millions of people in traffic accidents (3.6 million in the U.S. since 1899). If we had better understood the implications of the first automobile technology, perhaps we could have made better decisions. Similarly, society should avoid repeating past mistakes with AI. The market has driven AI’s development, often prioritizing those who stand to profit over consumers. You know, capitalism. The rapid adoption and expansion of AI, driven by commercial and nationalist competition, have created significant distortions. Companies like Nvidia have soared in value due to AI chip sales, and governments are heavily investing in AI technology to gain competitive advantages. Listening to AI experts highlights the enormity of the commitment being made and reveals that these experts, despite their knowledge, may not be the best sources for AI guidance. The size and impact of AI are already redirecting massive resources and creating new challenges. For example, AI’s demand for energy, chips, memory, and talent is immense, and the future of AI-driven applications depends on the availability of computing resources. The rise in demand for AI has already led to significant industry changes. Data centers are transforming into ‘AI data centers,’ and the demand for specialized AI chips and memory is skyrocketing. The U.S. government is investing billions to boost its position in AI, and countries like China are rapidly advancing in AI expertise. China may be behind in physical assets, but it is moving fast on expertise, generating almost half of the world’s top AI researchers (Source: New York Times). The U.S. has just announced it will provide chip maker Intel with $20 billion in grants and loans to boost the country’s position in AI. Nvidia is now the third largest company in the world, entirely because its specialized chips account for more than 70 percent of AI chip sales. Memory-maker Micro has mostly run out of high-bandwidth memory (HBM) stocks because of the chips’ usage in AI—one customer paid $600 million up-front to lock in supply, according to a story by Stack. Back in January, the International Energy Agency forecast that data centers may more than double their electrical consumption by 2026 (Source: Sandra MacGregor, Data Center Knowledge). AI is sucking up all the payroll: Those tech workers who don’t have AI skills are finding fewer roles and lower salaries—or their jobs disappearing entirely to automation and AI (Source: Belle Lin at WSJ). Sam Altman of OpenAI sees a future where demand for AI-driven apps is limited only by the amount of computing available at a price the consumer is willing o pay. “Compute is going to be the currency of the future. I think it will be maybe the most precious commodity in the world, and I think we should be investing heavily to make a lot more compute.” Sam Altman, OpenAI CEO This AI buildup is reminiscent of past technological transformations, where powerful interests shaped outcomes, often at the expense of broader societal considerations. Consider early car manufacturers. They focused on a need for factories, components, and roads.

Read More
gettectonic.com