Skip to main content
Posts tagged:

energy

Power and AI

The Growing Power Needs for Large Language Models

In 2024, AI is awesome, empowering and available to everyone. Unfortunately, while AI is free to consumers, these models are expensive to train and operate at scale. Training them is expected to be the most expensive thing ever. Yes, more than the Manhattan project, pyramids and the entire GDP of the world economy. No wonder companies with free compute are dominating this space.

By 2025, the cost to train an LLM surpasses the Apollo Project, a historical benchmark for significant expenditure. This projection emphasizes the increasing financial burden and resource demand associated with advancing AI capabilities, underscoring the need for more efficient and sustainable approaches in AI research and development. The data points to a future where the financial and energy requirements for AI could become unsustainable without significant technological breakthroughs or shifts in strategy.

Why?

Because of how deep learning works and how it’s trained. The first era, marked by steady progress, follows a trend aligned with Moore’s Law, where computing power doubled approximately every two years. Notable milestones during this period include the development of early AI models like the Perceptron and later advancements such as NETtalk and TD-Gammon.

The second era, beginning around 2012 with the advent of deep learning, demonstrates a dramatic increase in compute usage, following a much steeper trajectory where computational power doubles approximately every 3.4 months. This surge is driven by the development of more complex models like AlexNet, ResNets, and AlphaGoZero. Key factors behind this acceleration include the availability of massive datasets, advancements in GPU and specialized hardware, and significant investments in AI research. As AI models have become more sophisticated, the demand for computational resources has skyrocketed, leading to innovations and increased emphasis on sustainable and efficient energy sources to support this growth.

Training LLMs involves massive computational resources. For instance, models like GPT-3, with 175 billion parameters, require extensive parallel processing using GPUs. Training such a model on a single Nvidia V100 GPU would take an estimated 288 years, emphasizing the need for large-scale distributed computing setups to make the process feasible in a reasonable timeframe. This leads to higher costs, both financially and in terms of energy consumption.

Recent studies have highlighted the dramatic increase in computational power needed for AI training, which is rising at an unprecedented rate. Over the past seven years, compute usage has increased by 300,000-fold, underscoring the escalating costs associated with these advancements. This increase not only affects financial expenditures but also contributes to higher carbon emissions, posing environmental concerns.

Infrastructure and Efficiency Improvements

To address these challenges, companies like Cerebras and Cirrascale are developing specialized infrastructure solutions. For example, Cerebras’ AI Model Studio offers a rental model that leverages clusters of CS-2 nodes, providing a scalable and cost-effective alternative to traditional cloud-based solutions. This approach aims to deliver predictable pricing and reduce the costs associated with training large models.

Moreover, researchers are exploring various optimization techniques to improve the efficiency of LLMs. These include model approximation, compression strategies, and innovations in hardware architecture. For instance, advancements in GPU interconnects and supercomputing technologies are critical to overcoming bottlenecks related to data transfer speeds between servers, which remain a significant challenge.

Implications for Commodities and Nuclear Power

The increasing power needs for AI training have broader implications for commodities, particularly in the energy sector. As AI models grow, the demand for electricity to power the required computational infrastructure will likely rise. This could drive up the prices of energy commodities, especially in regions where data centers are concentrated. Additionally, the need for advanced hardware, such as GPUs and specialized processors, will impact the supply chains and pricing of these components.

To address the substantial energy needs of AI, particularly in powering the growing number of data centers, various approaches are being considered. One notable strategy involves leveraging nuclear power. This approach is championed by tech leaders like OpenAI CEO Sam Altman, who views AI and affordable, green energy as intertwined essentials for a future of abundance. Nuclear startups, such as Oklo, which Altman supports, are working on advanced nuclear reactors designed to be safer, more efficient, and smaller than traditional plants. Oklo’s projects include a 15-megawatt fission reactor and a grant-supported initiative to recycle nuclear waste into new fuel.

However, integrating nuclear energy into the tech sector faces significant regulatory challenges. The Nuclear Regulatory Commission (NRC) denied Oklo’s application for its Idaho plant design due to insufficient safety information, and the Air Force rescinded a contract for a microreactor pilot program in Alaska. These hurdles highlight the tension between the rapid development pace of AI technologies and the methodical, decades-long process traditionally required for nuclear energy projects .

The demand for sustainable energy solutions is underscored by the rising energy consumption of AI servers, which could soon exceed the annual energy use of some small nations. Major tech firms like Microsoft, Google, and Amazon are investing heavily in nuclear energy to secure stable, clean power for their operations. Microsoft has agreements to buy nuclear-generated electricity for its data centers, while Google and Amazon have invested in fusion startups .

By 0 Comments

Home Energy Usage

I decided to look at my home energy usage. Below are notes on how I’m approaching it. I wish I could get insight from the MIOS graphing plugin, but it is pretty basic and can’t throw out outliers and has terrible zoom capability. I also never figured out how to set the y-axis.

In order to build a plot, I considered the following options:

  • Use a custom javascript library that has zoom capability (i.e. re-purpose a stock chart)
  • Use MATLAB

I had to use MATLAB (or similar) to condition the data in any case. First, I had to scp the data over from my Vera. For example:

scp remote@micasa:/dataMine/database/10/raw/2302.txt tmp/10_2.txt

Which I was able to pull together using some MATLAB:

Since I monitor both legs of my electrical setup, I would have two plots. Here is what my MATLAB code currently produces (click on the plot below to do some analysis):

Booher Home Energy Use From 9 to 14 Feb

Booher Home Energy Use From 9 to 14 Feb

I’m not happy with this plot. It is hard to get at the events that are happening. I can’t easily put vertical bands on because the units are all strange with the time series plot. I definitely can’t browse the data and get insight for what is happening. I also want to add both datasets together so I can see total energy consumption.

Additional data from Dominion Power

Also, as part of my analysis. I pulled together some historical data from Dominion.

Energy Usage from Dominion Power

Energy Usage from Dominion Power

Actual Data

Meter Read Date Days Usage Daily Usage
02/04/2014 29 1240 43
01/06/2014 33 1060 32
12/04/2013 34 1271 37
10/31/2013 29 1144 39
10/02/2013 29 1181 41
09/03/2013 32 1523 48
08/02/2013 28 1342 48
07/05/2013 30 1608 54
06/05/2013 34 1318 39
05/02/2013 28 985 35
04/04/2013 30 1061 35
03/05/2013 29 1161 40
02/04/2013 31 1139 37

Table 2 — From Dominion Power

Meter Read Date Days Meter Reading Method Meter Read Usage (kWh) Demand Avg. Daily Usage
01/06/2014 33     6929     1060 0.0       32
12/04/2013 34     5869     1271 0.0       37
10/31/2013 29     4598     1144 0.0       39
10/02/2013 29     3454     1181 0.0       41
09/03/2013 32     2273     1523 0.0       48
08/02/2013 28      750     1342 0.0       48
07/05/2013 30 AMR – MOBILE READ BY VAN  19247     1608 0.0       54
06/05/2013 34 AMR – MOBILE READ BY VAN  17639     1318 0.0       39
05/02/2013 28 AMR – MOBILE READ BY VAN  16321      985 0.0       35
04/04/2013 30 AMR – MOBILE READ BY VAN  15336     1061 0.0       35
03/05/2013 29 AMR – MOBILE READ BY VAN  14275     1161 0.0       40
02/04/2013 31 AMR – MOBILE READ BY VAN  13114     1139 0.0       37
01/04/2013 32 AMR – MOBILE READ BY VAN  11975     1341 0.0       42
Totals
     
16,134
 

Some links:

I’m out of time this morning, but when I get more time, I’m going to be considering the following:

By 0 Comments