• The AI Pulse
  • Posts
  • šŸ§  Is AIā€™s Computing Demand Sustainable?

šŸ§  Is AIā€™s Computing Demand Sustainable?

PLUS: Are Your Prompts Using Too Much Energy?

Welcome back AI prodigies!

In todayā€™s Sunday Special:

  • šŸ“œThe Prelude

  • āš™ļøHow GenAI Uses Compute

  • šŸ“¦Demand vs. Supply

  • šŸ¤”Whatā€™s the Verdict

  • šŸ”‘Key Takeaway

Read Time: 7 minutes

šŸŽ“Key Terms

  • Generative AI (GenAI): When AI models create new content such as text, images, audio, video, or code.

  • Parameters: Internal settings like weights, biases, and numbers that help AI models understand and predict words.

  • Graphics Processing Units (GPUs): Specialized computer chips capable of performing mathematical calculations simultaneously.

  • Floating Point Operations per Second (FLOPs): How many operations (i.e., addition, subtraction, multiplication, or division) a computer can solve within a second.

šŸ©ŗ PULSE CHECK

Are you concerned about AIā€™s energy use?

Vote Below to View Live Results

Login or Subscribe to participate in polls.

šŸ“œTHE PRELUDE

For centuries, Human Labor fueled economies. In recent decades, Knowledge Workers have driven progress. Now, AI is poised to do the same.

Historically, weā€™ve faced human constraints; Production Workers could only assemble so many goods in a factory, and Financial Analysts could only analyze so much data in a spreadsheet.

AI faces a different constraint: Compute, which is how quickly and efficiently computers can perform calculations and process data.

This year, Google will spend more money on Compute than employee salaries. As one of the worldā€™s most valuable Big Tech companies, Google symbolizes a profound shift in how humans create economic valueā€”from Labor and Intelligence to AI.

Today, GenAI consumes the most Compute. So, how exactly does GenAI rely on Compute? Why does a lack of Compute hinder GenAI progress?

āš™ļøHOW GENAI USES COMPUTE

The size of the largest AI models has grown 2.8x every year since 2018. In 2024, OpenAIā€™s GPT-4 leveraged 1.8 trillion Parameters. By 2028, the largest AI models will likely reach 15 trillion Parameters, which is more than 7x Compute of GPT-4.

But the size of AI models isnā€™t the only factor influencing the demand for Compute. Here are the top three use cases of Compute and roughly how much each one uses:

  1. Training (i.e., 70%): This process involves feeding an AI model large amounts of high-quality datasets to help it perceive, interpret, and learn from the data. It requires tons of hardware, time, and mathematical calculations. For example, GPT-4 was trained on around 25,000 NVIDIA A100 Tensor Core GPUs over nearly 100 days, performing 2.15 Ɨ 1025 FLOPs, which is over 100x more mathematical calculations than the number of stars in the observable universe.

  2. Fine-Tuning (i.e., 15%): This process involves taking a pre-trained AI model and adjusting it to fit your specific use case better. For example, Reinforcement Learning From Human Feedback (RLHF) relies on human feedback to teach AI models to align with human preferences. Every time ChatGPT responds to your question, you can click the Thumbs Up button below the response to support OpenAIā€™s RLHF efforts.

  3. Inference (i.e., 15%): This process refers to everything that happens after you enter your prompt. For example, Reasoning Engines like OpenAI o3 use Chain-of-Thought (CoT) to break down complex problems into manageable steps.

šŸ“¦DEMAND VS. SUPPLY

The demand for Compute is enormous and ever-increasing, with no end in sight. The BCG Henderson Institute (BHI) outlined an aggressive scenario for GenAIā€™s impact on Compute through 2028.

Compute Demand?

When ChatGPT first launched in November 2022, it reached 100 million active users in just two months. Now, ChatGPT has passed 400 million weekly active users, and GenAI adoption across Big Tech companies has skyrocketed.

Fortunately, most businesses using GenAI donā€™t engage in computationally intensive Training. Instead, they use existing GenAI models for Inference. Since Inference doesnā€™t require the same cutting-edge GPUs as Training, it can be distributed across less specialized hardware. Despite this, Inference workloads are expected to eclipse Training workloads by 2028 because of the sheer scale of user adoption.

Meta Ads reaches 3.3 billion users per day. This reach represents 20% of U.S. digital ad spending. By 2028, Meta may use GenAI to generate 30 personalized ads per user. Using GenAI on Meta Ads alone would represent a major increase in Inference workloads.

Compute Supply?

Faced with growing demand, will supply run out? The answer lies in the expansion of global Compute capacity. In 2023, NVIDIA shipped approximately 3.8 million cutting-edge GPUs to data centers. About 650,000 of them were NVIDIA H100 Tensor Core GPUs. Assuming an Average Utilization Rate (AUR) of 50%, BHI estimates the 2023 global AI Compute supply will reach 7 Ɨ 10^28 FLOPs by 2028, which is equivalent to counting every single grain of sand on every beach in the world.

SemiAnalysis projects that Compute will increase 60x by the end of 2025. If these projections hold, global Compute capacity should reach approximately 4 x 10^30 FLOPs by 2028, which is more than enough to meet even the most aggressive GenAI adoption scenarios.

šŸ¤”WHATā€™S THE VERDICT

The worldā€™s aging power grid is ill-equipped to handle the sudden surge in energy consumption from GenAI. For context, new data centers demanded 195 terawatt-hours (TWh) of electricity last year. Thatā€™s as much electricity as 18 million households per year. By 2027, new data centers could require 500 TWh, or 46 million households, worth of electricity. If a Compute shortage occurs, energy is more likely the culprit than hardware.

šŸ”‘KEY TAKEAWAY

AIā€™s largest obstacle to widespread adoption is arguably Compute capacity. While Training has historically required the most Compute, demand will increasingly come from Inference. Fortunately, Compute capacity will likely remain sufficient, so AI adopters can breathe a sigh of relief. But the worldā€™s aging power grid could cause some issues.

šŸ“’FINAL NOTE

FEEDBACK

How would you rate todayā€™s email?

It helps us improve the content for you!

Login or Subscribe to participate in polls.

ā¤ļøTAIP Review of The Week

ā€œCan you guys cover computing demand? Great stuff btw!ā€

-Keaton (1ļøāƒ£ šŸ‘Nailed it!)
REFER & EARN

šŸŽ‰Your Friends Learn, You Earn!

You currently have 0 referrals, only 1 away from receiving āš™ļøUltimate Prompt Engineering Guide.

Reply

or to participate.