As of February 9, 2024, recent reports from the Wall Street Journal have unveiled OpenAI CEO Sam Altman’s ambitious plan to secure up to $7 trillion for a revolutionary project aimed at enhancing the global chip capacity. This groundbreaking initiative will be funded by investors, including the U.A.E., with the goal of significantly expanding the capability to power AI models.
While Altman’s vision may appear lofty or reminiscent of the hype generated by figures like Elon Musk, environmental expert Sasha Luccioni warns about the potential ecological ramifications of such a massive undertaking. According to Luccioni, even if the energy sources used are renewable, the sheer quantity of water and rare earth minerals required for this project is staggering.
For context, a Fortune article from September 2023 highlighted how AI tools contributed to a 34% surge in Microsoft’s water consumption. Similarly, Meta’s Llama 2 model reportedly consumed double the amount of water compared to its predecessor, Llama 1. Additionally, a study conducted in 2023 revealed that training OpenAI’s GPT-3 model consumed a whopping 700,000 liters of water.
The Implications of AI Development
Beyond the concerning environmental impact, the scarcity of rare earth minerals like gallium and germanium has exacerbated tensions in the global chip industry, particularly in relation to China. Luccioni expressed her frustration with Altman’s approach, criticizing the lack of emphasis on more efficient AI methodologies in favor of a brute force strategy that some applaud as visionary.
The Race for GPU Access in Silicon Valley
The pursuit of addressing the current GPU shortages and reshaping the semiconductor landscape, as demonstrated by Altman, is not an isolated initiative. Silicon Valley has been abuzz with discussions surrounding access to Nvidia’s coveted H100 GPU for large language model training, which has become a coveted asset.
In a recent earnings call, Meta’s CEO Mark Zuckerberg underscored the importance of world-class compute infrastructure in achieving AI advancements, emphasizing the company’s investment in designing custom silicon tailored to their specific workloads. Concurrently, Meta is on track to deploy approximately 350,000 H100 GPUs by the end of the year, with a total of around 600,000 H100 equivalents for computational purposes.
Meanwhile, criticisms have been raised regarding Nvidia’s transparency concerning the environmental impact of their product manufacturing processes, which are executed by the Taiwan Semiconductor Manufacturing Company. Luccioni has called for greater accountability in disclosing the carbon footprint associated with Nvidia’s products.
For further insights into the implications of Sam Altman’s innovative AI chip project and the substantial natural resources it entails, feel free to explore the full article on VentureBeat.

