|
1 | | -# Commercial Models – Green AI Analysis |
| 1 | +# Commercial Models: The Crisis of AI Environmental Visibility |
| 2 | + |
| 3 | +The environmental impact of Large Language Models (LLMs) has transitioned from |
| 4 | +an esoteric concern to a critical research priority. While early studies focused |
| 5 | +primarily on the energy cost of the massive **training phase** (e.g., GPT-4 |
| 6 | +training estimates: $\approx 51.8 \text{ – } 62.3 \text{ GWh}$) [Source: Sustainability |
| 7 | +by Numbers], recent, rigorous analyses—such as the landmark benchmarking paper |
| 8 | +*"How Hungry is AI? Benchmarking Energy, Water, and Carbon Footprint of LLM |
| 9 | +Inference"* (Jegham et al., 2025)—have confirmed two major points: |
| 10 | + |
| 11 | +* **Inference Dominance:** The energy consumed during **inference** |
| 12 | +* (running the model |
| 13 | +for every user query) is increasingly the dominant contributor to an LLM's total |
| 14 | +lifecycle environmental footprint due to global scale [Source: Google Cloud Blog]. |
| 15 | +* **Divergent Reporting:** LLM energy reporting is highly unstable. As |
| 16 | +demonstrated by the benchmarking paper, resource consumption can vary by over |
| 17 | +**70 times** between the most and least efficient models for an equivalent task. |
| 18 | +Furthermore, consumption figures depend not just on model size, but profoundly |
| 19 | +on **deployment infrastructure** (PUE, regional carbon intensity) [Source: AWS |
| 20 | +Sustainability Report] and **methodology** (batch size, token length, hardware |
| 21 | +generation) [Source: Optimal Sparsity of MoE, arXiv]. This divergence leads to |
| 22 | +figures where one study reports less consumption (e.g., Epoch AI's $0.3 \text{ Wh}$ |
| 23 | +per query) [Source: Epoch AI], while another (using different infrastructure |
| 24 | +assumptions) reports significantly more. |
2 | 25 |
|
3 | | -This folder contains the **comparative sustainability analysis** of several |
4 | | -large language models (LLMs) used in commercial applications. |
5 | | -It is part of the **ELO2 Green AI project**, focusing on estimating |
6 | | -the **energy, carbon, and water footprints** of each model. |
7 | | - |
8 | | ---- |
9 | | - |
10 | | -## 📄 Contents |
11 | | - |
12 | | -- **models.md** – Main document providing technical summaries and |
13 | | - sustainability estimates for: |
14 | | - - GPT-4 (OpenAI) |
15 | | - - Claude 3 Haiku (Anthropic) |
16 | | - - Gemini Nano (Google) |
17 | | - |
18 | 26 | --- |
19 | 27 |
|
20 | | -## 🎯 Purpose |
21 | | - |
22 | | -This documentation: |
23 | | - |
24 | | -- Highlights how **different LLM architectures and deployments** |
25 | | - affect energy and water use. |
26 | | -- Demonstrates how **model size and hosting** influence environmental impact. |
27 | | -- Supports ongoing evaluation of **Green AI strategies** for efficient computing. |
28 | | - |
29 | | ---- |
| 28 | +## Our Research Intervention |
| 29 | + |
| 30 | +This critical lack of standardized, reliable environmental data creates an |
| 31 | +**eco-efficiency paradox**: we cannot definitively prove which models are |
| 32 | +sustainable without standardized measurements, but the required performance |
| 33 | +trade-offs often push users toward the most power-hungry frontier models (e.g., |
| 34 | +GPT-4's high capability, estimated at $0.3 \text{ – } 1.8 \text{ Wh}$ per query) |
| 35 | +[Source: The Verge]. |
| 36 | + |
| 37 | +This project addresses this paradox with an **interventionist approach**. |
| 38 | +Rather than |
| 39 | +simply measuring existing systems, our research question directly investigates |
| 40 | +whether we can *engineer* open-source models to definitively close the gap in the |
| 41 | +efficiency balance: |
| 42 | + |
| 43 | +> **Research Question:** **Can open-source language models, enhanced with |
| 44 | +optimization techniques such as recursive editing and distillation, become |
| 45 | +environmentally and functionally viable alternatives to commercial models?** |
| 46 | + |
| 47 | +We hypothesize that by applying resource-saving techniques (e.g., **distillation** |
| 48 | +to compress model size and **recursive editing** to refine output with minimal |
| 49 | +re-computation), we can achieve an **eco-efficiency score** that consistently |
| 50 | +demonstrates their viability as a sustainable alternative, driving adoption of |
| 51 | +open-source models over proprietary, resource-intensive solutions. |
0 commit comments