Skip to content

Commit de33198

Browse files
authored
Merge pull request #14 from MIT-Emerging-Talent/Commercial-models
Milestone 2: Commercial Models Research β€” Energy, Carbon, and Accuracy Analysis #8
2 parents d2ae928 + 8f3a14f commit de33198

2 files changed

Lines changed: 226 additions & 0 deletions

File tree

β€Žcommercial_models/README.mdβ€Ž

Lines changed: 29 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,29 @@
1+
# Commercial Models – Green AI Analysis
2+
3+
This folder contains the **comparative sustainability analysis** of several
4+
large language models (LLMs) used in commercial applications.
5+
It is part of the **ELO2 Green AI project**, focusing on estimating
6+
the **energy, carbon, and water footprints** of each model.
7+
8+
---
9+
10+
## πŸ“„ Contents
11+
12+
- **models.md** – Main document providing technical summaries and
13+
sustainability estimates for:
14+
- GPT-4 (OpenAI)
15+
- Claude 3 Haiku (Anthropic)
16+
- Gemini Nano (Google)
17+
18+
---
19+
20+
## 🎯 Purpose
21+
22+
This documentation:
23+
24+
- Highlights how **different LLM architectures and deployments**
25+
affect energy and water use.
26+
- Demonstrates how **model size and hosting** influence environmental impact.
27+
- Supports ongoing evaluation of **Green AI strategies** for efficient computing.
28+
29+
---

β€Žcommercial_models/models.mdβ€Ž

Lines changed: 197 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,197 @@
1+
<!-- markdownlint-disable MD013 MD024 MD025 MD026 MD041 MD001 -->
2+
<!-- This disables line length, duplicate headings, multiple top-level headings, heading punctuation, first heading, and heading increment rules -->
3+
<!-- markdownlint-disable MD013 -->
4+
<!-- markdownlint-disable MD025 -->
5+
<!-- Multiple top-level headings needed for each model section -->
6+
7+
# Comparative Environmental and Technical Overview of Modern AI Models
8+
9+
This document presents verified technical and environmental data for three leading AI models β€” **OpenAI GPT-4**, **Anthropic Claude 3 Haiku**, and **Google Gemini Nano** β€” focusing on energy, water, and sustainability context.
10+
11+
---
12+
13+
## Model 1: GPT-4 (OpenAI)
14+
15+
### Model Name & Provider
16+
17+
**GPT-4**, developed by **OpenAI**.
18+
19+
### Hosting / Deployment
20+
21+
Hosted via **Microsoft Azure OpenAI Service**, which operates on Azure’s global data centers (specific regions not publicly disclosed).
22+
Source: [Introducing GPT-4 in Azure OpenAI Service – Microsoft Azure Blog](https://azure.microsoft.com/en-us/blog/introducing-gpt-4-in-azure-openai-service/)
23+
24+
### Estimated Model Size / Architecture
25+
26+
GPT-4 is widely considered a **frontier model** employing a **Sparse Mixture-of-Experts (MoE)** architecture.
27+
This structure activates only a subset of parameters per inference, optimizing efficiency while maintaining scale.
28+
Estimated total parameters exceed **1 trillion**.
29+
30+
Sources:
31+
32+
- [*Optimal Sparsity of Mixture-of-Experts Language Models for Reasoning Tasks* – arXiv (2508.18672v2)](https://arxiv.org/html/2508.18672v2)
33+
- [*Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models* – arXiv (2501.12370v2)](https://arxiv.org/html/2501.12370v2)
34+
35+
### Estimated Energy (Inference)
36+
37+
Estimates vary between studies:
38+
39+
- **Epoch AI (2024):** β‰ˆ 0.3 Wh (0.0003 kWh) per query under typical load.
40+
Source: [Epoch AI – *How Much Energy Does ChatGPT Use?*](https://epoch.ai/gradient-updates/how-much-energy-does-chatgpt-use)
41+
- **Other analysts:** 0.3 – 1.8 Wh (0.0003 – 0.0018 kWh), depending on token count and hardware.
42+
43+
**Note:** OpenAI has not released official inference energy data.
44+
Values depend on hardware (GPU vs TPU), data-center PUE, and carbon intensity.
45+
46+
### Training Energy Estimates
47+
48+
Extrapolated from compute budgets and model size:
49+
β‰ˆ 51,772,500 – 62,318,750 kWh (β‰ˆ 51.8 – 62.3 GWh) for full-scale training.
50+
Source: [*The Carbon Footprint of ChatGPT* – Sustainability by Numbers](https://sustainabilitybynumbers.com/how-much-energy-does-chatgpt-use/)
51+
52+
### Water Usage
53+
54+
Water consumption derives from data-center cooling processes:
55+
56+
- A single ChatGPT query may indirectly consume β‰ˆ 0.5 L of water.
57+
- Generating a 100-word email β‰ˆ 0.14 kWh energy + 0.52 L water.
58+
Source: [The Verge – *Sam Altman on ChatGPT Energy and Water Use*](https://www.theverge.com/news/685045/sam-altman-average-chatgpt-energy-water?)
59+
60+
### PUE / CI Context Used in Studies
61+
62+
Environmental analyses generally apply:
63+
64+
- **PUE (Power Usage Effectiveness):** β‰ˆ 1.1 – 1.3 (Azure hyperscale data centers)
65+
- **CI (Carbon Intensity):** β‰ˆ 0.3 – 0.4 kg COβ‚‚e / kWh (depending on regional grid mix)
66+
67+
---
68+
69+
## Model 2: Claude 3 Haiku (Anthropic)
70+
71+
### Model Description
72+
73+
**Claude 3 Haiku** is the smallest and fastest member of Anthropic’s Claude 3 model family ( Haiku, Sonnet, Opus ), released March 2024.
74+
It is optimized for low-latency, energy-efficient applications such as chatbots, summarization, and enterprise automation.
75+
76+
Sources:
77+
78+
- [Anthropic Blog – *Introducing the Claude 3 Model Family*](https://www.anthropic.com/news/claude-3-family)
79+
- [Anthropic 3 Announcement](https://www.anthropic.com/news/claude-3-7-sonnet)
80+
81+
### Hosting / Deployment
82+
83+
Claude 3 Haiku is available through Anthropic’s API and **AWS Bedrock**.
84+
AWS data centers maintain an average PUE of β‰ˆ 1.2.
85+
Sources:
86+
87+
- [AWS Bedrock – *Use Claude on Bedrock*](https://aws.amazon.com/bedrock/claude/)
88+
- [AWS Sustainability Report 2024](https://sustainability.aboutamazon.com/reporting)
89+
90+
### Estimated Model Size / Architecture
91+
92+
Community estimates place Claude 3 Haiku at β‰ˆ 20 B parameters.
93+
The largest model in the family (Claude 3 Opus) is β‰ˆ 2 T parameters.
94+
Source: [ClaudeAI Community Discussion (Reddit)](https://www.reddit.com/r/ClaudeAI/)
95+
96+
### Estimated Energy (Inference)
97+
98+
Anthropic does not publish per-query energy figures.
99+
Based on 10–30 B parameter transformers: β‰ˆ 0.05 – 0.1 Wh (0.00005 – 0.0001 kWh) per query.
100+
Haiku is β‰ˆ 5Γ— more efficient than Claude 3 Sonnet or Opus.
101+
Sources:
102+
103+
- [Epoch AI – Machine Learning Trends (for compute/power scaling)](https://epoch.ai/trends)
104+
- [Anthropic 3 Announcement](https://www.anthropic.com/news/claude-3-7-sonnet)
105+
106+
### Training Energy Estimates
107+
108+
Claude 3 models are trained on GPU clusters (NVIDIA A100/H100) via AWS.
109+
Typical training energy for models of this scale: β‰ˆ 3,000 – 10,000 MWh.
110+
Sources:
111+
112+
- [Epoch AI – Machine Learning Trends (for compute/power scaling)](https://epoch.ai/trends)
113+
- [Anthropic Responsible Scaling Policy](https://www-cdn.anthropic.com/872c653b2d0501d6ab44cf87f43e1dc4853e4d37.pdf)
114+
115+
### Water Usage
116+
117+
Anthropic does not publish direct figures; relies on AWS cooling efficiency and water recycling policies.
118+
Sources:
119+
120+
- [AWS Water Stewardship Report](https://sustainability.aboutamazon.com/2024-amazon-sustainability-report-aws-summary.pdf)
121+
- [Anthropic Responsible Scaling Policy](https://www-cdn.anthropic.com/872c653b2d0501d6ab44cf87f43e1dc4853e4d37.pdf)
122+
123+
### PUE / CI Context Used in Studies
124+
125+
- **PUE:** β‰ˆ 1.2 (AWS average)
126+
- **CI:** β‰ˆ 0 – 0.2 kg COβ‚‚e / kWh (based on regional renewable mix)
127+
AWS targets **100 % renewable energy by 2025**.
128+
Sources:
129+
- [AWS Global Infrastructure Efficiency Data](https://sustainability.aboutamazon.com/2024-amazon-sustainability-report-aws-summary.pdf)
130+
- [Anthropic Responsible Scaling Policy](https://www-cdn.anthropic.com/872c653b2d0501d6ab44cf87f43e1dc4853e4d37.pdf)
131+
132+
---
133+
134+
## Model 3: Gemini Nano (Google DeepMind)
135+
136+
### Model Name & Provider
137+
138+
**Gemini Nano**, developed by **Google DeepMind**, is the smallest member of the Gemini family (Nano, Pro, Ultra).
139+
Sources:
140+
141+
- [Google AI Blog – *Introducing Gemini*](https://blog.google/technology/ai/google-gemini-ai/)
142+
143+
### Hosting / Deployment
144+
145+
Runs **on-device** through Android’s **AICore** system (launched in Android 14).
146+
Deployed on optimized hardware (e.g., Pixel 8 Pro, Pixel 9 Series).
147+
This local processing approach eliminates cloud compute energy and network latency.
148+
Additional coverage: - [Android Developers – *Gemini Nano Overview*](https://developer.android.com/ai/gemini-nano)
149+
150+
### Estimated Model Size / Architecture
151+
152+
Deployed in quantized versions:
153+
154+
- **Nano-1:** β‰ˆ 1.8 B parameters
155+
- **Nano-2:** β‰ˆ 3.25 B parameters
156+
Reference: [Exploding Topics – AI Model Parameters Database](https://explodingtopics.com/blog/gpt-parameters)
157+
158+
### Estimated Energy (Inference)
159+
160+
- **Median Cloud Gemini Inference:** β‰ˆ 0.24 Wh per text prompt.
161+
- **On-Device Nano Estimate:** β‰ˆ 0.01 Wh per query (benchmarks + design targets).
162+
Note: Official Nano inference measurements are not yet public.
163+
Source: [Google Cloud Blog – *Measuring the Environmental Impact of AI Inference*](https://cloud.google.com/blog/products/infrastructure/measuring-the-environmental-impact-of-ai-inference/)
164+
165+
### Training Energy Estimates
166+
167+
Gemini Nano was distilled from larger Gemini models trained on **Google TPU v5e clusters**.
168+
Training energy estimated β‰ˆ 200 – 1,200 MWh (total, amortized across billions of devices).
169+
Sources:
170+
171+
- [Google Cloud TPU Documentation](https://cloud.google.com/tpu/docs/)
172+
- [Google Cloud Blog – Environmental Impact of AI Inference](https://cloud.google.com/blog/products/infrastructure/measuring-the-environmental-impact-of-ai-inference/)
173+
174+
### Water Usage
175+
176+
- **Inference:** Zero data-center water use (on-device).
177+
- **Training:** Uses Google data centers with average WUE β‰ˆ 0.26 mL per median cloud query.
178+
Source: [Google Cloud Blog – *Measuring the Environmental Impact of AI Inference*](https://cloud.google.com/blog/products/infrastructure/measuring-the-environmental-impact-of-ai-inference/)
179+
180+
### PUE / CI Context Used in Studies
181+
182+
- **Average PUE:** 1.10 – 1.12 (Google Data Centers)
183+
- **Carbon Intensity (CI):** β‰ˆ 0.03 g COβ‚‚e / query (market-based)
184+
Reflects Google’s near-total renewable energy purchasing.
185+
Source: [Google Cloud Blog – Environmental Impact of AI Inference](https://cloud.google.com/blog/products/infrastructure/measuring-the-environmental-impact-of-ai-inference/)
186+
187+
---
188+
189+
### Summary
190+
191+
| Model | Developer | Hosting Type | Est. Parameters | Inference Energy (Wh/query) | Training Energy (MWh) | PUE | CI (kg COβ‚‚e/kWh) |
192+
|:------|:-----------|:--------------|:----------------|:-----------------------------|:----------------------|:----|:----------------:|
193+
| GPT-4 | OpenAI | Cloud (Azure) | β‰ˆ 1 T + (MoE) | 0.3 – 1.8 | β‰ˆ 51 K – 62 K MWh | 1.1–1.3 | 0.3–0.4 |
194+
| Claude 3 Haiku | Anthropic | Cloud (AWS Bedrock) | β‰ˆ 20 B | 0.05 – 0.1 | 3 K – 10 K | β‰ˆ 1.2 | 0–0.2 |
195+
| Gemini Nano | Google DeepMind | On-Device (Android AICore) | 1.8–3.25 B | β‰ˆ 0.01 (on-device) | 200–1,200 | 1.10–1.12 | β‰ˆ 0.03 g COβ‚‚e /query |
196+
197+
---

0 commit comments

Comments
Β (0)