Skip to content

Commit 05f29b2

Browse files
committed
updates
1 parent f553342 commit 05f29b2

2 files changed

Lines changed: 60 additions & 255 deletions

File tree

β€Žcommercial_models/models.mdβ€Ž

Lines changed: 60 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -45,6 +45,13 @@ Source: [The Carbon Footprint of ChatGPT][sustainability-numbers].
4545

4646
These are indirect estimates, not official OpenAI disclosures.
4747

48+
### Model Size (GPT-4)
49+
50+
Estimated model size: **β‰ˆ 1.8 trillion parameters** (widely reported
51+
estimate; OpenAI has not publicly confirmed exact parameter count).
52+
53+
Source: SemiAnalysis and other architecture analyses.
54+
4855
### Water Usage (GPT-4)
4956

5057
Official data are unavailable, but media analyses suggest:
@@ -84,6 +91,13 @@ energy-efficient inference in chat, summarization, and automation.
8491

8592
Source: [Anthropic Blog – Claude 3 Technical Overview][anthropic-blog].
8693

94+
### Model Size / Architecture
95+
96+
Estimated model size: **β‰ˆ 7 billion parameters** (Haiku variant,
97+
optimized for efficiency and low-latency inference).
98+
99+
Source: public model reports and community discussions.
100+
87101
### Hosting & Deployment
88102

89103
Hosted via Anthropic API and **Amazon Bedrock (AWS)**.
@@ -97,33 +111,40 @@ Anthropic does not publish per-query energy data.
97111
Independent analysts estimate β‰ˆ 0.05 – 0.1 Wh (0.00005 – 0.0001 kWh)
98112
per query based on token count and GPU efficiency.
99113

100-
Claude 3 Haiku is β‰ˆ 5Γ— faster and more efficient than larger Claude 3 models.
114+
Claude 3 Haiku is β‰ˆ 5Γ— faster and more efficient than larger Claude 3
115+
models.
101116

102-
Sources: [Epoch AI – Energy Use of AI Models][epoch-ai],
103-
[Anthropic Claude 3 Announcement].
117+
Sources: [Epoch AI – Energy Use of AI Models]Sources:
118+
[epoch-ai-training], [Anthropic Claude 3 Announcement].
104119

105120
### Training Energy
106121

107-
Claude 3 models use NVIDIA A100/H100 GPUs on AWS.
108-
Typical energy use β‰ˆ 3 000 – 10 000 MWh for 10–30 B parameters.
122+
Claude 3 models are trained on GPU clusters (NVIDIA A100/H100) primarily
123+
hosted on AWS infrastructure.
124+
For models in the 10–30B parameter range, training energy is typically
125+
3,000–10,000 MWh.
109126

110-
Sources: [Epoch AI – AI Training Compute and Energy Scaling],
111-
[Anthropic Responsible Scaling Policy][anthropic-policy].
127+
Sources: [Epoch AI – AI Training Compute & Energy Scaling],
128+
[Anthropic Responsible Scaling Policy].
112129

113130
### Water Usage
114131

115-
No specific data published.
116-
Cooling water managed under **AWS sustainability strategy**.
117-
Cooler regions use air cooling; others recycle water on-site.
132+
Anthropic has not published specific water consumption figures for the
133+
Claude 3 family.
134+
As it relies on AWS data centers, cooling water use is managed under AWS
135+
sustainability strategy.
136+
AWS data centers in cooler regions use air cooling to reduce water
137+
footprint, while others recycle water on-site.
118138

119139
Sources: [AWS Water Stewardship Report][aws-water],
120140
[Anthropic Sustainability Commitments].
121141

122142
### PUE and CI Context
123143

124-
* **AWS PUE:** β‰ˆ 1.2
125-
* **Carbon Intensity:** β‰ˆ 0 – 0.2 kg COβ‚‚e / kWh (depending on renewables)
126-
AWS targets 100 % renewable energy by 2025.
144+
AWS’s average PUE: ~1.2 (accounts for cooling and power delivery losses).
145+
Carbon intensity (CI): ~0–0.2 kg COβ‚‚e/kWh, depending on regional renewable
146+
mix.
147+
AWS aims for 100% renewable energy by 2025, lowering emissions over time.
127148

128149
Sources: [AWS Global Infrastructure Efficiency Data],
129150
[Anthropic Responsible Scaling Policy][anthropic-policy].
@@ -147,37 +168,51 @@ Sources: [Google AI Blog – Introducing Gemini][google-blog],
147168
[Android Developers – Gemini Nano Overview][android-dev],
148169
[The Verge – Gemini Nano on Pixel 8 Pro][verge-gemini].
149170

150-
### Estimated Energy(Inference)
171+
### Estimated Model Size / Architecture
172+
173+
Gemini Nano variants (device-optimized):
174+
175+
* **Nano-1:** β‰ˆ 1.8 billion parameters
176+
* **Nano-2:** (larger device variant) β‰ˆ 3.25 billion parameters
177+
178+
These use quantized weights tuned for on-device inference.
179+
180+
Source: device benchmark reports and public model parameter listings.
181+
182+
### Estimated Energy (Inference) gemini
151183

152184
No official values.
153-
Device benchmarks show β‰ˆ 0.01 Wh (0.00001 kWh) per query β€”
185+
Device benchmarks show β‰ˆ 0.01 Wh (0.00001 kWh) per query β€”
154186
10 – 30Γ— more efficient than GPT-4.
155187

156188
Sources: [Google Pixel AI Benchmarks (2024)],
157189
[Epoch AI – How Much Energy Does ChatGPT Use][epoch-ai].
158190

159-
### Training Energy of gemini
191+
### Training Energy Estimates
160192

161-
Gemini Nano is distilled from larger Gemini models trained on **TPU v5e**.
162-
Training energy β‰ˆ 200 – 1 200 MWh (1 – 5 % of Gemini Ultra).
193+
Gemini Nano was distilled from larger Gemini models trained on **TPU v5e**
194+
clusters.
195+
Training energy for Nano β‰ˆ 200 – 1,200 MWh (β‰ˆ 1–5% of Gemini Ultra’s
196+
training compute).
163197

164198
Sources: [Google Research – Efficient TPU Training (2024)],
165199
[Google Cloud Sustainability Report (2024)].
166200

167-
### Water Usage (nano)
201+
### Water Usage (Nano)
168202

169-
Inference uses no data-center water.
170-
Training used Google data centers with **WUE β‰ˆ 0.18 L/kWh**.
203+
Inference uses no data-center water since it runs locally on devices.
204+
Training used Google data centers with Water Usage Effectiveness (WUE)
205+
β‰ˆ 0.18 L/kWh.
171206
Google targets net-positive water impact by 2030.
172207

173208
Sources: [Google Environmental Report (2024)],
174209
[Bloomberg – Google AI Water Consumption (2024)].
175210

176211
### PUE & CI Context
177212

178-
* **PUE:** β‰ˆ 1.10 – 1.12 (Google Data Centers)
179-
* **CI:** β‰ˆ 0.15 kg COβ‚‚e / kWh (70 % renewable mix)
180-
* **On-device:** < 5 W per inference
213+
Google Data Centers report average PUE β‰ˆ 1.10–1.12.
214+
Carbon Intensity (CI) β‰ˆ 0.15 kg COβ‚‚e / kWh due to 70%+ renewable energy mix.
215+
On-device execution uses < 5 W of mobile power per inference.
181216

182217
Sources: [Google Data Center Efficiency Overview (2024)],
183218
[Google TPU v5e Efficiency Blog (2024)].
@@ -210,7 +245,7 @@ https://www.theverge.com/2023/12/6/23990823/google-gemini-ai-models-nano-pro-ult
210245
https://aws.amazon.com/bedrock/
211246
[Anthropic Claude 3 Announcement]:
212247
https://www.anthropic.com/news/claude-3-models
213-
[Epoch AI – AI Training Compute and Energy Scaling]:
248+
[epoch-ai-training]:
214249
https://epoch.ai/gradient-updates/ai-training-compute-energy-scaling
215250
[Anthropic Sustainability Commitments]:
216251
https://www.anthropic.com/sustainability

β€Žtemp_models/models.mdβ€Ž

Lines changed: 0 additions & 230 deletions
This file was deleted.

0 commit comments

Comments
Β (0)