@@ -45,6 +45,13 @@ Source: [The Carbon Footprint of ChatGPT][sustainability-numbers].
4545
4646These are indirect estimates, not official OpenAI disclosures.
4747
48+ ### Model Size (GPT-4)
49+
50+ Estimated model size: ** β 1.8 trillion parameters** (widely reported
51+ estimate; OpenAI has not publicly confirmed exact parameter count).
52+
53+ Source: SemiAnalysis and other architecture analyses.
54+
4855### Water Usage (GPT-4)
4956
5057Official data are unavailable, but media analyses suggest:
@@ -84,6 +91,13 @@ energy-efficient inference in chat, summarization, and automation.
8491
8592Source: [ Anthropic Blog β Claude 3 Technical Overview] [ anthropic-blog ] .
8693
94+ ### Model Size / Architecture
95+
96+ Estimated model size: ** β 7 billion parameters** (Haiku variant,
97+ optimized for efficiency and low-latency inference).
98+
99+ Source: public model reports and community discussions.
100+
87101### Hosting & Deployment
88102
89103Hosted via Anthropic API and ** Amazon Bedrock (AWS)** .
@@ -97,33 +111,40 @@ Anthropic does not publish per-query energy data.
97111Independent analysts estimate β 0.05 β 0.1 Wh (0.00005 β 0.0001 kWh)
98112per query based on token count and GPU efficiency.
99113
100- Claude 3 Haiku is β 5Γ faster and more efficient than larger Claude 3 models.
114+ Claude 3 Haiku is β 5Γ faster and more efficient than larger Claude 3
115+ models.
101116
102- Sources: [ Epoch AI β Energy Use of AI Models] [ epoch-ai ] ,
103- [ Anthropic Claude 3 Announcement] .
117+ Sources: [ Epoch AI β Energy Use of AI Models] Sources:
118+ [ epoch-ai-training ] , [ Anthropic Claude 3 Announcement] .
104119
105120### Training Energy
106121
107- Claude 3 models use NVIDIA A100/H100 GPUs on AWS.
108- Typical energy use β 3 000 β 10 000 MWh for 10β30 B parameters.
122+ Claude 3 models are trained on GPU clusters (NVIDIA A100/H100) primarily
123+ hosted on AWS infrastructure.
124+ For models in the 10β30B parameter range, training energy is typically
125+ 3,000β10,000 MWh.
109126
110- Sources: [ Epoch AI β AI Training Compute and Energy Scaling] ,
111- [ Anthropic Responsible Scaling Policy] [ anthropic-policy ] .
127+ Sources: [ Epoch AI β AI Training Compute & Energy Scaling] ,
128+ [ Anthropic Responsible Scaling Policy] .
112129
113130### Water Usage
114131
115- No specific data published.
116- Cooling water managed under ** AWS sustainability strategy** .
117- Cooler regions use air cooling; others recycle water on-site.
132+ Anthropic has not published specific water consumption figures for the
133+ Claude 3 family.
134+ As it relies on AWS data centers, cooling water use is managed under AWS
135+ sustainability strategy.
136+ AWS data centers in cooler regions use air cooling to reduce water
137+ footprint, while others recycle water on-site.
118138
119139Sources: [ AWS Water Stewardship Report] [ aws-water ] ,
120140[ Anthropic Sustainability Commitments] .
121141
122142### PUE and CI Context
123143
124- * ** AWS PUE:** β 1.2
125- * ** Carbon Intensity:** β 0 β 0.2 kg COβe / kWh (depending on renewables)
126- AWS targets 100 % renewable energy by 2025.
144+ AWSβs average PUE: ~ 1.2 (accounts for cooling and power delivery losses).
145+ Carbon intensity (CI): ~ 0β0.2 kg COβe/kWh, depending on regional renewable
146+ mix.
147+ AWS aims for 100% renewable energy by 2025, lowering emissions over time.
127148
128149Sources: [ AWS Global Infrastructure Efficiency Data] ,
129150[ Anthropic Responsible Scaling Policy] [ anthropic-policy ] .
@@ -147,37 +168,51 @@ Sources: [Google AI Blog β Introducing Gemini][google-blog],
147168[ Android Developers β Gemini Nano Overview] [ android-dev ] ,
148169[ The Verge β Gemini Nano on Pixel 8 Pro] [ verge-gemini ] .
149170
150- ### Estimated Energy(Inference)
171+ ### Estimated Model Size / Architecture
172+
173+ Gemini Nano variants (device-optimized):
174+
175+ * ** Nano-1:** β 1.8 billion parameters
176+ * ** Nano-2:** (larger device variant) β 3.25 billion parameters
177+
178+ These use quantized weights tuned for on-device inference.
179+
180+ Source: device benchmark reports and public model parameter listings.
181+
182+ ### Estimated Energy (Inference) gemini
151183
152184No official values.
153- Device benchmarks show β 0.01 Wh (0.00001 kWh) per query β
185+ Device benchmarks show β 0.01 Wh (0.00001 kWh) per query β
15418610 β 30Γ more efficient than GPT-4.
155187
156188Sources: [ Google Pixel AI Benchmarks (2024)] ,
157189[ Epoch AI β How Much Energy Does ChatGPT Use] [ epoch-ai ] .
158190
159- ### Training Energy of gemini
191+ ### Training Energy Estimates
160192
161- Gemini Nano is distilled from larger Gemini models trained on ** TPU v5e** .
162- Training energy β 200 β 1 200 MWh (1 β 5 % of Gemini Ultra).
193+ Gemini Nano was distilled from larger Gemini models trained on ** TPU v5e**
194+ clusters.
195+ Training energy for Nano β 200 β 1,200 MWh (β 1β5% of Gemini Ultraβs
196+ training compute).
163197
164198Sources: [ Google Research β Efficient TPU Training (2024)] ,
165199[ Google Cloud Sustainability Report (2024)] .
166200
167- ### Water Usage (nano )
201+ ### Water Usage (Nano )
168202
169- Inference uses no data-center water.
170- Training used Google data centers with ** WUE β 0.18 L/kWh** .
203+ Inference uses no data-center water since it runs locally on devices.
204+ Training used Google data centers with Water Usage Effectiveness (WUE)
205+ β 0.18 L/kWh.
171206Google targets net-positive water impact by 2030.
172207
173208Sources: [ Google Environmental Report (2024)] ,
174209[ Bloomberg β Google AI Water Consumption (2024)] .
175210
176211### PUE & CI Context
177212
178- * ** PUE: ** β 1.10 β 1.12 (Google Data Centers)
179- * ** CI: ** β 0.15 kg COβe / kWh (70 % renewable mix)
180- * ** On-device: ** < 5 W per inference
213+ Google Data Centers report average PUE β 1.10β 1.12.
214+ Carbon Intensity (CI) β 0.15 kg COβe / kWh due to 70%+ renewable energy mix.
215+ On-device execution uses < 5 W of mobile power per inference.
181216
182217Sources: [ Google Data Center Efficiency Overview (2024)] ,
183218[ Google TPU v5e Efficiency Blog (2024)] .
@@ -210,7 +245,7 @@ https://www.theverge.com/2023/12/6/23990823/google-gemini-ai-models-nano-pro-ult
210245https://aws.amazon.com/bedrock/
211246[ Anthropic Claude 3 Announcement] :
212247https://www.anthropic.com/news/claude-3-models
213- [ Epoch AI β AI Training Compute and Energy Scaling ] :
248+ [ epoch-ai-training ] :
214249https://epoch.ai/gradient-updates/ai-training-compute-energy-scaling
215250[ Anthropic Sustainability Commitments] :
216251https://www.anthropic.com/sustainability
0 commit comments