You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
-**PUE:** ≈ 1.1 – 1.3 for Azure hyperscale centers
72
-
-**CI:** ≈ 0.3 – 0.4 kg CO₂e / kWh (depending on region)
79
+
-**CI:** ≈ 0.3 – 0.4 kg CO₂e/kWh (depending on region)
73
80
74
81
---
75
82
76
83
# Model 2: Claude 3 Haiku (Anthropic)
77
84
78
-
## Model Name & Provider
79
-
80
-
**Claude 3 Haiku**, developed by **Anthropic**.
81
-
82
85
### Model Description
83
86
84
-
Part of Anthropic’s Claude 3 family (Haiku, Sonnet, Opus).
85
-
Released March 2024. Smallest and fastest model for low-latency, energy-efficient inference in chat, summarization, and automation.
87
+
Claude 3 Haiku is part of Anthropic’s Claude 3 model family, released in March 2024. It is the smallest and fastest model in the Claude 3 lineup (Haiku, Sonnet, Opus) and is designed for low-latency, energy-efficient inference while maintaining strong reasoning capabilities. Haiku is optimized for lightweight commercial use cases, including chat applications, summarization, and enterprise automation.
86
88
87
-
Source: [Anthropic Blog – Claude 3 Technical Overview][anthropic-blog].
89
+
Sources:
90
+
[Anthropic – Claude 3 Model Family Overview][anthropic-claude3]
91
+
[Anthropic Blog – Claude 3 Announcement][anthropic-blog]
88
92
89
-
###Model Size and Architecture
93
+
#### Hosting / Deployment
90
94
91
-
Estimated model size: **≈ 7 billion parameters** (Haiku variant, optimized for efficiency and low-latency inference).
92
-
Source: public model reports and community discussions.
95
+
Claude 3 Haiku is hosted via Anthropic’s API and via AWS Amazon Bedrock (cloud). These data centers typically maintain a **PUE ≈ 1.2**.
93
96
94
-
### Hosting & Deployment
97
+
Sources:
98
+
[AWS Bedrock Claude Integration][aws-bedrock]
99
+
[AWS Sustainability Report][aws-sustainability]
95
100
96
-
Hosted via Anthropic API and **Amazon Bedrock (AWS)**.
97
-
These centers maintain **PUE ≈ 1.2**.
101
+
#### Estimated Model Size / Architecture
98
102
99
-
Sources: [AWS Bedrock Claude Integration][aws-bedrock], [AWS Sustainability Report 2024][aws-report].
103
+
Claude 3 Haiku is estimated to have **≈ 20 billion parameters**, making it significantly smaller than larger models in the family (e.g., Claude 3 Opus). This parameter estimate is based on community sources; Anthropic does not publicly confirm.
Independent analysts estimate ≈ 0.05 – 0.1 Wh (0.00005 – 0.0001 kWh) per query based on token count and GPU efficiency.
108
+
#### Estimated Energy
105
109
106
-
Claude 3 Haiku is ≈ 5× faster and more efficient than larger Claude 3 models.
110
+
Anthropic does not publish per-query energy data. Independent analysts estimate models of similar size (10–30 billion parameters) use ~0.05-0.1 Wh (0.00005-0.0001 kWh) per query, depending on hardware and tokens. Haiku is reportedly ~5× more efficient than larger Claude variants.
107
111
108
-
Sources: [Epoch AI – AI Training Compute and Energy Scaling][epoch-ai-training], [Anthropic Claude 3 Announcement][anthropic-announcement].
112
+
Sources:
113
+
[Epoch AI – Energy Use of AI Models][epoch-energy]
114
+
[Anthropic Claude 3 Announcement][anthropic-announcement]
109
115
110
-
### Training Energy
116
+
####Training Energy Estimates
111
117
112
-
Claude 3 models are trained on GPU clusters (NVIDIA A100/H100) primarily hosted on AWS infrastructure.
113
-
For models in the 10 – 30 B parameter range, training energy is typically **3 000 – 10 000 MWh**.
118
+
For models in the 10–30 billion parameter range, training energy is estimated at **3,000-10,000 MWh**, depending on runs and infrastructure.
114
119
115
-
Sources: [Epoch AI – AI Training Compute and Energy Scaling][epoch-ai-training], [Anthropic Responsible Scaling Policy][anthropic-policy].
120
+
Sources:
121
+
[Epoch AI – AI Training Compute & Energy Scaling][epoch-training]
Anthropic has not published specific water consumption figures for the Claude 3 family.
120
-
As it relies on AWS data centers, cooling water use is managed under AWS sustainability strategy.
121
-
AWS data centers in cooler regions use air cooling to reduce water footprint, while others recycle water on-site.
126
+
Anthropic has not published water-consumption data for Claude 3. AWS manages cooling water use via its sustainability programs. Some centers use air-cooling or recycle water on-site to reduce water usage.
122
127
123
-
Sources: [AWS Water Stewardship Report][aws-water], [Anthropic Sustainability Commitments][anthropic-sustainability].
0 commit comments