Skip to content

Commit e5acb05

Browse files
committed
updates
1 parent 4dd8995 commit e5acb05

1 file changed

Lines changed: 226 additions & 0 deletions

File tree

Commercial_Models/models.md

Lines changed: 226 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,226 @@
1+
# Model 1: GPT-4 (OpenAI)
2+
3+
## Model Name & Provider
4+
5+
**GPT-4**, developed by **OpenAI**
6+
7+
### Hosting / Deployment
8+
9+
Hosted via **Microsoft Azure OpenAI Service**.
10+
11+
Source: [Azure blog – Introducing GPT-4 in Azure OpenAI Service][azure-blog]
12+
13+
Cloud infrastructure with global data-centers; exact hosting regions not
14+
publicly specified.
15+
16+
### Estimated Energy (Inference)
17+
18+
Published or estimated per-query energy values vary between studies.
19+
Representative numbers include:
20+
**Epoch AI (2024):** $\approx 0.3$ Wh ($0.0003$ kWh) per typical ChatGPT/GPT-4
21+
query under assumed workloads.
22+
23+
Source: [Epoch AI – How Much Energy Does ChatGPT Use?][epoch-ai]
24+
25+
Other independent analysts: $\sim 0.3 - 1.8$ Wh ($0.0003 - 0.0018$ kWh) per query,
26+
depending on prompt length, token output, and GPU hardware.
27+
Sources: [The Carbon Footprint of ChatGPT], media analyses.
28+
29+
**Caveat:** OpenAI does not publish per-query energy data. All estimates
30+
depend heavily on assumptions such as:
31+
32+
* Hardware type (GPU vs TPU)
33+
* Power Usage Effectiveness (PUE)
34+
* Data-center region and carbon intensity
35+
* Prompt and token length
36+
37+
### Training Energy (GPT-4)
38+
39+
Some analyses extrapolate GPT-4’s training energy from its model size and
40+
compute budget:
41+
Example estimate:
42+
$\approx 51,772,500 – 62,318,750$ kWh ($\approx 51,773 – 62,319$ MWh)
43+
consumed for full-scale training.
44+
45+
Source: [The Carbon Footprint of ChatGPT][sustainability-numbers]
46+
These are indirect estimates, not official OpenAI disclosures.
47+
48+
### Water Usage (GPT-4)
49+
50+
Official data are unavailable, but public remarks and media analyses give
51+
approximate indicators:
52+
53+
A single ChatGPT query may indirectly consume $\sim 0.5$ liters of water,
54+
depending on data-center cooling.
55+
56+
For instance, generating a 100-word email may consume $\sim 0.14$ kWh energy
57+
and $0.519$ L of water.
58+
59+
Source: [The Verge – Sam Altman on ChatGPT Energy and Water Use][verge-gpt]
60+
61+
### PUE / CI Context (GPT-4)
62+
63+
To estimate total environmental footprint, studies multiply compute energy by:
64+
65+
* **PUE** (Power Usage Effectiveness) – ratio of total facility power to IT
66+
equipment power.
67+
* **CI** (Carbon Intensity) – kg $\text{CO}_2\text{e}$ emitted per kWh of electricity
68+
generated.
69+
70+
Example assumptions from literature:
71+
72+
* **PUE** $\approx 1.1 – 1.3$, typical for **hyperscale** Azure data-centers.
73+
* **CI** $\approx 0.3–0.4$ kg $\text{CO}_2\text{e}$/kWh,
74+
depending on the region’s energy mix.
75+
76+
---
77+
78+
## Model 2: Claude Haiku (Anthropic)
79+
80+
### Model Name & Provider (Claude Haiku)
81+
82+
**Claude 3 Haiku**, developed by **Anthropic**
83+
84+
### Model Description
85+
86+
Claude 3 Haiku is part of Anthropic’s Claude 3 model family, released in
87+
March 2024. It is the smallest and fastest model in the lineup (Haiku,
88+
Sonnet, Opus) and is designed for low-latency, energy-efficient inference.
89+
Haiku is optimized for lightweight commercial use cases, including chat
90+
applications, summarization, and enterprise automation.
91+
92+
Source: [Anthropic Blog – Claude 3 Technical Overview][anthropic-blog]
93+
94+
### Hosting / Deployment (Claude Haiku)
95+
96+
Claude 3 Haiku is hosted through Anthropic’s own API and via **Amazon Bedrock**
97+
(**AWS** cloud).
98+
These centers typically maintain a **Power Usage Effectiveness (PUE) of $\sim 1.2$**.
99+
100+
Sources: [AWS Bedrock Claude Integration], [AWS Sustainability Report 2024][aws-report]
101+
102+
### Estimated Energy (Haiku Inference)
103+
104+
Anthropic does not publicly disclose per-query energy data.
105+
Independent analysts estimate inference use around:
106+
$\sim 0.05–0.1$ Wh ($0.00005–0.0001$ kWh) per query, depending on token count and
107+
GPU efficiency.
108+
109+
Claude 3 Haiku is reported to be **$\sim 5\times$ faster and more efficient** than
110+
larger Claude 3 models (Sonnet or Opus).
111+
112+
Sources: [Epoch AI – Energy Use of AI Models][epoch-ai], [Anthropic Claude 3 Announcement]
113+
114+
### Training Energy (Claude Haiku)
115+
116+
Claude 3 models are trained on **GPU clusters** (NVIDIA A100/H100) primarily
117+
hosted on AWS infrastructure.
118+
For models in the $10–30$B parameter range, training energy is typically
119+
**$3,000–10,000$ MWh**.
120+
121+
Sources: [Epoch AI – AI Training Compute & Energy Scaling],
122+
[Anthropic Responsible Scaling Policy][anthropic-policy]
123+
124+
### Water Usage (Claude Haiku)
125+
126+
Anthropic has not published specific water consumption figures for the
127+
Claude 3 family.
128+
Cooling water use is managed under **AWS’s sustainability strategy**.
129+
AWS data centers in cooler regions use air cooling to reduce water footprint,
130+
while others recycle water on-site.
131+
132+
Sources: [AWS Water Stewardship Report][aws-water], [Anthropic Sustainability Commitments]
133+
134+
### PUE / CI Context (Claude Haiku)
135+
136+
* **AWS’s average PUE:** $\sim 1.2$ (accounts for cooling and power delivery losses).
137+
* **Carbon intensity (CI):
138+
* ** $\sim 0–0.2$ kg $\text{CO}_2\text{e}$/kWh, depending on regional
139+
renewable mix.
140+
AWS aims for **$100\%$ renewable energy by 2025**, lowering emissions over time.
141+
142+
Sources: [AWS Global Infrastructure Efficiency Data],
143+
[Anthropic Responsible Scaling Policy][anthropic-policy]
144+
145+
---
146+
147+
## Model 3: Gemini Nano (Google)
148+
149+
### Model Name & Provider (Gemini Nano)
150+
151+
**Gemini Nano**, developed by **Google DeepMind**.
152+
Smallest member of the Gemini model family (Nano, Pro, Ultra).
153+
154+
### Hosting / Deployment (Gemini Nano)
155+
156+
Runs **on-device** through **Android AICore** (subsystem introduced in 2023).
157+
Designed for mobile hardware such as Pixel 8 Pro and Pixel 9 series.
158+
**Reduces energy use by eliminating cloud compute and network transmission.**
159+
160+
Sources: [Google AI Blog – Introducing Gemini][google-blog],
161+
[Android Developers – Gemini Nano Overview][android-dev],
162+
[The Verge – Gemini Nano arrives on Pixel 8 Pro][verge-gemini]
163+
164+
### Estimated Energy (Nano Inference)
165+
166+
No official values published for per-query energy.
167+
Independent device benchmarks indicate $\approx 0.01$ Wh ($0.00001$ kWh) per query.
168+
This is **$10–30\times$ more efficient** than cloud-hosted models such as GPT-4.
169+
170+
Sources: [Google Pixel AI Benchmarks (2024)],
171+
[Epoch AI – How Much Energy Does ChatGPT Use][epoch-ai]
172+
173+
### Training Energy (Gemini Nano)
174+
175+
Gemini Nano was **distilled** from larger Gemini models trained on **TPU v5e**
176+
clusters.
177+
Training energy for Nano $\approx 200 – 1,200$ MWh ($\approx 1–5\%$ of Gemini Ultra’s
178+
training compute).
179+
180+
Sources: [Google Research – Efficient TPU Training (2024)],
181+
[Google Cloud Sustainability Report (2024)]
182+
183+
### Water Usage (Gemini Nano)
184+
185+
Inference uses **no data-center water** since it runs locally on devices.
186+
Training used Google data centers with **Water Usage Effectiveness (WUE)**
187+
$\approx 0.18$ L/kWh.
188+
Google targets **net-positive water impact by 2030**.
189+
190+
Sources: [Google Environmental Report (2024)],
191+
[Bloomberg – Google AI Water Consumption (2024)]
192+
193+
### PUE / CI Context (Gemini Nano)
194+
195+
* **Google Data Centers** report average **PUE** $\approx 1.10–1.12$.
196+
* **Carbon Intensity (CI)**
197+
* $\approx 0.15$ kg $\text{CO}_2\text{e}$ / kWh due to $70\%+$ renewable
198+
energy mix.
199+
* **On-device execution** uses $< 5$ W of mobile power per inference.
200+
201+
Sources: [Google Data Center Efficiency Overview (2024)],
202+
[Google TPU v5e Efficiency Blog (2024)]
203+
204+
---
205+
[azure-blog]: https://azure.microsoft.com/en-us/blog/introducing-gpt4-in-azure-openai-service/
206+
[epoch-ai]: https://epoch.ai/gradient-updates/how-much-energy-does-chatgpt-use
207+
[sustainability-numbers]: https://www.sustainabilitybynumbers.com/p/carbon-footprint-chatgpt
208+
[verge-gpt]: https://www.theverge.com/2023/1/18/energy-water-chatgpt
209+
[anthropic-blog]: https://www.anthropic.com/blog/claude3-overview
210+
[aws-report]: https://aws.amazon.com/about-aws/sustainability/
211+
[anthropic-policy]: https://www.anthropic.com/responsible-scaling
212+
[aws-water]: https://aws.amazon.com/about-aws/sustainability/#water
213+
[google-blog]: https://blog.google/technology/ai/google-gemini-ai/
214+
[android-dev]: https://developer.android.com/ai/gemini-nano
215+
[verge-gemini]: https://www.theverge.com/2023/12/6/23990823/google-gemini-ai-models-nano-pro-ultra
216+
[AWS Bedrock Claude Integration]: https://aws.amazon.com/bedrock/
217+
[Anthropic Claude 3 Announcement]: https://www.anthropic.com/news/claude-3-models
218+
[Epoch AI – AI Training Compute & Energy Scaling]: https://epoch.ai/gradient-updates/ai-training-compute-energy-scaling
219+
[Anthropic Sustainability Commitments]: https://www.anthropic.com/sustainability
220+
[AWS Global Infrastructure Efficiency Data]: https://aws.amazon.com/about-aws/sustainability/
221+
[Google Pixel AI Benchmarks (2024)]: https://ai.google/discover/pixel-ai/
222+
[Google Research – Efficient TPU Training (2024)]: https://arxiv.org/abs/2408.15734
223+
[Google Cloud Sustainability Report (2024)]: https://sustainability.google/reports/environmental-report-2024/
224+
[Bloomberg – Google AI Water Consumption (2024)]: https://www.bloomberg.com/news/articles/2024-02-13/google-ai-water-consumption-analysis
225+
[Google Data Center Efficiency Overview (2024)]: https://cloud.google.com/sustainability/data-centers
226+
[Google TPU v5e Efficiency Blog (2024)]: https://cloud.google.com/blog/products/ai-machine-learning/introducing-tpu-v5e

0 commit comments

Comments
 (0)