Skip to content

Commit 7a607be

Browse files
committed
updates
1 parent 34104ac commit 7a607be

1 file changed

Lines changed: 230 additions & 0 deletions

File tree

Commercial_Models/models.md

Lines changed: 230 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,230 @@
1+
# Model 1: GPT-4 (OpenAI)
2+
3+
## Model Name and Provider
4+
5+
**GPT-4**, developed by **OpenAI**.
6+
7+
### Hosting and Deployment
8+
9+
Hosted via **Microsoft Azure OpenAI Service**.
10+
11+
Source: [Azure blog – Introducing GPT-4 in Azure OpenAI Service][azure-blog].
12+
13+
Cloud infrastructure uses global data centers; regions are not public.
14+
15+
### Estimated Energy (Inference)
16+
17+
Published or estimated per-query energy values vary between studies.
18+
Representative numbers include:
19+
20+
**Epoch AI (2024):** ≈ 0.3 Wh (0.0003 kWh) per ChatGPT/GPT-4 query.
21+
22+
Source: [Epoch AI – How Much Energy Does ChatGPT Use?][epoch-ai].
23+
24+
Other analysts estimate ≈ 0.3 – 1.8 Wh (0.0003 – 0.0018 kWh)
25+
depending on prompt length, token output, and GPU hardware.
26+
27+
Sources: “The Carbon Footprint of ChatGPT,” media analyses.
28+
29+
**Caveat:** OpenAI does not publish per-query energy data.
30+
All estimates depend on assumptions such as:
31+
32+
* Hardware type (GPU vs TPU)
33+
* Power Usage Effectiveness (PUE)
34+
* Data center region and carbon intensity
35+
* Prompt and token length
36+
37+
### Training Energy (GPT-4)
38+
39+
Some analyses extrapolate GPT-4’s training energy from model size and
40+
compute budget:
41+
42+
≈ 51 – 62 GWh (51 772 500 – 62 318 750 kWh) for full-scale training.
43+
44+
Source: [The Carbon Footprint of ChatGPT][sustainability-numbers].
45+
46+
These are indirect estimates, not official OpenAI disclosures.
47+
48+
### Water Usage (GPT-4)
49+
50+
Official data are unavailable, but media analyses suggest:
51+
52+
A single ChatGPT query may indirectly consume ≈ 0.5 L of water,
53+
depending on data-center cooling.
54+
55+
Generating a 100-word email may use ≈ 0.14 kWh energy and 0.52 L water.
56+
57+
Source: [The Verge – Sam Altman on ChatGPT Energy and Water Use][verge-gpt].
58+
59+
### PUE and CI Context (GPT-4)
60+
61+
Studies multiply compute energy by:
62+
63+
* **PUE** – Power Usage Effectiveness (total facility power / IT power)
64+
* **CI** – Carbon Intensity (kg CO₂e / kWh electricity)
65+
66+
Example assumptions:
67+
68+
* **PUE:** ≈ 1.1 – 1.3 for Azure hyperscale centers
69+
* **CI:** ≈ 0.3 – 0.4 kg CO₂e / kWh (depending on region)
70+
71+
---
72+
73+
## Model 2: Claude Haiku (Anthropic)
74+
75+
### Model Name & Provider
76+
77+
**Claude 3 Haiku**, developed by **Anthropic**.
78+
79+
### Model Description
80+
81+
Part of Anthropic’s Claude 3 family (Haiku, Sonnet, Opus).
82+
Released March 2024. Smallest and fastest model for low-latency,
83+
energy-efficient inference in chat, summarization, and automation.
84+
85+
Source: [Anthropic Blog – Claude 3 Technical Overview][anthropic-blog].
86+
87+
### Hosting & Deployment
88+
89+
Hosted via Anthropic API and **Amazon Bedrock (AWS)**.
90+
These centers maintain **PUE ≈ 1.2**.
91+
92+
Sources: [AWS Bedrock Claude Integration], [AWS Sustainability Report 2024][aws-report].
93+
94+
### Estimated Energy
95+
96+
Anthropic does not publish per-query energy data.
97+
Independent analysts estimate ≈ 0.05 – 0.1 Wh (0.00005 – 0.0001 kWh)
98+
per query based on token count and GPU efficiency.
99+
100+
Claude 3 Haiku is ≈ 5× faster and more efficient than larger Claude 3 models.
101+
102+
Sources: [Epoch AI – Energy Use of AI Models][epoch-ai],
103+
[Anthropic Claude 3 Announcement].
104+
105+
### Training Energy
106+
107+
Claude 3 models use NVIDIA A100/H100 GPUs on AWS.
108+
Typical energy use ≈ 3 000 – 10 000 MWh for 10–30 B parameters.
109+
110+
Sources: [Epoch AI – AI Training Compute and Energy Scaling],
111+
[Anthropic Responsible Scaling Policy][anthropic-policy].
112+
113+
### Water Usage
114+
115+
No specific data published.
116+
Cooling water managed under **AWS sustainability strategy**.
117+
Cooler regions use air cooling; others recycle water on-site.
118+
119+
Sources: [AWS Water Stewardship Report][aws-water],
120+
[Anthropic Sustainability Commitments].
121+
122+
### PUE and CI Context
123+
124+
* **AWS PUE:** ≈ 1.2
125+
* **Carbon Intensity:** ≈ 0 – 0.2 kg CO₂e / kWh (depending on renewables)
126+
AWS targets 100 % renewable energy by 2025.
127+
128+
Sources: [AWS Global Infrastructure Efficiency Data],
129+
[Anthropic Responsible Scaling Policy][anthropic-policy].
130+
131+
---
132+
133+
## Model 3: Gemini Nano (Google)
134+
135+
### Provider
136+
137+
**Gemini Nano**, developed by **Google DeepMind**.
138+
Smallest member of Gemini family (Nano, Pro, Ultra).
139+
140+
### Hosting
141+
142+
Runs on-device via **Android AICore** (subsystem introduced 2023).
143+
Designed for mobile hardware like Pixel 8 Pro and Pixel 9.
144+
Reduces energy use by eliminating cloud compute and network load.
145+
146+
Sources: [Google AI Blog – Introducing Gemini][google-blog],
147+
[Android Developers – Gemini Nano Overview][android-dev],
148+
[The Verge – Gemini Nano on Pixel 8 Pro][verge-gemini].
149+
150+
### Estimated Energy(Inference)
151+
152+
No official values.
153+
Device benchmarks show ≈ 0.01 Wh (0.00001 kWh) per query —
154+
10 – 30× more efficient than GPT-4.
155+
156+
Sources: [Google Pixel AI Benchmarks (2024)],
157+
[Epoch AI – How Much Energy Does ChatGPT Use][epoch-ai].
158+
159+
### Training Energy of gemini
160+
161+
Gemini Nano is distilled from larger Gemini models trained on **TPU v5e**.
162+
Training energy ≈ 200 – 1 200 MWh (1 – 5 % of Gemini Ultra).
163+
164+
Sources: [Google Research – Efficient TPU Training (2024)],
165+
[Google Cloud Sustainability Report (2024)].
166+
167+
### Water Usage (nano)
168+
169+
Inference uses no data-center water.
170+
Training used Google data centers with **WUE ≈ 0.18 L/kWh**.
171+
Google targets net-positive water impact by 2030.
172+
173+
Sources: [Google Environmental Report (2024)],
174+
[Bloomberg – Google AI Water Consumption (2024)].
175+
176+
### PUE & CI Context
177+
178+
* **PUE:** ≈ 1.10 – 1.12 (Google Data Centers)
179+
* **CI:** ≈ 0.15 kg CO₂e / kWh (70 % renewable mix)
180+
* **On-device:** < 5 W per inference
181+
182+
Sources: [Google Data Center Efficiency Overview (2024)],
183+
[Google TPU v5e Efficiency Blog (2024)].
184+
185+
---
186+
187+
[azure-blog]:
188+
https://azure.microsoft.com/en-us/blog/introducing-gpt4-in-azure-openai-service/
189+
[epoch-ai]:
190+
https://epoch.ai/gradient-updates/how-much-energy-does-chatgpt-use
191+
[sustainability-numbers]:
192+
https://www.sustainabilitybynumbers.com/p/carbon-footprint-chatgpt
193+
[verge-gpt]:
194+
https://www.theverge.com/2023/1/18/energy-water-chatgpt
195+
[anthropic-blog]:
196+
https://www.anthropic.com/blog/claude3-overview
197+
[aws-report]:
198+
https://aws.amazon.com/about-aws/sustainability/
199+
[anthropic-policy]:
200+
https://www.anthropic.com/responsible-scaling
201+
[aws-water]:
202+
https://aws.amazon.com/about-aws/sustainability/#water
203+
[google-blog]:
204+
https://blog.google/technology/ai/google-gemini-ai/
205+
[android-dev]:
206+
https://developer.android.com/ai/gemini-nano
207+
[verge-gemini]:
208+
https://www.theverge.com/2023/12/6/23990823/google-gemini-ai-models-nano-pro-ultra
209+
[AWS Bedrock Claude Integration]:
210+
https://aws.amazon.com/bedrock/
211+
[Anthropic Claude 3 Announcement]:
212+
https://www.anthropic.com/news/claude-3-models
213+
[Epoch AI – AI Training Compute and Energy Scaling]:
214+
https://epoch.ai/gradient-updates/ai-training-compute-energy-scaling
215+
[Anthropic Sustainability Commitments]:
216+
https://www.anthropic.com/sustainability
217+
[AWS Global Infrastructure Efficiency Data]:
218+
https://aws.amazon.com/about-aws/sustainability/
219+
[Google Pixel AI Benchmarks (2024)]:
220+
https://ai.google/discover/pixel-ai/
221+
[Google Research – Efficient TPU Training (2024)]:
222+
https://arxiv.org/abs/2408.15734
223+
[Google Cloud Sustainability Report (2024)]:
224+
https://sustainability.google/reports/environmental-report-2024/
225+
[Bloomberg – Google AI Water Consumption (2024)]:
226+
https://www.bloomberg.com/news/articles/2024-02-13/google-ai-water-consumption-analysis
227+
[Google Data Center Efficiency Overview (2024)]:
228+
https://cloud.google.com/sustainability/data-centers
229+
[Google TPU v5e Efficiency Blog (2024)]:
230+
https://cloud.google.com/blog/products/ai-machine-learning/introducing-tpu-v5e

0 commit comments

Comments
 (0)