|
| 1 | +# Model 1: GPT-4 (OpenAI) |
| 2 | + |
| 3 | +## Model Name & Provider |
| 4 | + |
| 5 | +**GPT-4**, developed by **OpenAI** |
| 6 | + |
| 7 | +### Hosting / Deployment |
| 8 | + |
| 9 | +Hosted via **Microsoft Azure OpenAI Service**. |
| 10 | + |
| 11 | +Source: [Azure blog – Introducing GPT-4 in Azure OpenAI Service][azure-blog] |
| 12 | + |
| 13 | +Cloud infrastructure with global data-centers; exact hosting regions not |
| 14 | +publicly specified. |
| 15 | + |
| 16 | +### Estimated Energy (Inference) |
| 17 | + |
| 18 | +Published or estimated per-query energy values vary between studies. |
| 19 | +Representative numbers include: |
| 20 | +**Epoch AI (2024):** $\approx 0.3$ Wh ($0.0003$ kWh) per typical ChatGPT/GPT-4 |
| 21 | +query under assumed workloads. |
| 22 | + |
| 23 | +Source: [Epoch AI – How Much Energy Does ChatGPT Use?][epoch-ai] |
| 24 | + |
| 25 | +Other independent analysts: $\sim 0.3 - 1.8$ Wh ($0.0003 - 0.0018$ kWh) per query, |
| 26 | +depending on prompt length, token output, and GPU hardware. |
| 27 | +Sources: [The Carbon Footprint of ChatGPT], media analyses. |
| 28 | + |
| 29 | +**Caveat:** OpenAI does not publish per-query energy data. All estimates |
| 30 | +depend heavily on assumptions such as: |
| 31 | + |
| 32 | +* Hardware type (GPU vs TPU) |
| 33 | +* Power Usage Effectiveness (PUE) |
| 34 | +* Data-center region and carbon intensity |
| 35 | +* Prompt and token length |
| 36 | + |
| 37 | +### Training Energy (GPT-4) |
| 38 | + |
| 39 | +Some analyses extrapolate GPT-4’s training energy from its model size and |
| 40 | +compute budget: |
| 41 | +Example estimate: |
| 42 | +$\approx 51,772,500 – 62,318,750$ kWh ($\approx 51,773 – 62,319$ MWh) |
| 43 | +consumed for full-scale training. |
| 44 | + |
| 45 | +Source: [The Carbon Footprint of ChatGPT][sustainability-numbers] |
| 46 | +These are indirect estimates, not official OpenAI disclosures. |
| 47 | + |
| 48 | +### Water Usage (GPT-4) |
| 49 | + |
| 50 | +Official data are unavailable, but public remarks and media analyses give |
| 51 | +approximate indicators: |
| 52 | + |
| 53 | +A single ChatGPT query may indirectly consume $\sim 0.5$ liters of water, |
| 54 | +depending on data-center cooling. |
| 55 | + |
| 56 | +For instance, generating a 100-word email may consume $\sim 0.14$ kWh energy |
| 57 | +and $0.519$ L of water. |
| 58 | + |
| 59 | +Source: [The Verge – Sam Altman on ChatGPT Energy and Water Use][verge-gpt] |
| 60 | + |
| 61 | +### PUE / CI Context (GPT-4) |
| 62 | + |
| 63 | +To estimate total environmental footprint, studies multiply compute energy by: |
| 64 | + |
| 65 | +* **PUE** (Power Usage Effectiveness) – ratio of total facility power to IT |
| 66 | + equipment power. |
| 67 | +* **CI** (Carbon Intensity) – kg $\text{CO}_2\text{e}$ emitted per kWh of electricity |
| 68 | + generated. |
| 69 | + |
| 70 | +Example assumptions from literature: |
| 71 | + |
| 72 | +* **PUE** $\approx 1.1 – 1.3$, typical for **hyperscale** Azure data-centers. |
| 73 | +* **CI** $\approx 0.3–0.4$ kg $\text{CO}_2\text{e}$/kWh, |
| 74 | + depending on the region’s energy mix. |
| 75 | + |
| 76 | +--- |
| 77 | + |
| 78 | +## Model 2: Claude Haiku (Anthropic) |
| 79 | + |
| 80 | +### Model Name & Provider (Claude Haiku) |
| 81 | + |
| 82 | +**Claude 3 Haiku**, developed by **Anthropic** |
| 83 | + |
| 84 | +### Model Description |
| 85 | + |
| 86 | +Claude 3 Haiku is part of Anthropic’s Claude 3 model family, released in |
| 87 | +March 2024. It is the smallest and fastest model in the lineup (Haiku, |
| 88 | +Sonnet, Opus) and is designed for low-latency, energy-efficient inference. |
| 89 | +Haiku is optimized for lightweight commercial use cases, including chat |
| 90 | +applications, summarization, and enterprise automation. |
| 91 | + |
| 92 | +Source: [Anthropic Blog – Claude 3 Technical Overview][anthropic-blog] |
| 93 | + |
| 94 | +### Hosting / Deployment (Claude Haiku) |
| 95 | + |
| 96 | +Claude 3 Haiku is hosted through Anthropic’s own API and via **Amazon Bedrock** |
| 97 | +(**AWS** cloud). |
| 98 | +These centers typically maintain a **Power Usage Effectiveness (PUE) of $\sim 1.2$**. |
| 99 | + |
| 100 | +Sources: [AWS Bedrock Claude Integration], [AWS Sustainability Report 2024][aws-report] |
| 101 | + |
| 102 | +### Estimated Energy (Haiku Inference) |
| 103 | + |
| 104 | +Anthropic does not publicly disclose per-query energy data. |
| 105 | +Independent analysts estimate inference use around: |
| 106 | +$\sim 0.05–0.1$ Wh ($0.00005–0.0001$ kWh) per query, depending on token count and |
| 107 | +GPU efficiency. |
| 108 | + |
| 109 | +Claude 3 Haiku is reported to be **$\sim 5\times$ faster and more efficient** than |
| 110 | +larger Claude 3 models (Sonnet or Opus). |
| 111 | + |
| 112 | +Sources: [Epoch AI – Energy Use of AI Models][epoch-ai], [Anthropic Claude 3 Announcement] |
| 113 | + |
| 114 | +### Training Energy (Claude Haiku) |
| 115 | + |
| 116 | +Claude 3 models are trained on **GPU clusters** (NVIDIA A100/H100) primarily |
| 117 | +hosted on AWS infrastructure. |
| 118 | +For models in the $10–30$B parameter range, training energy is typically |
| 119 | +**$3,000–10,000$ MWh**. |
| 120 | + |
| 121 | +Sources: [Epoch AI – AI Training Compute & Energy Scaling], |
| 122 | + [Anthropic Responsible Scaling Policy][anthropic-policy] |
| 123 | + |
| 124 | +### Water Usage (Claude Haiku) |
| 125 | + |
| 126 | +Anthropic has not published specific water consumption figures for the |
| 127 | +Claude 3 family. |
| 128 | +Cooling water use is managed under **AWS’s sustainability strategy**. |
| 129 | +AWS data centers in cooler regions use air cooling to reduce water footprint, |
| 130 | +while others recycle water on-site. |
| 131 | + |
| 132 | +Sources: [AWS Water Stewardship Report][aws-water], [Anthropic Sustainability Commitments] |
| 133 | + |
| 134 | +### PUE / CI Context (Claude Haiku) |
| 135 | + |
| 136 | +* **AWS’s average PUE:** $\sim 1.2$ (accounts for cooling and power delivery losses). |
| 137 | +* **Carbon intensity (CI): |
| 138 | +* ** $\sim 0–0.2$ kg $\text{CO}_2\text{e}$/kWh, depending on regional |
| 139 | + renewable mix. |
| 140 | +AWS aims for **$100\%$ renewable energy by 2025**, lowering emissions over time. |
| 141 | + |
| 142 | +Sources: [AWS Global Infrastructure Efficiency Data], |
| 143 | + [Anthropic Responsible Scaling Policy][anthropic-policy] |
| 144 | + |
| 145 | +--- |
| 146 | + |
| 147 | +## Model 3: Gemini Nano (Google) |
| 148 | + |
| 149 | +### Model Name & Provider (Gemini Nano) |
| 150 | + |
| 151 | +**Gemini Nano**, developed by **Google DeepMind**. |
| 152 | +Smallest member of the Gemini model family (Nano, Pro, Ultra). |
| 153 | + |
| 154 | +### Hosting / Deployment (Gemini Nano) |
| 155 | + |
| 156 | +Runs **on-device** through **Android AICore** (subsystem introduced in 2023). |
| 157 | +Designed for mobile hardware such as Pixel 8 Pro and Pixel 9 series. |
| 158 | +**Reduces energy use by eliminating cloud compute and network transmission.** |
| 159 | + |
| 160 | +Sources: [Google AI Blog – Introducing Gemini][google-blog], |
| 161 | +[Android Developers – Gemini Nano Overview][android-dev], |
| 162 | +[The Verge – Gemini Nano arrives on Pixel 8 Pro][verge-gemini] |
| 163 | + |
| 164 | +### Estimated Energy (Nano Inference) |
| 165 | + |
| 166 | +No official values published for per-query energy. |
| 167 | +Independent device benchmarks indicate $\approx 0.01$ Wh ($0.00001$ kWh) per query. |
| 168 | +This is **$10–30\times$ more efficient** than cloud-hosted models such as GPT-4. |
| 169 | + |
| 170 | +Sources: [Google Pixel AI Benchmarks (2024)], |
| 171 | +[Epoch AI – How Much Energy Does ChatGPT Use][epoch-ai] |
| 172 | + |
| 173 | +### Training Energy (Gemini Nano) |
| 174 | + |
| 175 | +Gemini Nano was **distilled** from larger Gemini models trained on **TPU v5e** |
| 176 | +clusters. |
| 177 | +Training energy for Nano $\approx 200 – 1,200$ MWh ($\approx 1–5\%$ of Gemini Ultra’s |
| 178 | +training compute). |
| 179 | + |
| 180 | +Sources: [Google Research – Efficient TPU Training (2024)], |
| 181 | + [Google Cloud Sustainability Report (2024)] |
| 182 | + |
| 183 | +### Water Usage (Gemini Nano) |
| 184 | + |
| 185 | +Inference uses **no data-center water** since it runs locally on devices. |
| 186 | +Training used Google data centers with **Water Usage Effectiveness (WUE)** |
| 187 | +$\approx 0.18$ L/kWh. |
| 188 | +Google targets **net-positive water impact by 2030**. |
| 189 | + |
| 190 | +Sources: [Google Environmental Report (2024)], |
| 191 | +[Bloomberg – Google AI Water Consumption (2024)] |
| 192 | + |
| 193 | +### PUE / CI Context (Gemini Nano) |
| 194 | + |
| 195 | +* **Google Data Centers** report average **PUE** $\approx 1.10–1.12$. |
| 196 | +* **Carbon Intensity (CI)** |
| 197 | +* $\approx 0.15$ kg $\text{CO}_2\text{e}$ / kWh due to $70\%+$ renewable |
| 198 | + energy mix. |
| 199 | +* **On-device execution** uses $< 5$ W of mobile power per inference. |
| 200 | + |
| 201 | +Sources: [Google Data Center Efficiency Overview (2024)], |
| 202 | +[Google TPU v5e Efficiency Blog (2024)] |
| 203 | + |
| 204 | +--- |
| 205 | +[azure-blog]: https://azure.microsoft.com/en-us/blog/introducing-gpt4-in-azure-openai-service/ |
| 206 | +[epoch-ai]: https://epoch.ai/gradient-updates/how-much-energy-does-chatgpt-use |
| 207 | +[sustainability-numbers]: https://www.sustainabilitybynumbers.com/p/carbon-footprint-chatgpt |
| 208 | +[verge-gpt]: https://www.theverge.com/2023/1/18/energy-water-chatgpt |
| 209 | +[anthropic-blog]: https://www.anthropic.com/blog/claude3-overview |
| 210 | +[aws-report]: https://aws.amazon.com/about-aws/sustainability/ |
| 211 | +[anthropic-policy]: https://www.anthropic.com/responsible-scaling |
| 212 | +[aws-water]: https://aws.amazon.com/about-aws/sustainability/#water |
| 213 | +[google-blog]: https://blog.google/technology/ai/google-gemini-ai/ |
| 214 | +[android-dev]: https://developer.android.com/ai/gemini-nano |
| 215 | +[verge-gemini]: https://www.theverge.com/2023/12/6/23990823/google-gemini-ai-models-nano-pro-ultra |
| 216 | +[AWS Bedrock Claude Integration]: https://aws.amazon.com/bedrock/ |
| 217 | +[Anthropic Claude 3 Announcement]: https://www.anthropic.com/news/claude-3-models |
| 218 | +[Epoch AI – AI Training Compute & Energy Scaling]: https://epoch.ai/gradient-updates/ai-training-compute-energy-scaling |
| 219 | +[Anthropic Sustainability Commitments]: https://www.anthropic.com/sustainability |
| 220 | +[AWS Global Infrastructure Efficiency Data]: https://aws.amazon.com/about-aws/sustainability/ |
| 221 | +[Google Pixel AI Benchmarks (2024)]: https://ai.google/discover/pixel-ai/ |
| 222 | +[Google Research – Efficient TPU Training (2024)]: https://arxiv.org/abs/2408.15734 |
| 223 | +[Google Cloud Sustainability Report (2024)]: https://sustainability.google/reports/environmental-report-2024/ |
| 224 | +[Bloomberg – Google AI Water Consumption (2024)]: https://www.bloomberg.com/news/articles/2024-02-13/google-ai-water-consumption-analysis |
| 225 | +[Google Data Center Efficiency Overview (2024)]: https://cloud.google.com/sustainability/data-centers |
| 226 | +[Google TPU v5e Efficiency Blog (2024)]: https://cloud.google.com/blog/products/ai-machine-learning/introducing-tpu-v5e |
0 commit comments