Skip to content

Commit 15957b2

Browse files
authored
Merge pull request #25 from MIT-Emerging-Talent/gemma
Gemma
2 parents 6f07957 + 3851536 commit 15957b2

3 files changed

Lines changed: 48 additions & 35 deletions

File tree

commercial_models/README.md

Lines changed: 48 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -1,29 +1,51 @@
1-
# Commercial Models – Green AI Analysis
1+
# Commercial Models: The Crisis of AI Environmental Visibility
2+
3+
The environmental impact of Large Language Models (LLMs) has transitioned from
4+
an esoteric concern to a critical research priority. While early studies focused
5+
primarily on the energy cost of the massive **training phase** (e.g., GPT-4
6+
training estimates: $\approx 51.8 \text{ – } 62.3 \text{ GWh}$) [Source: Sustainability
7+
by Numbers], recent, rigorous analyses—such as the landmark benchmarking paper
8+
*"How Hungry is AI? Benchmarking Energy, Water, and Carbon Footprint of LLM
9+
Inference"* (Jegham et al., 2025)—have confirmed two major points:
10+
11+
* **Inference Dominance:** The energy consumed during **inference**
12+
* (running the model
13+
for every user query) is increasingly the dominant contributor to an LLM's total
14+
lifecycle environmental footprint due to global scale [Source: Google Cloud Blog].
15+
* **Divergent Reporting:** LLM energy reporting is highly unstable. As
16+
demonstrated by the benchmarking paper, resource consumption can vary by over
17+
**70 times** between the most and least efficient models for an equivalent task.
18+
Furthermore, consumption figures depend not just on model size, but profoundly
19+
on **deployment infrastructure** (PUE, regional carbon intensity) [Source: AWS
20+
Sustainability Report] and **methodology** (batch size, token length, hardware
21+
generation) [Source: Optimal Sparsity of MoE, arXiv]. This divergence leads to
22+
figures where one study reports less consumption (e.g., Epoch AI's $0.3 \text{ Wh}$
23+
per query) [Source: Epoch AI], while another (using different infrastructure
24+
assumptions) reports significantly more.
225

3-
This folder contains the **comparative sustainability analysis** of several
4-
large language models (LLMs) used in commercial applications.
5-
It is part of the **ELO2 Green AI project**, focusing on estimating
6-
the **energy, carbon, and water footprints** of each model.
7-
8-
---
9-
10-
## 📄 Contents
11-
12-
- **models.md** – Main document providing technical summaries and
13-
sustainability estimates for:
14-
- GPT-4 (OpenAI)
15-
- Claude 3 Haiku (Anthropic)
16-
- Gemini Nano (Google)
17-
1826
---
1927

20-
## 🎯 Purpose
21-
22-
This documentation:
23-
24-
- Highlights how **different LLM architectures and deployments**
25-
affect energy and water use.
26-
- Demonstrates how **model size and hosting** influence environmental impact.
27-
- Supports ongoing evaluation of **Green AI strategies** for efficient computing.
28-
29-
---
28+
## Our Research Intervention
29+
30+
This critical lack of standardized, reliable environmental data creates an
31+
**eco-efficiency paradox**: we cannot definitively prove which models are
32+
sustainable without standardized measurements, but the required performance
33+
trade-offs often push users toward the most power-hungry frontier models (e.g.,
34+
GPT-4's high capability, estimated at $0.3 \text{ – } 1.8 \text{ Wh}$ per query)
35+
[Source: The Verge].
36+
37+
This project addresses this paradox with an **interventionist approach**.
38+
Rather than
39+
simply measuring existing systems, our research question directly investigates
40+
whether we can *engineer* open-source models to definitively close the gap in the
41+
efficiency balance:
42+
43+
> **Research Question:** **Can open-source language models, enhanced with
44+
optimization techniques such as recursive editing and distillation, become
45+
environmentally and functionally viable alternatives to commercial models?**
46+
47+
We hypothesize that by applying resource-saving techniques (e.g., **distillation**
48+
to compress model size and **recursive editing** to refine output with minimal
49+
re-computation), we can achieve an **eco-efficiency score** that consistently
50+
demonstrates their viability as a sustainable alternative, driving adoption of
51+
open-source models over proprietary, resource-intensive solutions.
103 KB
Binary file not shown.

google_gemma/qa_log.txt

Lines changed: 0 additions & 9 deletions
This file was deleted.

0 commit comments

Comments
 (0)