Skip to content

Commit c22ace6

Browse files
committed
RSI
1 parent 26fd366 commit c22ace6

File tree

11 files changed

+2312
-1587
lines changed

11 files changed

+2312
-1587
lines changed

_freeze/posts/2025-09-13-recursive-self-improvement-explosion-optimization/execute-results/html.json

Lines changed: 2 additions & 2 deletions
Large diffs are not rendered by default.

docs/index.html

Lines changed: 622 additions & 619 deletions
Large diffs are not rendered by default.

docs/index.xml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
<link>tecunningham.github.io/</link>
1010
<atom:link href="tecunningham.github.io/index.xml" rel="self" type="application/rss+xml"/>
1111
<description>{{&lt; meta description-meta &gt;}}</description>
12-
<generator>quarto-1.5.57</generator>
12+
<generator>quarto-1.8.25</generator>
1313
<lastBuildDate>Sat, 07 Mar 2026 08:00:00 GMT</lastBuildDate>
1414
<item>
1515
<title>When You Overtake More Runners than You’re Overtaken by</title>

docs/listings.json

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -3,6 +3,7 @@
33
"listing": "/index.html",
44
"items": [
55
"/posts/2024-10-27-from-citlali.html",
6+
"/posts/2026-03-19-heebie-jeebies.html",
67
"/posts/2023-08-13-runner-percentile.html",
78
"/posts/2026-01-29-knowledge-creating-llms.html",
89
"/posts/2025-12-30-llm-verification.html",

docs/posts/2025-09-13-recursive-self-improvement-explosion-optimization-offcuts.html

Lines changed: 363 additions & 376 deletions
Large diffs are not rendered by default.

docs/posts/2025-09-13-recursive-self-improvement-explosion-optimization.html

Lines changed: 453 additions & 463 deletions
Large diffs are not rendered by default.

docs/posts/2026-03-19-heebie-jeebies.html

Lines changed: 726 additions & 0 deletions
Large diffs are not rendered by default.

docs/search.json

Lines changed: 84 additions & 91 deletions
Large diffs are not rendered by default.

docs/sitemap.xml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2,14 +2,14 @@
22
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9">
33
<url>
44
<loc>tecunningham.github.io/index.html</loc>
5-
<lastmod>2023-07-26T14:50:16.883Z</lastmod>
5+
<lastmod>2025-10-29T15:46:32.992Z</lastmod>
66
</url>
77
<url>
88
<loc>tecunningham.github.io/posts/2023-01-31-social-media-suspensions-data.html</loc>
99
<lastmod>2025-03-10T06:05:27.346Z</lastmod>
1010
</url>
1111
<url>
12-
<loc>tecunningham.github.io/posts/2025-09-13-recursive-self-improvement-explosion-optimization.html</loc>
13-
<lastmod>2026-03-21T12:33:54.246Z</lastmod>
12+
<loc>tecunningham.github.io/posts/2026-03-19-heebie-jeebies.html</loc>
13+
<lastmod>2026-03-20T03:28:02.301Z</lastmod>
1414
</url>
1515
</urlset>

posts/2025-09-13-recursive-self-improvement-explosion-optimization.qmd

Lines changed: 50 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -18,11 +18,49 @@ execute:
1818

1919
# Summary
2020

21+
22+
Capabilities are growing rapidly, but they're hard to quantify.
23+
: There's no consensus for a *scale* in AI capabilities. We do see fairly steady growth across many metrics: time horizon, average benchmark scores (ECI), effective compute, and predictive loss.
24+
25+
Algorithmic progress seems to be around 4X/year.
26+
: We can quantify algorithmic progress with compute efficiency, i.e. the reduction in cost required to reach a given capability score.
27+
28+
AI researchers have been making a very consistent series of discoveries, typically estimated at increasing compute efficiency by around 4X/year (with many qualifications, discussed below).
29+
30+
The 3X/year increase in algorithmic efficiency seems to be coming from a roughly 2X/year increase in researchers.
31+
32+
AI speedups will cause a loop, but unclear how strong.
33+
:
34+
Historically algorithmic progress has been .
35+
$$
36+
\xymatrix@C=1.4em@R=1.4em{
37+
& *++[F]{R\&D_t}\ar[d] & *++[F]{R\&D_{t+1}}\ar[d] & *++[F]{R\&D_{t+2}}\ar[d] \\
38+
\ar[r]& *++[F]{Algorithms_t}\ar[r] & *++[F]{Algorithms_{t+1}}\ar[r]\ar[ur]|{??} & *++[F]{Algorithms_{t+2}}
39+
}
40+
$$
41+
42+
AI systems are now contributing to algorithmic progress.
43+
: Until recently most AI R&D was done without help by LLMs, but we now see evidence for two channels:
44+
45+
1. _Augmenting AI researchers:_ AI researchers self-report big efficiency gains, e.g. @anthropic2025claude_work self-report approximately 50% productivity gains.
46+
2. _Automating AI research:_ E.g. AlphaEvolve, TTT-Discover, autoresearch.
47+
48+
Both of these effects are hard to measure, & we have a great deal of uncertainty.
49+
50+
51+
52+
53+
54+
55+
56+
57+
# Summary (OLD)
58+
2159
1. **Baseline model:**
60+
- Frontier model capability is growing at 9X/year (measured in effective compute)
2261
- Frontier training compute is growing at 3X/year
2362
- Algorithmic efficiency is growing at 3X/year
24-
- Frontier model capability is growing at 9X/year (measured in effective compute)
25-
- R&D staff growing at 2X/year.
63+
- R&D staff is growing at 2X/year.
2664

2765
2. **Two ways we can get RSI:** (1) augmentation of AI R&D; (2) automation of AI R&D.
2866

@@ -120,17 +158,14 @@ Q: where is AI likely to help?
120158

121159

122160

123-
# Data
161+
# Data on Compute Growth
124162

125-
My best estimates.
126-
127-
1. Training compute expenditure ($) has been growing around 3X/year, but will slow to 1.1X/year over 2026-2030.
128-
2. Training compute (FLOP) has been growing around 4X/year, but will
129-
3. Algorithmic efficiency has been growing around 3X/year, not clear if it will slow down or accelerate.
130-
131-
Takeaway: the outside-view seems roughly like `4-5X/year` for frontier training compute, `~3.5X/year` for training cost, `~2.3X/year` for installed compute stock, and `~1.37X/year` for hardware price-performance.
163+
Best estimates.
164+
:
165+
1. Training compute expenditure has been growing around 3X/year, but will gradually fall to 1.1X/year over 2026-2030.
166+
2. Training compute (FLOP) has been growing around 4X/year, but will fall to around 1.5X/year.
167+
3. Algorithmic efficiency has been growing around 3X/year, it is hard to forecast future trends.
132168

133-
### Training compute growth
134169

135170
| source | scope | growth | years | quantity |
136171
| ------------------------------------- | ------------------------------------- | ---------- | ----------------- | -------------------------- |
@@ -223,9 +258,6 @@ Takeaway: the outside-view seems roughly like `4-5X/year` for frontier training
223258

224259
The interactive graph currently shows about 4.6x/year growth in FLOPs of notable models, over 2020 - July 2025 (the latest datapoint).
225260

226-
@epoch2026canaicompaniesprofitable "Can AI companies become profitable?"
227-
:
228-
Useful mainly as a reminder that frontier-model economics are not just final training compute.
229261

230262
@you2025openaicomputespend "Most of OpenAI's 2024 compute went to experiments"
231263
:
@@ -234,12 +266,12 @@ Takeaway: the outside-view seems roughly like `4-5X/year` for frontier training
234266
- only a minority of R&D compute appears to have gone to the final training runs of released models
235267
- GPT-4.5 final training run was only a modest share of the total R&D bucket
236268

237-
This seems relevant to the RSI story: training runs are scaling fast, but frontier labs are also spending enormous amounts on experiments, unreleased models, and inference.
238-
239269

240-
## Estimates of LLM Algorithmic Progress
270+
# Data on Algorithmic Progress
241271

242-
These estimates are also not all measuring the same thing. Some are narrow pre-training efficiency estimates; others are all-in software-progress estimates; others are small-scale case studies.
272+
Best estimates.
273+
:
274+
- Around 4X/year, including the entire stack (GPU, pretraining, posttraining, elicitation).
243275

244276
| source | scope | progress | years | quantity |
245277
| --------------------------------------- | --------------------------------------------------------- | ------------ | -------------- | --------------------------------------------------------- |

0 commit comments

Comments
 (0)