You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/posts/2025-09-13-recursive-self-improvement-explosion-optimization.html
+30-13Lines changed: 30 additions & 13 deletions
Original file line number
Diff line number
Diff line change
@@ -271,18 +271,18 @@ <h1 class="title">Recursive Self-Improvement, Literature Review</h1>
271
271
<ul>
272
272
<li>Basic introduction to the problem</li>
273
273
<li>Survey of data: training expenditure (3X/year); training efficiency (4X/year); R&D effort (2X/year).</li>
274
-
<li>An overview of ~10 different models of RSI.</li>
274
+
<li>A single canonical model, & then an overview of ~10 different models of RSI</li>
275
275
</ul></li>
276
276
<li>TO DO:
277
277
<ul>
278
-
<li>A clearer distinction between two stages: (1) now->R&D automation; (2) automation->singularity.</li>
279
-
<li>A single canonical model upfront, & then the review can talk about departures from that model.</li>
280
278
<li>Survey of evidence on speedup & autonomous optimization ability.</li>
281
279
<li>Crisper statement of the balance of evidence.</li>
282
280
</ul></li>
283
281
</ol>
284
282
</div>
285
283
</div>
284
+
<sectionid="summary" class="level1">
285
+
<h1>Summary</h1>
286
286
<dl>
287
287
<dt>Summary.</dt>
288
288
<dd>
@@ -294,18 +294,34 @@ <h1 class="title">Recursive Self-Improvement, Literature Review</h1>
294
294
<p>We are now beginning to see signs of feedback, where AI is sufficiently capable to accelerate progress in R&D: <spanclass="math display">\[
295
295
\xymatrix@C=3em@R=1.4em{
296
296
*++[F]{\text{R\&D}}\ar[r]|(0.4)r
297
-
& *++[F]{\text{AI Capabilities}}\ar@{.>}@/_3em/[l]|{?}
297
+
& *++[F]{\text{AI Capabilities}}\ar@{.>}@/_3em/[l]|{a}
298
298
}
299
299
\]</span></p>
300
-
<p>We have reason to believe that capability progress is fairly sensitive to the quantity of R&D input. We can quantify the effect with <spanclass="math inline">\(r\)</span>, representing the . (<spanclass="math inline">\(r\approx 1\)</span>). However there is a great deal of uncertainty on how capabilities will affect R&D.</p>
301
-
<p>We can break the effects into two types, and make some very tentative guesses as of April 2026:<ahref="#fn1" class="footnote-ref" id="fnref1" role="doc-noteref"><sup>1</sup></a></p>
300
+
</dd>
301
+
<dt>In the pre-automation world we care about <spanclass="math inline">\(a\)</span>.</dt>
302
+
<dd>
303
+
<p>The discussion of recursive self-improvement breaks down into two very distinct questions depending on whether we have already achieved AI research automation, i.e. agents that are at least as good as human researchers.<ahref="#fn1" class="footnote-ref" id="fnref1" role="doc-noteref"><sup>1</sup></a></p>
304
+
<p>In the pre-automation world we are most interested in the degree to which AI accelerates human researchers, the variable <spanclass="math inline">\(a\)</span> in the diagram above. We can distinguish between augmentation of human researchers (“uplift”, “acceleration”), and partial-automation of research (autonomous research agents).</p>
305
+
<p>We can make a very rough guess at our current state, as of April 2026:<ahref="#fn2" class="footnote-ref" id="fnref2" role="doc-noteref"><sup>2</sup></a></p>
302
306
<oltype="1">
303
307
<li>Augmentation: agents are accelerating researcher productivity by 50%.</li>
304
-
<li>Automation: agents are able to make autonomous contributions equivalent to around 1 month of a researcher’s work.</li>
308
+
<li>Partial Automation: agents are able to make autonomous contributions equivalent to around 1 month of a researcher’s work.</li>
305
309
</ol>
306
-
<p>The critical uncertainty is how these effects <em>scale</em>. If R&D acceleration continues to scale with capabilities growth then we would expect an imminent substantial acceleration over the historical rate of capabilities growth.</p>
310
+
<p>Even if we confidently knew the levels of augmentation and partial-automation, it’s also important to know how these scale with AI capabilities progress.</p>
311
+
</dd>
312
+
<dt>In the post-automation world we care about <spanclass="math inline">\(r\)</span>.</dt>
313
+
<dd>
314
+
<p>In the post-automation world the effective stock of R&D labor will suddenly expand, and so there will clearly be a rapid advance in AI capabilities, the question is how quickly that will explode. Many formal models assume that <spanclass="math inline">\(a=1\)</span>, and switch their attention to <spanclass="math inline">\(r\)</span>. A justification of <spanclass="math inline">\(a=1\)</span> is, if we interpret AI capabilities as algorithmic efficiency, then a 1% increase in efficiency implies at least a 1% increase in effective R&D labor.</p>
315
+
<p>The critical questions for <spanclass="math inline">\(r\)</span> then become potential bottlenecks on R&D inputs:</p>
316
+
<oltype="1">
317
+
<li>The limited supply of inference compute for automated running researchers;</li>
318
+
<li>The limited supply of compute for running research experiments;</li>
319
+
<li>Logical limits on research progress, e.g. statistical ceilings on algorithmic efficiency.</li>
320
+
</ol>
321
+
<p>Most work on estimating post-automation effects doesn’t focus on measuring the abilites of AI, but instead measuring historical returns to R&D inputs, and the substitutability between different types of inputs.</p>
307
322
</dd>
308
323
</dl>
324
+
</section>
309
325
<sectionid="argument" class="level1">
310
326
<h1>Argument</h1>
311
327
<dl>
@@ -328,7 +344,7 @@ <h1>Argument</h1>
328
344
<dd>
329
345
<p>Until recently AI R&D was mostly done without significant help from AI, but we now see evidence for two channels:</p>
330
346
<oltype="1">
331
-
<li><em>Augmenting researchers:</em> AI researchers self-report big efficiency gains, e.g. <spanclass="citation" data-cites="anthropic2025claude_work">Anthropic (<ahref="#ref-anthropic2025claude_work" role="doc-biblioref">2025</a>)</span> self-report approximately 50% productivity gains, and <spanclass="citation" data-cites="anthropic2026risk">Anthropic (<ahref="#ref-anthropic2026risk" role="doc-biblioref">2026</a>)</span> estimate 100% productivity gains.<ahref="#fn2" class="footnote-ref" id="fnref2" role="doc-noteref"><sup>2</sup></a></li>
347
+
<li><em>Augmenting researchers:</em> AI researchers self-report big efficiency gains, e.g. <spanclass="citation" data-cites="anthropic2025claude_work">Anthropic (<ahref="#ref-anthropic2025claude_work" role="doc-biblioref">2025</a>)</span> self-report approximately 50% productivity gains, and <spanclass="citation" data-cites="anthropic2026risk">Anthropic (<ahref="#ref-anthropic2026risk" role="doc-biblioref">2026</a>)</span> estimate 100% productivity gains.<ahref="#fn3" class="footnote-ref" id="fnref3" role="doc-noteref"><sup>3</sup></a></li>
332
348
<li><em>Automating research:</em> Autonomous systems are making contributions to frontier R&D, e.g. AlphaEvolve, TTT-Discover, autoresearch.</li>
333
349
</ol>
334
350
<p>Both of these effects are hard to measure, & we have a great deal of uncertainty.</p>
<p>So if <spanclass="math inline">\(R\)</span> is constant, <spanclass="math inline">\(g_A\)</span> declines as <spanclass="math inline">\(A\)</span> rises.<ahref="#fn3" class="footnote-ref" id="fnref3" role="doc-noteref"><sup>3</sup></a> If <spanclass="math inline">\(R\)</span> grows at rate <spanclass="math inline">\(g_R\)</span> along a balanced growth path, then <spanclass="math inline">\(g_A = \frac{\lambda}{\beta}g_R.\)</span></p></li>
1062
+
<p>So if <spanclass="math inline">\(R\)</span> is constant, <spanclass="math inline">\(g_A\)</span> declines as <spanclass="math inline">\(A\)</span> rises.<ahref="#fn4" class="footnote-ref" id="fnref4" role="doc-noteref"><sup>4</sup></a> If <spanclass="math inline">\(R\)</span> grows at rate <spanclass="math inline">\(g_R\)</span> along a balanced growth path, then <spanclass="math inline">\(g_A = \frac{\lambda}{\beta}g_R.\)</span></p></li>
1047
1063
<li><p>Recursive self-improvement, where knowledge directly raises research input: <spanclass="math display">\[\begin{gathered}
<liid="fn1"><p>Ryan Greenblatt in April 2026 <ahref="https://www.lesswrong.com/posts/WjaGAA4xCAXeFpyWm/my-picture-of-the-present-in-ai">estimates</a> the speedup to engineering to be around 1.6x, and the autonomous capability to be around 5 hours (“the task duration at which AIs match a randomly selected AI company engineer (who is familiar with that part of the code base) is around 5 hours”). Note that a more robust statistic would be the number of hours.<ahref="#fnref1" class="footnote-back" role="doc-backlink">↩︎</a></p></li>
1843
-
<liid="fn2"><p>“Productivity uplift estimates ranged from 30% to 700%, with a mean of 152% and median of 100%.”<ahref="#fnref2" class="footnote-back" role="doc-backlink">↩︎</a></p></li>
1844
-
<liid="fn3"><p><spanclass="citation" data-cites="jones1995rd">Charles I. Jones (<ahref="#ref-jones1995rd" role="doc-biblioref">1995</a>)</span> introduced diminishing returns to knowledge, whereas Romer (1990) had assumed no diminishing returns to knowledge, <spanclass="math inline">\(\beta=0\)</span>.<ahref="#fnref3" class="footnote-back" role="doc-backlink">↩︎</a></p></li>
1858
+
<liid="fn1"><p><spanclass="citation" data-cites="davidson2025howquickandbigwo">Davidson and Houlden (<ahref="#ref-davidson2025howquickandbigwo" role="doc-biblioref">2025</a>)</span> calls this ASARA, “AI System for AI R&D Automation”, <spanclass="citation" data-cites="kokotajlo2025aifuturesmodel">Kokotajlo et al. (<ahref="#ref-kokotajlo2025aifuturesmodel" role="doc-biblioref">2025</a>)</span> calls this SAR, “superhuman AI researcher”.<ahref="#fnref1" class="footnote-back" role="doc-backlink">↩︎</a></p></li>
1859
+
<liid="fn2"><p>Ryan Greenblatt in April 2026 <ahref="https://www.lesswrong.com/posts/WjaGAA4xCAXeFpyWm/my-picture-of-the-present-in-ai">estimates</a> the speedup to engineering to be around 1.6x, and the autonomous capability to be around 5 hours (“the task duration at which AIs match a randomly selected AI company engineer (who is familiar with that part of the code base) is around 5 hours”). Note that a more robust statistic would be the number of hours.<ahref="#fnref2" class="footnote-back" role="doc-backlink">↩︎</a></p></li>
1860
+
<liid="fn3"><p>“Productivity uplift estimates ranged from 30% to 700%, with a mean of 152% and median of 100%.”<ahref="#fnref3" class="footnote-back" role="doc-backlink">↩︎</a></p></li>
1861
+
<liid="fn4"><p><spanclass="citation" data-cites="jones1995rd">Charles I. Jones (<ahref="#ref-jones1995rd" role="doc-biblioref">1995</a>)</span> introduced diminishing returns to knowledge, whereas Romer (1990) had assumed no diminishing returns to knowledge, <spanclass="math inline">\(\beta=0\)</span>.<ahref="#fnref4" class="footnote-back" role="doc-backlink">↩︎</a></p></li>
0 commit comments