You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/posts/2026-03-13-apple-picking-ai.html
+17-18Lines changed: 17 additions & 18 deletions
Original file line number
Diff line number
Diff line change
@@ -263,11 +263,7 @@ <h1 class="title">An Apple-Picking Model of Agents</h1>
263
263
<dt>A simple model for AI R&D.</dt>
264
264
<dd>
265
265
<p>An agent helping you to optimize an algorithm is like a robot helping you pick apples. It will take care of all the apples up to a certain height, and it may find apples you haven’t found yet, but there will still be apples out of its reach.</p>
266
-
<p>The motivation for writing this model was to help think through the implications of recent evidence that AI can push forward the frontier on various optimization and AI R&D problems (see my earlier post on <ahref="https://tecunningham.github.io/posts/2026-01-29-knowledge-creating-llms.html">AI knowledge creation</a>). If you can spend $100 in tokens to increase the efficiency of an AI training algorithm by 0.1% then, on its surface, this looks like the path to self-improvement, and you can replace humans with AI. But realistically the agents have been discovering <em>shallow</em> improvements to algorithms. This apple-picking model is my attempt to help think through the distinction, and figure out how to measure agents’ optimization ability.</p>
267
266
<p>Below I give a formal model but the basic ideas can all be seen with a drawing: here both the human and robot have picked four apples, but they’ve left the tree in a very different state, so the robot isn’t ready to replace the human yet:</p>
268
-
</dd>
269
-
</dl>
270
-
<p> </p>
271
267
<divclass="cell" data-layout-align="center">
272
268
<divclass="cell-output-display">
273
269
<divclass="quarto-figure quarto-figure-center">
@@ -277,37 +273,40 @@ <h1 class="title">An Apple-Picking Model of Agents</h1>
277
273
</div>
278
274
</div>
279
275
</div>
276
+
<p>The motivation for writing this model was to help think through the implications of recent evidence that AI can push forward the frontier on various optimization and AI R&D problems (see my earlier post on <ahref="https://tecunningham.github.io/posts/2026-01-29-knowledge-creating-llms.html">AI knowledge creation</a>). If you can spend $100 in tokens to increase the efficiency of an AI training algorithm by 0.1% then, on its surface, this looks like the path to self-improvement, and you can replace humans with AI. But realistically the agents have been discovering <em>shallow</em> improvements to algorithms. This apple-picking model is my attempt to help think through the distinction, and figure out how to measure agents’ optimization ability.</p>
277
+
</dd>
278
+
</dl>
280
279
<tableclass="caption-top table">
281
280
<colgroup>
282
281
<colstyle="width: 50%">
283
282
<colstyle="width: 50%">
284
283
</colgroup>
285
284
<thead>
286
285
<trclass="header">
287
-
<th>Agents finding optimizations</th>
288
286
<th>Robots picking apples</th>
287
+
<th>Agents finding optimizations</th>
289
288
</tr>
290
289
</thead>
291
290
<tbody>
292
291
<trclass="odd">
293
-
<td>Agents can autonomously advance the state-of-the-art on an optimization problem</td>
294
292
<td>Robots can find low apples that humans have not picked yet</td>
293
+
<td>Agents can autonomously advance the state-of-the-art on optimization problems</td>
295
294
</tr>
296
295
<trclass="even">
297
-
<td>Agents will asymptote to a lower maximum value</td>
298
296
<td>Robots will not be able to pick all the apples humans can</td>
297
+
<td>Agents will asymptote to a lower maximum value</td>
299
298
</tr>
300
299
<trclass="odd">
301
-
<td>Agents will have greater relative value for problems that are not yet heavily optimized</td>
302
300
<td>Robots are more useful for trees that have never been picked</td>
301
+
<td>Agents will have greater relative value for problems that are not yet heavily optimized</td>
303
302
</tr>
304
303
<trclass="even">
305
-
<td>An algorithm will be more optimized if agents are started after some rounds of human optimization</td>
306
304
<td>A tree will yield more apples if it’s harvested by both a human and robots</td>
305
+
<td>An algorithm will be more optimized if agents are started after some rounds of human optimization</td>
307
306
</tr>
308
307
<trclass="odd">
309
-
<td>To gauge the value of agents we want to test for the maximum <em>depth</em> of optimization they can do</td>
310
308
<td>To gauge the ability of robots we want to measure the highest apple they can reach</td>
309
+
<td>To gauge the value of agents we want to test for the maximum <em>depth</em> of optimization they can do</td>
311
310
</tr>
312
311
</tbody>
313
312
</table>
@@ -318,8 +317,8 @@ <h1 class="title">An Apple-Picking Model of Agents</h1>
318
317
</dd>
319
318
<dt>Relation to other models of AI R&D.</dt>
320
319
<dd>
321
-
<p>The critical distinction between this model and others is how we represent the state. Most existing models summarize the level of productivity (or stock of knowledge) with a single number, meaning there’s no distinction between a shallow and deep contribution. The model I’m using here allows us to represent the state with two numbers: the share of apples picked above and below <spanclass="math inline">\(\lambda\)</span>. For the RSI version of the model, we track <spanclass="math inline">\(\lambda\)</span> and the share of apples picked between <spanclass="math inline">\(\lambda\)</span> and 1.</p>
322
-
<p>Most existing models of recurse self-improvement assume that either AI accelerates or replaces human R&D researchers. I believe this is roughly true for <spanclass="citation" data-cites="aghion2019artificial">Aghion, Jones, and Jones (<ahref="#ref-aghion2019artificial" role="doc-biblioref">2019</a>)</span>, <spanclass="citation" data-cites="davidson2021could">Davidson (<ahref="#ref-davidson2021could" role="doc-biblioref">2021</a>)</span>, <spanclass="citation" data-cites="erdil2025gate">Erdil et al. (<ahref="#ref-erdil2025gate" role="doc-biblioref">2025</a>)</span>, <spanclass="citation" data-cites="davidson2026automatingairesearch">Davidson et al. (<ahref="#ref-davidson2026automatingairesearch" role="doc-biblioref">2026</a>)</span>, <spanclass="citation" data-cites="jones2025aird">Jones (<ahref="#ref-jones2025aird" role="doc-biblioref">2025</a>)</span>, <spanclass="citation" data-cites="kwa2026simpleraitimelines">Kwa (<ahref="#ref-kwa2026simpleraitimelines" role="doc-biblioref">2026</a>)</span>. These models then calibrate the effect through (1) how much does AI accelerate R&D workers; (2) how much do R&D workers contribute to our stock of knowledge.</p>
320
+
<p>The critical distinction between this model and others is how we represent the state. Most existing models summarize the level of productivity (or stock of knowledge) with a single number, meaning there’s no distinction between a shallow and deep contribution. The model I’m using here allows us to represent the state with two numbers: the share of apples picked above <spanclass="math inline">\(\lambda\)</span>, and the share below <spanclass="math inline">\(\lambda\)</span>. For the RSI version of the model we instead track the current level of capabilities (<spanclass="math inline">\(\lambda_n\)</span>) and the share of apples picked above that threshold.</p>
321
+
<p>Most existing models of recursive self-improvement assume that either AI accelerates or replaces human R&D researchers. I believe this is roughly true for <spanclass="citation" data-cites="aghion2019artificial">Aghion, Jones, and Jones (<ahref="#ref-aghion2019artificial" role="doc-biblioref">2019</a>)</span>, <spanclass="citation" data-cites="davidson2021could">Davidson (<ahref="#ref-davidson2021could" role="doc-biblioref">2021</a>)</span>, <spanclass="citation" data-cites="erdil2025gate">Erdil et al. (<ahref="#ref-erdil2025gate" role="doc-biblioref">2025</a>)</span>, <spanclass="citation" data-cites="davidson2026automatingairesearch">Davidson et al. (<ahref="#ref-davidson2026automatingairesearch" role="doc-biblioref">2026</a>)</span>, <spanclass="citation" data-cites="jones2025aird">Jones (<ahref="#ref-jones2025aird" role="doc-biblioref">2025</a>)</span>, <spanclass="citation" data-cites="kwa2026simpleraitimelines">Kwa (<ahref="#ref-kwa2026simpleraitimelines" role="doc-biblioref">2026</a>)</span>. These models then calibrate the effect through (1) how much does AI accelerate R&D workers; (2) how much do R&D workers contribute to our stock of knowledge.</p>
323
322
<p>However there is some awkwardness in fitting these models to the data:</p>
324
323
<oltype="1">
325
324
<li>It is hard to reconcile these models with evidence that AI is already autonomously contributing to AI research, yet still not replacing humans (i.e. autonomous agents are not perfect substitutes for humans).</li>
@@ -331,11 +330,11 @@ <h1 class="title">An Apple-Picking Model of Agents</h1>
331
330
<dd>
332
331
<ul>
333
332
<li><p><em>Landscape.</em> A more general version would model the entire <em>landscape</em>. You can represent an optimization problem as <spanclass="math inline">\(y=f(\bm{x})\)</span>, where you’re trying to choose an <spanclass="math inline">\(\bm{x}\)</span> to maximize <spanclass="math inline">\(y\)</span>, given some unknown <spanclass="math inline">\(f(\cdot)\)</span>. (talk about non-additivity of optimizations; talk about conditions under which landscape is separable, and so each subspace is an independent apple; talk about path dependence).</p></li>
334
-
<li><p><em>Relation to time horizon.</em> You can think of the high apples as long-time-horizon tasks.</p></li>
333
+
<li><p><em>Relation to time horizon.</em> You could think of the high apples as long-time-horizon tasks, i.e. tasks that AI agents are relatively worse at performing.</p></li>
335
334
<li><p><em>Shape of the tree.</em> You can extend the model such that apples are non-uniformly distributed, then we can replace <spanclass="math inline">\(\lambda\)</span> with <spanclass="math inline">\(F(\lambda)\)</span> below. We can then talk about types of domain which are bottom-heavy (most optimizations are pretty easy to find) vs top-heavy (most optimizations are hard to find). It then becomes important to know whether AI R&D is relatively more bottom-heavy or top-heavy, if the former then we might already be on the brink of an intelligence explosion.</p></li>
336
-
<li><p><em>Directed search.</em> We assume that the probability of finding an apple/optimization is independent of other apples already found.</p></li>
<li><p><em>Sketch of a quantitative model of LLM training.</em> LLM training is a big stack of algorithms, which we’ve been optimizing at perhaps 10X/year. [add some speculation about which parts of the stack have low-hanging fruit]</p></li>
335
+
<li><p><em>Directed search.</em> We assumed that the probability of finding an apple is independent of other apples already found. Realistically people have an ability to put direct their attention to finding new innovations. This implies lower diminishing returns to expenditure, and higher complementarity between agents and humans, I’m not sure whether it would change the qualitative conclusions of the model.</p></li>
336
+
<li><p><em>Bottlenecks.</em>Some R&D is bottlenecked not just by thinking (which agents can do), but also by running experiments.</p></li>
337
+
<li><p><em>Sketch of a quantitative model of LLM training.</em> LLM training is a big stack of algorithms, which we’ve been optimizing at perhaps 10X/year. Would be useful to add some speculation about which parts of the stack have low-hanging fruit.</p></li>
339
338
</ul>
340
339
</dd>
341
340
</dl>
@@ -346,11 +345,11 @@ <h1>Open Model</h1>
346
345
<dd>
347
346
<p>There is a continuum of apples spread uniformly on the real line.</p>
348
347
<p>A human can find apples over <spanclass="math inline">\([0,1]\)</span>, but an agent can only find apples over <spanclass="math inline">\([0,\lambda]\)</span>, with <spanclass="math inline">\(\lambda < 1\)</span> (at least for now).</p>
349
-
<p>Humans find apples at rate <spanclass="math inline">\(r_H\)</span>, agents find apples at rate <spanclass="math inline">\(r_A\)</span>, and we use <spanclass="math inline">\(t_H\)</span> and <spanclass="math inline">\(t_A\)</span> to represent the time humans and agents spend searching (you can also interpret <spanclass="math inline">\(t_H\)</span> and <spanclass="math inline">\(t_A\)</span> as expenditure on the problem).</p>
348
+
<p>Humans find apples at rate <spanclass="math inline">\(r_H\)</span>, agents find apples at rate <spanclass="math inline">\(r_A\)</span>, and we use <spanclass="math inline">\(t_H\)</span> and <spanclass="math inline">\(t_A\)</span> to represent the time humans and agents spend searching (you can also interpret <spanclass="math inline">\(t_H\)</span> and <spanclass="math inline">\(t_A\)</span> as monetary expenditure on the problem).</p>
350
349
</dd>
351
350
</dl>
352
351
<p><strong>We can then derive apples found:</strong></p>
353
-
<p><spanclass="math display">\[\text{share apples found}= \underbrace{\lambda(1-e^{-r_Ht_H-r_At_A})}_{\text{apples from bottom of tree}}+\underbrace{(1-\lambda)(1-e^{-r_Ht_H})}_{\text{apples from top of tree}}.\]</span></p>
352
+
<p><spanclass="math display">\[\text{apples picked}= \underbrace{\lambda(1-e^{-r_Ht_H-r_At_A})}_{\text{apples from bottom of tree}}+\underbrace{(1-\lambda)(1-e^{-r_Ht_H})}_{\text{apples from top of tree}}.\]</span></p>
354
353
<dl>
355
354
<dt>Implication: agents asymptote to a lower level than humans.</dt>
0 commit comments