<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Token for Token]]></title><description><![CDATA[language models and the future of AI]]></description><link>https://blog.jxmo.io</link><image><url>https://substackcdn.com/image/fetch/$s_!ySE4!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b4d4a83-97c1-455a-a39b-46563262fcdb_1024x1024.png</url><title>Token for Token</title><link>https://blog.jxmo.io</link></image><generator>Substack</generator><lastBuildDate>Tue, 28 Apr 2026 10:55:51 GMT</lastBuildDate><atom:link href="https://blog.jxmo.io/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Jack Morris]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[jxmnop@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[jxmnop@substack.com]]></itunes:email><itunes:name><![CDATA[Jack Morris]]></itunes:name></itunes:owner><itunes:author><![CDATA[Jack Morris]]></itunes:author><googleplay:owner><![CDATA[jxmnop@substack.com]]></googleplay:owner><googleplay:email><![CDATA[jxmnop@substack.com]]></googleplay:email><googleplay:author><![CDATA[Jack Morris]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[How to train the best embedding model in the world]]></title><description><![CDATA[one PhD later, I'm giving my secrets away for free]]></description><link>https://blog.jxmo.io/p/how-to-train-the-best-embedding-model</link><guid isPermaLink="false">https://blog.jxmo.io/p/how-to-train-the-best-embedding-model</guid><dc:creator><![CDATA[Jack Morris]]></dc:creator><pubDate>Mon, 09 Mar 2026 16:55:54 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/26a2ad08-3602-4db8-8e7c-8cb5745a821e_706x713.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>An unexpected side effect of graduate school was that I became the best person in the world at training text embedding models. Midway through my PhD, I trained a<a href="https://huggingface.co/jxm/cde-small-v2">state-of-the-art model</a> (using a still-unbeaten method) and I had a clear plan for how to make state-of-the-art models much better. </p><p>Sadly, the method turned out to be too expensive for an academic lab. Eventually I stopped working on these models.</p><p>Since I never got around to executing my grand vision, in this article, I&#8217;m just going to give the knowledge away for free. If you are training embedding models for your job, and want to make them better, you should use this method.</p><p><em>(Note: this post is more on the technical side. If you&#8217;re just here for my showerthoughts about AGI, you can skip it. But if you want to learn something, read on&#8230;)</em></p><h2><strong>A brief aside: how I became a world expert</strong></h2><p>This isn&#8217;t an exaggeration, by the way.</p><p><a href="https://arxiv.org/abs/2410.02525">Our model</a>, CDE, worked so well because of two different techniques we developed, both of which are somewhat difficult to implement. It beat all the other embedding models of its size in every category.</p><p>How did I become the world expert on this topic? First of all, it&#8217;s not that competitive. People don&#8217;t care about embedding models that much, and most people who do are in industry. OpenAI, Anthropic, Google, and Cursor all train internal embedding models, but mostly don&#8217;t publish; when they do, <a href="https://arxiv.org/html/2503.07891v1">they mostly just change the data</a>. </p><p>Within academia, embedding models aren&#8217;t that popular, and there aren&#8217;t many new ideas. I think what really worked for me was having a supportive advisor who forced me to go deep. I didn&#8217;t just look at how people were training these models (boring); I spent my time understanding the fundamental underlying techniques like <a href="https://www.jmlr.org/papers/volume13/gutmann12a/gutmann12a.pdf">noise contrastive estimation</a> and <a href="https://www.jmlr.org/papers/volume3/bengio03a/bengio03a.pdf">softmax approximations</a>. </p><h1><strong>The big idea</strong></h1><p>Now that I&#8217;ve established such impressive credentials, let me explain the big idea. </p><h2><strong>It all comes back to scaling</strong></h2><p>Why haven&#8217;t we scaled embedding models? </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!L9p3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!L9p3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png 424w, https://substackcdn.com/image/fetch/$s_!L9p3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png 848w, https://substackcdn.com/image/fetch/$s_!L9p3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!L9p3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!L9p3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:132023,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/190410320?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!L9p3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png 424w, https://substackcdn.com/image/fetch/$s_!L9p3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png 848w, https://substackcdn.com/image/fetch/$s_!L9p3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!L9p3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe9d59db-b4e6-45c5-84fc-f6f368a73b63_1800x1200.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The best LLMs get bigger every year, while embedding models haven't grown past the eight-billion parameter scale. (Thanks for the graph, chatGPT!)</figcaption></figure></div><p>The best LLMs get bigger every year, while embedding models haven&#8217;t grown past the eight-billion parameter scale. (Thanks for the graph, chatGPT!)</p><p>Language models have seen extreme success from training longer on more data with more parameters. Embedding models haven&#8217;t seen similar returns to investment: typically we see embedding models don&#8217;t improve much from training on larger document sizes. It depends on the setting, but performance tends to plateau on training sets around 10M document-document pairs. No one is training on billions of pairs (and if it worked, they surely would be).</p><p>So why is it that don&#8217;t embedding models scale? <em> </em>First, let&#8217;s see a brief overview of <em>contrastive learning</em>, the canonical training algorithm for these models.</p><h2><strong>How embedding models work</strong></h2><p>The training objective for embedding models often looks something like this:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\mathcal{L} = -\\frac{1}{N}\\sum_{i=1}^{N} \\log \\frac{\\exp \\left( f(q_i) \\cdot f(d_i) \\right) }{\\sum_{j=1}^{N} \\exp \\left( f(q_i) \\cdot f(d_j) \\right) }&quot;,&quot;id&quot;:&quot;LYWBLSUUZW&quot;}" data-component-name="LatexBlockToDOM"></div><p>here, f(q) is an embedding vector, and f(q) &#183; f(d) is the similarity between a document <em>d</em> and query <em>q</em>. The effect is to make query-document pairs close in embedding space.</p><p>Critically, this loss is applied to each <em>batch</em> of d&#8217;s and q&#8217;s. To train an embedding model, we sample a batch, apply this loss within the batch, and iterate.</p><p>If you go back to the basics, you&#8217;ll realize that this is essentially a form of <em>sampled softmax</em>, an old technique people used to use when they didn&#8217;t have enough GPU memory to compute a full softmax. It&#8217;s gone out of fashion, but what we <em>really  </em>want is to compute scores over everything in an entire dataset:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\mathcal{L} = -\\frac{1}{N}\\sum_{i=1}^{N} \\log \\frac{\\exp \\left( f(q_i) \\cdot f(d_i) \\right) }{\\sum_{j=1}^{N} \\exp \\left( f(q_i) \\cdot f(d_j) \\right) } \\approx -\\frac{1}{N}\\sum_{i=1}^{N} \\log \\frac{\\exp \\left( f(q_i) \\cdot f(d_i) \\right) }{\\sum_{d \\in \\mathcal{D}} \\exp \\left( f(q_i) \\cdot f(d) \\right)}&quot;,&quot;id&quot;:&quot;LQTLEOQVVT&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>Conventional embedding model-training optimizes the left equation. This is a Monte-Carlo approximation of the right quantity, which involves computing embedding similarities over the entire dataset.</p><p>(Note: Once, <a href="https://arxiv.org/abs/2210.11528">for another paper, I trained an embedding model where we *did* have perfect labels</a>, since it was matching up two views of the same person. In this case, I optimized the exact softmax via <a href="https://en.wikipedia.org/wiki/Coordinate_descent">coordinate ascent</a> and worked much better than NCE loss.)</p><h2><strong>The data problem</strong></h2><p>Ideally we would compute the quantity on the right &#8211; for each query, we would compare its embedding to the embedding of the proper document <em>as well as all other documents in the dataset</em>. If this worked perfectly, we could scale ad nauseam and produce better-and-better embedding models.</p><p>The reason naive scaling doesn&#8217;t work here is that as we reduce the approximate-ness of the softmax, we introduce a lot of noise. There are many <a href="https://arxiv.org/abs/2010.08191">papers</a> <a href="https://arxiv.org/abs/2402.16829">about</a> <a href="https://arxiv.org/abs/2301.02280">this</a> in both vision and text.</p><p>Let&#8217;s take an example query from the popular <a href="https://arxiv.org/abs/1611.09268">MSMarco search dataset</a>:</p><p><em>was ronald reagan a democrat?</em></p><p>This query corresponds to multiple labeled documents from Wikipedia that answer the question. And it&#8217;s clear that most documents on the web do not answer this question.</p><p>But as you scale the number of documents, you&#8217;re likely to run into more and more &#8220;collisions&#8221; that answer this question. And that means in our exact softmax above, you&#8217;ve introduced more noise into its gradient, since some of the terms in the denominator are incorrect. </p><p>In other words: say we scale our data and accidentally include a new document that mentions Ronald Reagan&#8217;s political affiliation. This document will automatically be marked as a non-answer to the pre-existing query, and we will incorrectly encourage the model to push the embedding of the new document away from the embedding of <em>was ronald reagan a democrat?</em></p><p>Over time, this type of noise compounds and scaling embedding data can actually <em>hurt</em>.</p><p>Incidentally, this is the exact problem tackled in many recent retrieval papers; in our CDE work, the reason why we were able to make contextual batching work at all is because we aggressively filtered these fake negatives within the batch. </p><p>In that setting, we observed nearly a 10% improvement by implementing a very crude embedding-based filtering mechanism:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!G3rv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!G3rv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg 424w, https://substackcdn.com/image/fetch/$s_!G3rv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg 848w, https://substackcdn.com/image/fetch/$s_!G3rv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!G3rv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!G3rv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg" width="1200" height="734" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:734,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:113493,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/190410320?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!G3rv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg 424w, https://substackcdn.com/image/fetch/$s_!G3rv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg 848w, https://substackcdn.com/image/fetch/$s_!G3rv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!G3rv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108ab51f-115d-414a-9891-3e4cee6f40dd_1200x734.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">In our CDE work, training with filtering gave us nearly a 10% improvement (compare left graph to right graph scale).</figcaption></figure></div><p>In CDE, training with filtering gave us nearly a 10% improvement (compare left graph to right graph scale).</p><h2><strong>LLMs to the rescue</strong></h2><p>The difficult realization here is that for D training query-document pairs, we technically need D^2 labels to compute the exact softmax. This is an intractable number of labels for human labeling pipeline: e.g. for a billion-pair corpus, we&#8217;d need 10^18 <strong>&#8776; one quintillion human labels</strong> to deduplicate the entire corpus. (We could reduce this number with some filtering + reranking, but this relies on another retrieval model, which perhaps defeats the purpose...)</p><p>Luckily, we can exploit a property that&#8217;s become common for RL-training, known as <a href="https://www.jasonwei.net/blog/asymmetry-of-verification-and-verifiers-law">verification asymmetry</a> or the Generator-Verifier gap. The main idea here is the same: <strong>regular LLMs are much, much smarter than embedding models</strong> (and more expensive!). If we could train an embedding model that agrees with the decision of an LLM for every possible query-document pair, that would be an enormous accomplishment.</p><p>So let&#8217;s forget human labels. We can use LLMs to label this kind of massive dataset. This way we can end up with a nearly <em>perfect</em> deduplicated dataset for training embedding models. (Using LLMs on every query-document pair also opens the door for training even better models using <a href="https://hturner.github.io/PlackettLuce/">more precise ranking algorithms</a>...)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nyCf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nyCf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg 424w, https://substackcdn.com/image/fetch/$s_!nyCf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg 848w, https://substackcdn.com/image/fetch/$s_!nyCf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!nyCf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nyCf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg" width="1199" height="629" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:629,&quot;width&quot;:1199,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:21517,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/190410320?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nyCf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg 424w, https://substackcdn.com/image/fetch/$s_!nyCf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg 848w, https://substackcdn.com/image/fetch/$s_!nyCf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!nyCf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e989e09-e1aa-4b40-aec2-be8b232b1052_1199x629.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Every experiment I ran looks something like this: embedding model training scales, but only when you properly filter for hard negatives.</figcaption></figure></div><p>Every experiment I ran looks something like this: embedding model training scales, but only when you properly filter for hard negatives.</p><h2><strong>The full procedure</strong></h2><p>I&#8217;d be remiss if I didn&#8217;t end by explaining *exactly* how to train this model. Here&#8217;s a short recipe for training a state-of-the-art embedding model, guaranteed to beat any existing embedder:</p><ul><li><p>Gather all the text pairs you can find.</p><p><a href="https://huggingface.co/datasets/nomic-ai/nomic-embed-unsupervised-data">This data</a></p><p> is a good start. Recent work has generated large amounts of synthetic data</p></li><li><p>For each query Q, filter the top-k using BM25 and embeddings. Gain some set of K documents {D_1, ..., D_K},  relevant to Q. (This is just a practical step to make labeling tractable.)</p></li><li><p>Run each pair (D_*, Q) through an LLM to figure out if D is truly a negative example or not.</p></li><li><p>Output: a perfectly-labeled retrieval dataset that&#8217;s arbitrarily scalable.</p></li><li><p>Train the model using exact softmax and coordinate ascent.</p></li></ul><p>If you have the money to run this &#8212; let me know how it goes! Happy to talk over DM or email. </p>]]></content:encoded></item><item><title><![CDATA['AI' just means LLMs now]]></title><description><![CDATA[There used to be many possible candidates for how humans might build intelligent systems. Now there's only one.]]></description><link>https://blog.jxmo.io/p/ai-just-means-llms-now-superintelligence</link><guid isPermaLink="false">https://blog.jxmo.io/p/ai-just-means-llms-now-superintelligence</guid><dc:creator><![CDATA[Jack Morris]]></dc:creator><pubDate>Sat, 02 Aug 2025 16:57:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!QsG5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Since the industrial revolution, humans have dreamed of building intelligent machines.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading <em>Token for Token</em>! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>In the 1800s, Charles Babbage and Ada Lovelace built <a href="https://en.wikipedia.org/wiki/Analytical_engine">the first general-purpose computer</a> and designed the first ever computer algorithm. They also speculated about whether computers might be <em>creative</em> one day and compose their own music or art.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QsG5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QsG5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!QsG5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!QsG5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!QsG5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QsG5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png" width="456" height="304.1043956043956" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:456,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!QsG5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!QsG5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!QsG5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!QsG5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9478342f-3577-480b-98e3-f43ac01a1579_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Fast forward to the twentieth century and lots of people started thinking about what it could look like to build intelligence artificially. The most famous early such thinker is Alan Turing, the father of the Turing Test and pioneer of early logic.</p><p>Even as AI spent several decades as a marginalized topic for scientific study, the idea of AI as science fiction blossomed. It&#8217;s now played a central role in movies (2001: A Space Odyssey, Blade Runner, The Terminator, The Matrix, Ex Machina) and TV (Star Trek, Knight Rider, The Twilight Zone, Westworld, Black Mirror). It&#8217;s safe to assume that anyone who&#8217;s watched Western media over the last fifty years knows what AI is.</p><h3>The Form of Superintelligence</h3><p>And through science fiction we&#8217;ve seen AI take many forms. I don&#8217;t think there&#8217;s a definition of <em>exactly</em> what an AI should look like. Perhaps in defining AI we can take a cue from the 1964 U.S. Supreme Court: &#8220;we&#8217;ll know it when we see it.&#8221;</p><p>Things get even murkier when we consider superintelligence. Are all these AIs <em>super</em>intelligent&#8211;smarter than humans? C-3PO certainly was; that guy knew everything. But many media portrayals of AI consider them as human-level, but not smarter.</p><p>For the purpose of this series of posts, I&#8217;m going to define &#8216;superintelligence&#8217; as whatever we saw in Iron Man (Jarvis) and Star Wars (C-3PO):</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6-Wr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6-Wr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png 424w, https://substackcdn.com/image/fetch/$s_!6-Wr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png 848w, https://substackcdn.com/image/fetch/$s_!6-Wr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png 1272w, https://substackcdn.com/image/fetch/$s_!6-Wr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6-Wr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png" width="1454" height="476" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:476,&quot;width&quot;:1454,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1134210,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/169856309?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6-Wr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png 424w, https://substackcdn.com/image/fetch/$s_!6-Wr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png 848w, https://substackcdn.com/image/fetch/$s_!6-Wr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png 1272w, https://substackcdn.com/image/fetch/$s_!6-Wr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ed52f7e-4190-40dd-ac1b-385486db3e43_1454x476.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>So we&#8217;ll take a reductionary position that superintelligence is simply a helpful machine that&#8217;s as capable and more knowledgeable than a human.</p><p>Do we have this yet? Unfortunately no. The newest, smartest models from OpenAI are very good&#8211; they can code like professional programmers and achieved gold medal on the International Math Olympiad&#8211; but they&#8217;re not considered yet human-level in many areas. (This skill profile is what some have described as <em>jagged intelligence</em>.)</p><h3>Biological Candidates for Superintelligence</h3><p>For a long time we didn&#8217;t know how to build JARVIS. We had no idea what superintelligence would look like. Many saw that some kind of connectionism might be the path, but didn&#8217;t know how to scale it.</p><p>In fact, lots of early AI research was inspired by the human brain. The most popular arguments, such as <a href="https://nickbostrom.com/superintelligence">Nick Bostrom&#8217;s original arguments for superintelligence</a>, all consider the fact that we already have proof that it&#8217;s possible to build (non-super) intelligence, if you consider the human brain as an existence proof. So you can make arguments like this one:</p><blockquote><p>The human brain contains about 10^11 neurons. Each neuron has about 5*10^3 synapses, and signals are transmitted along these synapses at an average frequency of about 10^2 Hz. Each signal contains, say, 5 bits. This equals 10^17 ops. (&#8230;)</p></blockquote><p>and its ultimate conclusion:</p><blockquote><p>Depending on degree of optimization assumed, human-level intelligence probably requires between 10^14 and 10^17 ops.</p></blockquote><p>Although I deeply admire the author&#8217;s conviction (and prescience, since these beliefs were published in 1998) it seems flawed to equate the operations performed by a human brain with simple &#8216;ops&#8217; (such as floating-point operations) that happen inside a computer. (I think this is the basic point that <a href="https://en.wikipedia.org/wiki/The_Emperor%27s_New_Mind">Roger Penrose made</a> when he hypothesized that quantum mechanics plays an important role in the development of human consciousness inside the brain.)</p><p>This is all worsened by the fact that the systems that ended up getting us closest to superintelligence are called <em>neural networks</em>, which is simultaneously an excellent name and a dreadful misnomer. They&#8217;re systems of interconnected dynamically-updatable components, but they&#8217;re also not related to biological neurons in any way.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!N2_X!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!N2_X!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png 424w, https://substackcdn.com/image/fetch/$s_!N2_X!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png 848w, https://substackcdn.com/image/fetch/$s_!N2_X!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png 1272w, https://substackcdn.com/image/fetch/$s_!N2_X!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!N2_X!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1980908,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/169856309?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!N2_X!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png 424w, https://substackcdn.com/image/fetch/$s_!N2_X!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png 848w, https://substackcdn.com/image/fetch/$s_!N2_X!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png 1272w, https://substackcdn.com/image/fetch/$s_!N2_X!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1b59ee1-9537-4fb0-ac82-3840f6ed76fd_1484x1060.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Anyway, we spent a long time making these types of biological analogies, and they didn&#8217;t get us far. What worked was at best very loosely inspired by the functionality of the human brain, machine learning and neural networks:</p><h3>Digital Candidates for Superintelligence</h3><p>Since 2011 (around when <a href="https://en.wikipedia.org/wiki/AlexNet">AlexNet was introduced</a>) every year has gotten us closer to building Jarvis or C-3PO. We developed better machine learning and neural networks, got good at training them to mimic human text, and now we&#8217;re making steady progress on incentivizing them to grow smarter than humans in some areas. This is really exceptional progress.</p><p>As mentioned, the best systems that we have right now are already very useful. They know how to code, give directions, and write recipes. They&#8217;re decent therapists and life coaches. They&#8217;re coming around in the creativity department, too, and writing better prose and poetry each year.</p><p>One might expect, then, that we as a field have <em>several</em> candidate systems for superintelligent AI. Perhaps we have a really good video simulator, a few companies have embodied continually-learning robots, and there&#8217;s a speech system out there that also exhibits Jarvis-like capabilities.</p><p>But this turns out not to be the case. There is only one existing technology that&#8217;s close to Jarvis: large language models. We haven&#8217;t built models that get smarter by exploring the world or watching lots of movies. We&#8217;ve only built models that get smarter by reading lots of text (say, all the words on the Internet) and then marginally smarter after doing a few thousand math problems.</p><p>All we have is language models.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Token for Token! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[All AI Models Might Be The Same]]></title><description><![CDATA[what can language model embeddings tell us about whales speech, and decoding ancient texts?&#160; &#160; &#160; (on The Platonic Representation Hypothesis and the idea of *universality* in AI models)]]></description><link>https://blog.jxmo.io/p/there-is-only-one-model</link><guid isPermaLink="false">https://blog.jxmo.io/p/there-is-only-one-model</guid><dc:creator><![CDATA[Jack Morris]]></dc:creator><pubDate>Thu, 17 Jul 2025 17:17:19 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/08727245-d2dc-4d17-84d9-258d335e27bb_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!h3NX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!h3NX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png 424w, https://substackcdn.com/image/fetch/$s_!h3NX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png 848w, https://substackcdn.com/image/fetch/$s_!h3NX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png 1272w, https://substackcdn.com/image/fetch/$s_!h3NX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!h3NX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png" width="487" height="432.3076171875" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:909,&quot;width&quot;:1024,&quot;resizeWidth&quot;:487,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Can We Learn To Talk With Whales? Introducing Project CETI - InDEPTH&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Can We Learn To Talk With Whales? Introducing Project CETI - InDEPTH" title="Can We Learn To Talk With Whales? Introducing Project CETI - InDEPTH" srcset="https://substackcdn.com/image/fetch/$s_!h3NX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png 424w, https://substackcdn.com/image/fetch/$s_!h3NX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png 848w, https://substackcdn.com/image/fetch/$s_!h3NX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png 1272w, https://substackcdn.com/image/fetch/$s_!h3NX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33aec54-92e7-410c-a964-1464aa5675d8_1024x909.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://www.projectceti.org/">Project CETI</a> is a large-scale effort to decode whale speech. If AI models do learn a universal language, we might be able to use it to talk to whales.</em></figcaption></figure></div><p>Growing up, I sometimes played a game with my friends called &#8220;Mussolini or Bread.&#8221;</p><p>It&#8217;s a guessing game, kind of like Twenty Questions. The funny name comes from the idea that, in the space of everything, &#8216;Mussolini&#8217; and &#8216;bread&#8217; are about as far away from each other as you can get.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Token for Token! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>One round might go like this:</p><ul><li><p>Is it closer to Mussolini or bread? <em>Mussolini.</em></p></li><li><p>Is it closer to Mussolini or David Beckham? <em>Uhh, I guess Mussolini.</em> (Ok, they&#8217;re definitely thinking of a person.)</p></li><li><p>Is it closer to Mussolini or Bill Clinton? <em>Bill Clinton</em>.</p></li><li><p>Is it closer to Bill Clinton or Pel&#233;? <em>Bill Clinton, I think</em>.</p></li><li><p>Is it closer to Bill Clinton or Grace Hopper? <em>Grace Hopper</em>.</p></li><li><p>Is it closer to Grace Hopper or Richard Hamming? <em>Richard Hamming</em>.</p></li><li><p>Is it closer to Richard Hamming or Claude Shannon? <em>You got it, I was thinking of Claude Shannon.</em></p></li></ul><p>Hopefully you get the point. By successively narrowing down the space of possible things or people, we&#8217;re able to guess almost anything.</p><p>How is this game possible? Mussolini or Bread only works because you and I have a shared sense of semantics. Before we played this game, we never talked about whether Claude Shannon is semantically &#8216;closer&#8217; to Mussolini or Beckham. We never even talked about what it means for two things to be &#8216;close&#8217;, even, or agreed on rules to the game.</p><p>As you might imagine, the edge cases in M or B can be controversial. But I&#8217;ve played this game with many people and people tend to &#8220;just get it&#8221; on their first try. How is that possible?</p><h3>A universal sense of semantics</h3><p>One explanation for why this game works is that <em>there is only one way in which things are related</em>, and this comes from the underlying world we live in. Put another way, our brains build up complicated models of the world in which we live, and the model of the world that my brain relies on is very similar to the one in yours. In fact, our brains&#8217; models of the world are so similar that we can narrow down almost any concept by successively refining the questions we ask, a-la Mussolini or Bread.</p><p>Let&#8217;s try to explain this through the lens of compression. One perspective on AI is that we&#8217;re just learning to compress all the data in the world. In fact, the task of language modeling (predicting the next word) can be seen as a compression task, ever since <a href="https://en.wikipedia.org/wiki/Shannon%27s_source_coding_theorem">Shannon&#8217;s source coding theorem</a> formalized the relationship between probability distributions and compression algorithms.</p><p>In recent years, we&#8217;ve developed much more accurate probability distributions of the world; this turned out to be easy, since <a href="https://arxiv.org/abs/2001.08361">bigger and bigger language models</a> give us <a href="https://arxiv.org/abs/1712.00409">better and better probability distributions</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!APsC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!APsC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png 424w, https://substackcdn.com/image/fetch/$s_!APsC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png 848w, https://substackcdn.com/image/fetch/$s_!APsC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png 1272w, https://substackcdn.com/image/fetch/$s_!APsC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!APsC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png" width="294" height="234.80090497737555" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:706,&quot;width&quot;:884,&quot;resizeWidth&quot;:294,&quot;bytes&quot;:85548,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168573586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!APsC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png 424w, https://substackcdn.com/image/fetch/$s_!APsC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png 848w, https://substackcdn.com/image/fetch/$s_!APsC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png 1272w, https://substackcdn.com/image/fetch/$s_!APsC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69374442-353c-4f39-9ec7-270034bb6b2b_884x706.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Intelligence is compression, and compression follows scaling laws. I like reminding people that the original work on scaling laws <a href="https://arxiv.org/abs/1712.00409">came from Baidu in 2017</a>.</figcaption></figure></div><p>And with better probability distributions comes better compression. In practice, we find that a model that can compress real data better knows more about the world. And thus there is a duality between compression and intelligence. Compression <em>is</em> intelligence. Some have even said <a href="https://www.youtube.com/watch?v=dO4TPJkeaaU">compression may be the way to AGI</a>. Ilya gave a <a href="https://www.lesswrong.com/posts/KqgujtM3vSAfZE2dR/on-ilya-sutskever-s-a-theory-of-unsupervised-learning#Kolmogorov_Complexity__20_48_">famously incomprehensible talk</a> about the connections between intelligence and compression.</p><p>Last year some folks at DeepMind wrote a paper simply titled <a href="https://arxiv.org/abs/2309.10668">Language Modeling Is Compression</a> and actually tested different language models&#8217; ability to compress various data modalities. Across the board, they found that smarter language models are better compressors. (Of course, this is what we&#8217;d expect, given the source coding theorem.)</p><p>And learning to compress is exactly how models end up generalizing. Some of <a href="https://arxiv.org/abs/2505.24832">our recent work</a> has analyzed models&#8217; compression behavior in the limit of training: we train models for infinitely long on datasets of varying size.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OJir!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OJir!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png 424w, https://substackcdn.com/image/fetch/$s_!OJir!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png 848w, https://substackcdn.com/image/fetch/$s_!OJir!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png 1272w, https://substackcdn.com/image/fetch/$s_!OJir!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OJir!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png" width="1456" height="743" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:743,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:636353,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168573586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OJir!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png 424w, https://substackcdn.com/image/fetch/$s_!OJir!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png 848w, https://substackcdn.com/image/fetch/$s_!OJir!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png 1272w, https://substackcdn.com/image/fetch/$s_!OJir!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5536a11-1f7f-45f9-96c6-01027a2c228a_1992x1016.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figures from our recent work, <em><a href="https://arxiv.org/abs/2505.24832">How much can language models memorize?</a></em> Generalization only begins when compression is no longer possible, since the model can&#8217;t store data points separately and is forced to combine things.</figcaption></figure></div><p>When a model can fit the training dataset perfectly (left side of both graphs) we see that it memorizes data really well, and totally fails to generalize. But when the dataset gets too big, and the model can no longer fit all of the data in its parameters, it&#8217;s forced to &#8220;combine&#8221; information from multiple datapoints in order to get the best training loss. This is where generalization occurs.</p><p>And the central idea I&#8217;ll push here is that when generalization occurs, <em>it usually occurs in the same way</em>, even within different models. From the compression perspective, under a given architecture and within a fixed number of parameters, <em>there is only one way to compress the data well</em>. This sounds like a crazy idea&#8211;and it is&#8211; but across different domains and models, there turns out to be a lot of evidence for this phenomenon.</p><h3>The Platonic Representation Hypothesis</h3><p>So how can different models be learning the shared representations? Given <a href="https://transformer-circuits.pub/2023/privileged-basis/index.html">the massive number of ~equivalent ways in which a model can represent things</a>, why should two models ever converge to analogous representations?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!D0oH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!D0oH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png 424w, https://substackcdn.com/image/fetch/$s_!D0oH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png 848w, https://substackcdn.com/image/fetch/$s_!D0oH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png 1272w, https://substackcdn.com/image/fetch/$s_!D0oH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!D0oH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png" width="338" height="465.2470588235294" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:936,&quot;width&quot;:680,&quot;resizeWidth&quot;:338,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!D0oH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png 424w, https://substackcdn.com/image/fetch/$s_!D0oH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png 848w, https://substackcdn.com/image/fetch/$s_!D0oH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png 1272w, https://substackcdn.com/image/fetch/$s_!D0oH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3753a05b-fda0-4084-8653-18ab0cba898a_680x936.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">A terse description and illustration of the headlining theory from <a href="https://arxiv.org/abs/2405.07987">The Platonic Representation Hypothesis</a> (2024).</figcaption></figure></div><p>Remember what these models are really doing is modeling the relationships between things in the world. In some sense there&#8217;s only one correct way to model things, and that&#8217;s the <em>true</em> model, the one that perfectly reflects the reality in which we live. Perhaps an infinitely large model with infinite training data would be a perfect simulator of the world itself.</p><p>As models have gotten bigger, their similarities have become more apparent. The theory that models are converging to a shared underlying representation space was formalized in <strong>The Platonic Representation Hypothesis,</strong> <a href="https://arxiv.org/abs/2405.07987">a position paper written by a group of MIT researchers in 2024</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ID7w!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ID7w!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png 424w, https://substackcdn.com/image/fetch/$s_!ID7w!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png 848w, https://substackcdn.com/image/fetch/$s_!ID7w!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png 1272w, https://substackcdn.com/image/fetch/$s_!ID7w!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ID7w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png" width="1456" height="833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:833,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!ID7w!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png 424w, https://substackcdn.com/image/fetch/$s_!ID7w!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png 848w, https://substackcdn.com/image/fetch/$s_!ID7w!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png 1272w, https://substackcdn.com/image/fetch/$s_!ID7w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ff78e39-5852-4c94-a1d5-7d7a5e1e4091_1496x856.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The Platonic Representation Hypothesis argues that as models get bigger, they&#8217;re learning more and more of the same features. They provide evidence for this in vision and language.</figcaption></figure></div><p>The Platonic Representation Hypothesis argues that models are converging to a shared representation space, and this is becoming more true as we make models bigger and smarter. This is true in text and language, at a minimum,</p><p>Remember <a href="https://situational-awareness.ai/">the trends in scaling</a> show that models are getting all three of bigger, smarter, and more efficient every year. That means that we can expect models to get <em>more similar</em>, too, as the years go on.</p><h3>A brief aside on embedding inversion</h3><p>The evidence for the Platonic Representation Hypothesis is compelling. But is it useful? Before I explain how to take advantage of the PRH, I have to give a bit of background on a problem of <a href="https://arxiv.org/abs/2310.06816">embedding inversion</a>.</p><p>I worked for a year or so of my PhD on this problem: given a representation vector from a neural network, can we infer what text was inputted to the network?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I8Gv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I8Gv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png 424w, https://substackcdn.com/image/fetch/$s_!I8Gv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png 848w, https://substackcdn.com/image/fetch/$s_!I8Gv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png 1272w, https://substackcdn.com/image/fetch/$s_!I8Gv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I8Gv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png" width="1456" height="844" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:844,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:707814,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168573586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I8Gv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png 424w, https://substackcdn.com/image/fetch/$s_!I8Gv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png 848w, https://substackcdn.com/image/fetch/$s_!I8Gv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png 1272w, https://substackcdn.com/image/fetch/$s_!I8Gv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aee165a-db0d-4c32-a3d1-5ce124be1b76_1722x998.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualization of a network that reconstructs images astonishingly well given only the 1000 class probability predictions from an ImageNet classifier (from <em><a href="https://arxiv.org/abs/2103.07470">Understanding Invariance via Feedforward Inversion of Discriminatively Trained Classifiers)</a></em>.</figcaption></figure></div><p>We thought inversion should be possible because <a href="https://arxiv.org/abs/2103.07470">results on ImageNet</a> showed that they could do very effective reconstruction given only a model&#8217;s output of 1000 class probabilities. This is extremely unintuitive. Apparently knowing that an image is 0.0001% parakeet and 0.0017% baboon is useful enough to infer not only the true class but lots of irrelevant information like facial structure, pose, and background details.</p><p>In the realm of text, the problem looks easy on its face, because typical embedding vectors have ~1000 floating-point numbers in them, or around 16 KB of data. If you store 16KB of text, it can represent quite a lot. Since we were working with datapoints on the level of long sentences or short documents, it seemed reasonable that we would be able to do inversion quite well.</p><p>But it turns out to be really hard. This mostly comes about because embeddings are in some sense extremely compressed: since similar texts have similar embeddings, it becomes very difficult to distinguish between two embeddings that represent similar-but-different data. So our models could output something <em>close</em> to the embedding, but almost never the exactly-correct text.</p><p>We ended up getting around this problem by using a primitive form of test-time compute: we made many queries to the embedding space and built a model that could &#8220;narrow down&#8221; the true text by iteratively improving itself in embedding space. Our system looks kind of like a learned optimizer that takes text-based steps to move position in embedding space.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H7Gz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H7Gz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png 424w, https://substackcdn.com/image/fetch/$s_!H7Gz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png 848w, https://substackcdn.com/image/fetch/$s_!H7Gz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png 1272w, https://substackcdn.com/image/fetch/$s_!H7Gz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H7Gz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png" width="1456" height="657" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:657,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:459806,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168573586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H7Gz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png 424w, https://substackcdn.com/image/fetch/$s_!H7Gz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png 848w, https://substackcdn.com/image/fetch/$s_!H7Gz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png 1272w, https://substackcdn.com/image/fetch/$s_!H7Gz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73b1066d-c14f-4fdf-86e4-9c8c40c2a616_1834x828.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Iterative refinement is an extremely effective method for embedding inversion (<a href="https://thegradient.pub/text-embedding-inversion/">read more here</a>).</figcaption></figure></div><p>This new approach turns out to work very well. Given an embedding model, we were able to invert text at the level of a long sentence with 94% exact accuracy.</p><h3>Harnessing Plato for embedding inversion</h3><p>We were very pleased with ourselves after making that method work. This had a whole lot of implications for the new model of the <a href="https://en.wikipedia.org/wiki/Vector_database">vector database</a>: sharing vectors, apparently, is equivalent to sharing the text those vectors represent.</p><p>But unfortunately our method was embedding-specific. It wasn&#8217;t clear that it could transfer to future embedding models or private fine-tunes that we didn&#8217;t have access to. And it required making a lot of queries to the embedding model we knew: training the models took millions of embeddings.</p><p>We thought that this shouldn&#8217;t be the case. If the Platonic Representation Hypothesis is true, and different models (in some sense) are learning the same thing, we should be able to build one <em>universal</em> embedding inverter and use it for any kind of model. This idea set us off on a multi-year quest to &#8220;harness&#8221; the PRH and build a universal embedding inverter.</p><p>We started by expressing our problem as a mathematical one. Given a bunch of embeddings from model A, and a bunch of embeddings from model B, can we learn to map from A&#8594;B (or B&#8594;A)?</p><p>Importantly, we don&#8217;t have any <em>correspondence</em>, i.e. pairs of texts with representations in both A and B. That&#8217;s why this problem is hard. We want to learn to align the spaces of A and B in some way so that we can &#8216;magically&#8217; learn how to convert between their spaces.</p><p>We realized after a while that this problem has been solved at least once in the deep learning world: work on a model called CycleGAN proposed a way to translate between spaces without correspondence using a method called cycle consistency:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2EH7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2EH7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png 424w, https://substackcdn.com/image/fetch/$s_!2EH7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png 848w, https://substackcdn.com/image/fetch/$s_!2EH7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png 1272w, https://substackcdn.com/image/fetch/$s_!2EH7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2EH7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png" width="661" height="318.24244505494505" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:701,&quot;width&quot;:1456,&quot;resizeWidth&quot;:661,&quot;bytes&quot;:2387580,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168573586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2EH7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png 424w, https://substackcdn.com/image/fetch/$s_!2EH7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png 848w, https://substackcdn.com/image/fetch/$s_!2EH7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png 1272w, https://substackcdn.com/image/fetch/$s_!2EH7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62dab6f3-1bfc-4f2f-b199-e5f3e2443866_1866x898.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Unpaired image translations from <em><a href="https://arxiv.org/abs/1703.10593">Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks</a> </em>(2017). If you squint, you might see extremely preliminary evidence for the Platonic Representation Hypothesis.</figcaption></figure></div><p>Just imagine that the horses and zebras above are a piece of text from model A being translated into the space of model B and back. If this works for zebras and horses, why shouldn&#8217;t it work for text?</p><p>And, after at least a year of ruthlessly debugging our own embedding-specific version of CycleGAN, we started to see signs of life. In our unsupervised matching task we started to produce GIFs like this:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!btgP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!btgP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif 424w, https://substackcdn.com/image/fetch/$s_!btgP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif 848w, https://substackcdn.com/image/fetch/$s_!btgP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif 1272w, https://substackcdn.com/image/fetch/$s_!btgP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!btgP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif" width="404" height="353.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:700,&quot;width&quot;:800,&quot;resizeWidth&quot;:404,&quot;bytes&quot;:2008705,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/gif&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168573586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!btgP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif 424w, https://substackcdn.com/image/fetch/$s_!btgP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif 848w, https://substackcdn.com/image/fetch/$s_!btgP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif 1272w, https://substackcdn.com/image/fetch/$s_!btgP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcebf3891-44d8-4ebb-9fb0-c9987169c752_800x700.gif 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">After training a CycleGAN-like model for mapping between embedding spaces, <a href="https://vec2vec.github.io/">vec2vec</a> learns to &#8216;magically&#8217; align them. Hooray for the Platonic Representation Hypothesis!</figcaption></figure></div><p>To us, this was an incredible step forward, and proof for an even stronger claim we call the &#8220;Strong Platonic Representation Hypothesis&#8221;. Models&#8217; representations share so much structure that <em>we can translate between them</em>, even without having knowledge of individual points in either of the spaces. This meant that we could do unsupervised conversion between models, as well as invert embeddings mined from databases where we know nothing about the underlying model.</p><h3>Universality in Circuits</h3><p>Some additional evidence for the PRH comes from the world of <em>mechanistic interpretability</em>, where researchers attempt to reverse-engineer the inner workings of models. Work on <a href="https://distill.pub/2020/circuits/zoom-in/#claim-3">Circuits</a> in 2020 found very similar functionalities in very different models:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lo8W!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lo8W!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png 424w, https://substackcdn.com/image/fetch/$s_!lo8W!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png 848w, https://substackcdn.com/image/fetch/$s_!lo8W!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png 1272w, https://substackcdn.com/image/fetch/$s_!lo8W!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lo8W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png" width="1456" height="749" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:749,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2380808,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168573586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lo8W!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png 424w, https://substackcdn.com/image/fetch/$s_!lo8W!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png 848w, https://substackcdn.com/image/fetch/$s_!lo8W!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png 1272w, https://substackcdn.com/image/fetch/$s_!lo8W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37c23ccf-6f08-4fe6-8cdd-ee97170fb245_1994x1026.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Universal feature dectors from <em><a href="https://distill.pub/2020/circuits/zoom-in/#claim-3">Circuits</a></em><a href="https://distill.pub/2020/circuits/zoom-in/#claim-3"> (2020</a>). Different networks exhibit remarkably similar behaviors.</figcaption></figure></div><p>More recently, there&#8217;s been some action around a method for feature discretization known as sparse autoencoders (SAEs). SAEs take a bunch of embeddings and learn a dictionary of interpretable features that can reproduce those embeddings with minimal loss.</p><p>Many are observing that if you train SAEs on two different models, they often learn many of the same features. There&#8217;s even been some recent work on &#8216;unsupervised concept discovery&#8217;, a suite of methods that can compare two SAEs to find feature overlap:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!s35k!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!s35k!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png 424w, https://substackcdn.com/image/fetch/$s_!s35k!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png 848w, https://substackcdn.com/image/fetch/$s_!s35k!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png 1272w, https://substackcdn.com/image/fetch/$s_!s35k!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!s35k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png" width="1456" height="753" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:753,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2262466,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168573586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!s35k!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png 424w, https://substackcdn.com/image/fetch/$s_!s35k!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png 848w, https://substackcdn.com/image/fetch/$s_!s35k!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png 1272w, https://substackcdn.com/image/fetch/$s_!s35k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d6bed3b-03ea-4132-970a-557561bca105_2048x1059.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Universal features from Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment (2025).</figcaption></figure></div><p>Since the PRH conjectures that models become more aligned as they get stronger, I suspect this type of common circuit discovery will only grow more common.</p><h3>What can we make of all this?</h3><p>Besides being a deep philosophical idea, the Platonic Representation Hypothesis turns out to be an important practical insight with real-world implications. As the mechanistic interpretability community develops better tools for reverse-engineering models, I expect them to find more and more similarities; as models get bigger, this will become more common.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HUzN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HUzN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png 424w, https://substackcdn.com/image/fetch/$s_!HUzN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png 848w, https://substackcdn.com/image/fetch/$s_!HUzN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png 1272w, https://substackcdn.com/image/fetch/$s_!HUzN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HUzN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png" width="294" height="282.828" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:481,&quot;width&quot;:500,&quot;resizeWidth&quot;:294,&quot;bytes&quot;:97739,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168573586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HUzN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png 424w, https://substackcdn.com/image/fetch/$s_!HUzN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png 848w, https://substackcdn.com/image/fetch/$s_!HUzN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png 1272w, https://substackcdn.com/image/fetch/$s_!HUzN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4003053e-89b6-4dcb-aeaa-53897f723b6e_500x481.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://en.wikipedia.org/wiki/Linear_A">Linear A</a> is an ancient Greek text that humans have never been able to decrypt. Perhaps the Platonic Representation Hypothesis gives us hope for one day decoding it back to English.</figcaption></figure></div><p></p><p>As for our method (vec2vec), we found strong evidence, but things are still brittle. It seems clear that we can learn an unsupervised mapping between text-based models that are trained on the Internet, as well as <a href="https://openai.com/index/clip/">CLIP-like</a> image-text embeddings. </p><p>It&#8217;s not obvious whether we can map between languages with high fidelity. If this turns out to be true, we may be able to decode ancient texts such as <a href="https://en.wikipedia.org/wiki/Linear_A">Linear A</a> or <a href="https://www.projectceti.org/">convert whale speech back to a human language</a>. Only time will tell.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Token for Token! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How to scale RL to 10^26 FLOPs]]></title><description><![CDATA[A roadmap for RL-ing LLMs on the entire Internet]]></description><link>https://blog.jxmo.io/p/how-to-scale-rl-to-1026-flops</link><guid isPermaLink="false">https://blog.jxmo.io/p/how-to-scale-rl-to-1026-flops</guid><dc:creator><![CDATA[Jack Morris]]></dc:creator><pubDate>Thu, 10 Jul 2025 20:47:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6hjW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>TLDR: Reinforcement learning (RL) is the next training technique for building frontier-level AI models. To make it better, we need to train on more data. The current approach of scaling many environments simultaneously is messy and complicated. Instead, I propose we find a way to do next-token prediction on the Web using RL. This way, we learn to reason from general web data, instead of just math and code.</em></p><div><hr></div><p>I&#8217;ve spent a good part of the past year in denial.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>I was in denial because when <a href="https://openai.com/index/learning-to-reason-with-llms/">OpenAI released o1</a>, and explained their paradigm of <em>test-time compute</em>, I thought it was a good idea but mostly a way to get better performance out of models of fixed size. After all, letting models &#8216;think for longer&#8217; by generating more tokens lets them do more internal computation.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Dxsi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Dxsi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png 424w, https://substackcdn.com/image/fetch/$s_!Dxsi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png 848w, https://substackcdn.com/image/fetch/$s_!Dxsi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png 1272w, https://substackcdn.com/image/fetch/$s_!Dxsi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Dxsi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png" width="1456" height="818" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:818,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:94405,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168023871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Dxsi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png 424w, https://substackcdn.com/image/fetch/$s_!Dxsi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png 848w, https://substackcdn.com/image/fetch/$s_!Dxsi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png 1272w, https://substackcdn.com/image/fetch/$s_!Dxsi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbe8a574-8953-446a-b6fc-dc3fb43915f0_1980x1113.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://openai.com/index/learning-to-reason-with-llms/">The o1 release</a> from OpenAI was the first demonstration of a new type of language model, one that could think for longer to generate better answers.</figcaption></figure></div><p>So I wasn&#8217;t that surprised that these new models, termed <em>reasoning models</em>, gave better answers. And I especially wasn&#8217;t surprised when I found out these answers mostly came on problems that inherently require lots of computation, like difficult math and engineering test questions.</p><p>Don&#8217;t get me wrong: I always thought reasoning models were <em>interesting</em>. It&#8217;s cool to me that they generate &#8220;thinking traces&#8221; before giving answers (although the thinking traces might not be very reliable). And it&#8217;s amazing that the models were trained with reinforcement learning, a foundational technique in machine learning that was generally understood to be difficult to use effectively for real problems.</p><p>But I still thought of myself as a scale maximalist: all that really mattered, I thought, was training bigger models on more data. Anything else (read: reasoning models) appeared to be a coping mechanism, just a way to get by while we wait for the hardware needed to train bigger models.</p><p>I&#8217;ve spent the past few months working on RL research at Meta. It took a bit of time but I&#8217;ve come full-circle: something far more nuanced is happening with reasoning models. RL isn&#8217;t just a way to give models more compute. <strong>RL training really is teaching models something </strong><em><strong>different</strong></em><strong>, a way to use compute to generate better answers</strong> given finite model capacity. Through RL, models are clearly learning something that they&#8217;re not getting from pretraining.</p><p></p><h3>Two waves of AI scaling</h3><p>The AI research-into-production cycle moves through a few distinct phases. First, we as a community identify a new learning paradigm. Second, we find the correct datasets for training and design evaluations to know when our models are getting better, and by how much. And third, we scale it to all hell.</p><p>This cycle has happened already, exactly once. Pretraining. It started with the innocuous observation that models can learn quite a lot when trained on internet text data using next-token prediction. We realized that this gives intelligence improvements in just about every domain. And then we scaled.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6hjW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6hjW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!6hjW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!6hjW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!6hjW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6hjW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3363962,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168023871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6hjW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!6hjW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!6hjW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!6hjW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23e84ae2-5f9b-4748-aa99-11200f2ea70b_1536x1024.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">We spent 2022&#8211;2024 scaling language model pretraining: first making models bigger, and now working to pack as much knowledge we could into models of various scales. We&#8217;ll spend the next several years scaling post-training using RL.</figcaption></figure></div><p>And to be clear, pretraining research is ongoing. We&#8217;re still figuring out how to scale our models via bigger datacenters and better hardware and more efficient algorithms. And we&#8217;re gathering more and better data every year. But the upper bound of pretraining performance is really clear. To build better models, we need to give them more parameters and train them on bigger datasets. This is what the AI labs have been working on for three years or so now.</p><p>But as the dust settles on the pretraining frenzy, reasoning models are showing us a new way to scale. We&#8217;ve found a way to make models better that&#8217;s independent of the number of training tokens or model size.</p><p></p><h3>The murky path to RL scaling starts with <em>data</em></h3><p>We&#8217;ve identified a new paradigm: learning to <em>reason</em>. But reasoning models are <a href="https://www.mechanize.work/blog/the-upcoming-gpt-3-moment-for-rl/">in their GPT-3 era</a>: they&#8217;re trained on small datasets to do a narrow selection of tasks. We have a brittle proof-of-concept in the reasoning models of 2025. These models have achieved state-of-the-art scores on a small number of tasks, mostly expert-level math and coding questions.</p><p>In the case of pretraining, the path to progress was very clear. Models can learn via next-token prediction on just about <em>any</em> data, so we could simply scrape the entire Web and feed it to the models. And once we&#8217;d done that it became clear that our models were too small and we needed to make them much, much bigger.</p><p>But RL training is different. Let&#8217;s briefly remind ourselves how RL works:</p><p>Models like o1 are trained with <em>verifiable rewards</em>, meaning that after thinking and generating answers, we teach models by encouraging them to think more of the thoughts that led to correct answers, and less of the thoughts that led to incorrect answers. This is how RL algorithms like <a href="https://arxiv.org/abs/1707.06347">PPO</a> (what o1 probably uses) and <a href="https://arxiv.org/abs/2402.03300">GRPO</a> (the algorithm behind DeepSeek R1) work. They don&#8217;t teach, they incentivize.</p><p>So clearly we can only train RL models on tasks where we can score answers based on correctness. This is the idea behind <em>verifiability</em>, an RL buzzword used to describe tasks with a well-defined automatic scoring function. (The o1-style RL training paradigm is usually called <em>RLVR</em>, reinforcement learning with verifiable rewards, as to be distinguished from <em>RLHF</em>, or reinforcement learning from human feedback.)</p><p>Unfortunately, most things aren&#8217;t automatically verifiable. There aren&#8217;t perfect computer programs that can tell you whether an essay is good, for example, or an explanation.</p><p>In fact, things that we know how to automatically verify tend to be in the scientific domain. For example, <a href="https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k">OpenThoughts</a>, a recently-released dataset of training data for reasoning models, contains four categories, Code, Math, Science, and &#8216;Puzzle&#8217;:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CQ0O!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CQ0O!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png 424w, https://substackcdn.com/image/fetch/$s_!CQ0O!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png 848w, https://substackcdn.com/image/fetch/$s_!CQ0O!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png 1272w, https://substackcdn.com/image/fetch/$s_!CQ0O!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CQ0O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png" width="465" height="513.3219761499148" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8df13457-bc45-4202-a49a-d524523416be_1174x1296.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1296,&quot;width&quot;:1174,&quot;resizeWidth&quot;:465,&quot;bytes&quot;:118444,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168023871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CQ0O!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png 424w, https://substackcdn.com/image/fetch/$s_!CQ0O!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png 848w, https://substackcdn.com/image/fetch/$s_!CQ0O!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png 1272w, https://substackcdn.com/image/fetch/$s_!CQ0O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8df13457-bc45-4202-a49a-d524523416be_1174x1296.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The recent <a href="https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k">OpenThoughts dataset</a> contains verifiable tasks in Math, Science, Coding, as well as a small dataset of puzzles. (What is the Puzzle task, I wonder?)</figcaption></figure></div><p>Ok, so we can see that there are at least four domains that contain verifiable problems that we can train on. But there are many open problems here. Are those all the verifiable things that exist? Are they equally valuable? During training should we randomly alternate between them or <a href="https://arxiv.org/abs/2203.05482">train separate models and then average</a>?</p><p>In fact, in typical RL setups, we don&#8217;t even understand the marginal value of a <em>single</em> training example. One recent paper, <a href="https://arxiv.org/abs/2504.20571">Reinforcement Learning for Reasoning in Large Language Models with One Training Example</a>, demonstrated that training on just a single example with thousands of different attempts of reasoning can actually produce a very good model:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!p87E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!p87E!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png 424w, https://substackcdn.com/image/fetch/$s_!p87E!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png 848w, https://substackcdn.com/image/fetch/$s_!p87E!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png 1272w, https://substackcdn.com/image/fetch/$s_!p87E!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!p87E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png" width="1364" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:1364,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:212047,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168023871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!p87E!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png 424w, https://substackcdn.com/image/fetch/$s_!p87E!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png 848w, https://substackcdn.com/image/fetch/$s_!p87E!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png 1272w, https://substackcdn.com/image/fetch/$s_!p87E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22ce64ec-421f-4564-873e-6e813b016a59_1364x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://arxiv.org/abs/2504.20571">A recent paper</a> showed close to on-par reasoning performance from just learning a reasoning process from a <em>single training example</em>.</figcaption></figure></div><p>It&#8217;s also interesting to note the x-axis in the above graph: training only runs for 2000 steps. And that&#8217;s typical. Right now, these models are typically only trained for a few hundred or at most a few thousand steps. In the pretraining era, we often trained models on <em>trillions</em> of tokens, which meant millions of training steps.</p><p>This is mostly a compute issue: each step of RL requires <em>sampling reasoning tokens</em>, which is expensive and just difficult from a software perspective. The infrastructure to do this sort of thing is challenging and requires a lot of new engineering, since we aren&#8217;t used to doing generation at all during training, let alone at each step.</p><p>Mark my words: before we know it we&#8217;ll be running millions of steps of RLVR too.</p><p></p><h3>RL compute scales unevenly</h3><p>There are many practical engineering problems that need to be solved to scale RL.</p><p>In the pretraining days, training was a very homogenous workload, a very real-time continuous process. Batch of text passes through the model, we compute losses and backpropagate once. Queue the next back of text. This was simple and straightforward to optimize.</p><p>When we do RL, training infrastructure has to get more complicated. Gradient steps happen much less frequently and (depending on our chosen hyperparameters) we spend a lot more time generating thinking tokens.</p><p>Luckily, we&#8217;ve spent the last year or two making LLM inference super fast, and we can take advantage of these improvements here. In particular there are two really good libraries for doing inference (<a href="https://github.com/sgl-project/sglang">SGLang</a> and <a href="https://github.com/vllm-project/vllm">vLLM</a>) that make this part ~10x faster than naive python inference code. That helps a lot.</p><p>Another systems problem arises when we actually compute verifiable rewards. In the case of math problems this is usually pretty easy. Most datasets have answers computed ahead of time, so we can simply check if the final answer is correct and score accordingly (In practice, <a href="https://gist.github.com/willccbb/4676755236bb08cab5f4e54a0475d6fb">formatting makes this process slightly more complicated</a>.)</p><p>But in domains besides math, verification quickly becomes expensive. This is especially noticeable in the code domain, which is where a lot of AI labs are focusing their efforts right now.</p><p>Remember that each domain needs a domain-specific &#8220;verifier&#8221;, a system that provides rewards that guide LLMs to generate better outputs. In the case of code, this usually involves running some code and scoring based on the code&#8217;s output. Given an LLM-generated answer to a coding problem, we may need to run a bunch of unit tests and count the number of ones that pass to provide a reward.</p><p>There are a lot of people working on this right now. Doing better and faster verification, running it in parallel, scaling it properly. In some cases, the training bottleneck isn&#8217;t anything related to the model &#8211; it&#8217;s not inference or backpropagation that slows things down, but the time it takes to compile and execute model code.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YfNa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YfNa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png 424w, https://substackcdn.com/image/fetch/$s_!YfNa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png 848w, https://substackcdn.com/image/fetch/$s_!YfNa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png 1272w, https://substackcdn.com/image/fetch/$s_!YfNa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YfNa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png" width="1445" height="1182" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1182,&quot;width&quot;:1445,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:623720,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168023871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YfNa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png 424w, https://substackcdn.com/image/fetch/$s_!YfNa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png 848w, https://substackcdn.com/image/fetch/$s_!YfNa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png 1272w, https://substackcdn.com/image/fetch/$s_!YfNa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7fe7d41-0a9c-4de0-a97e-58631aaa0576_1445x1182.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The new <a href="https://www.nvidia.com/en-us/data-center/dgx-b200/">DGX B200 GPU servers from NVIDIA</a> cost $500K a pop and provide around 10^17 FLOPS for training or inference. Unfortunately, our systems for doing RL on LLMs are pretty primitive and can&#8217;t get anywhere near this level of performance yet.</figcaption></figure></div><p></p><p>Since <a href="https://wccftech.com/nvidia-blackwell-dgx-b200-price-half-a-million-dollars-top-of-the-line-ai-hardware/">a single B200 costs over $500K</a>, any time training spends bottlenecked by CPU execution is a big waste of money.</p><p>One path to scaling RL is optimizing this kind of system: making verifiers faster and more reliable, optimizing the new generate-before-backprop <a href="https://github.com/volcengine/verl">training pipeline</a>, and designing <a href="https://arxiv.org/html/2412.01152v1">clever systems</a> that let us scale RL across datacenters.</p><p>In addition to all these systems-level improvements, we&#8217;ll need to build <a href="https://github.com/open-thought/reasoning-gym">lots of new environments</a> to try and learn diverse skills via RL training. Oh, and no one really knows what the right skills are to learn via RL, or how to combine them. So we&#8217;ll have to try lots of different combinations and train many models. And we can try averaging them in different ways via some kind of <a href="https://arxiv.org/abs/2203.05482">model souping</a> (also known as model merging). We&#8217;ll just run lots of evaluations to find which combination of environments and souping produces the best model. This sounds difficult, doesn&#8217;t it? And quite messy.</p><p>What if I told you there was another way?</p><p></p><h3>What does it mean to be verifiable?</h3><p>If we want to scale RL in verifiable settings, we should probably start by figuring out which things are verifiable in the first place. It&#8217;s my feeling that people have been throwing this word around a lot, but there&#8217;s no clear definition.</p><p>It all comes down to what we can train into the models. If we can check a model output and provide a score, that&#8217;s good enough.</p><p>Wait&#8211; but isn&#8217;t this how language modeling works <em>already</em>?</p><p></p><h3>Pretraining for reasoning with next-token prediction</h3><p>Before making my proposal, let me start with listing a few core tenets that I believe about the current state of AI:</p><ol><li><p><strong>The only data we&#8217;ve found that really &#8220;works&#8221; (i.e. helps us build more intelligent models) is web-scale pretraining data</strong>. Ilya Sutskever famously compared all the human text on the internet to <a href="https://www.youtube.com/watch?v=YD-9NG1Ke5Y">a reserve of fossil fuel</a>: it&#8217;s exceptionally useful, but finite.</p></li><li><p>Reasoning, at its core, is a way to get better performance out of smaller models. It&#8217;s not doing anything more magical. Crucially, we&#8217;re getting limited new signal from the verifier itself; RL with verification is just a way to elicit capabilities that already exist within models. (<a href="https://limit-of-rlvr.github.io/">This is a common belief about RL.</a>)</p></li><li><p><strong>There is nothing special about math and code</strong>. These modalities happen to lie in a space that&#8217;s difficult to model with a reward model (so prior approaches didn&#8217;t work super well) but easy to verify. And we happen to care about them (<a href="https://cursor.com/en/blog/series-c">automating coding seems especially valuable)</a>. But we should be able to learn to reason from any type of data.</p></li><li><p>And finally, <strong>we haven&#8217;t fully saturated models with Internet data</strong>. Today&#8217;s models don&#8217;t seem to <a href="https://arxiv.org/abs/2505.24832">have enough capacity</a> to memorize the entire Internet. Additional pretraining on Web data should still give us a performance boost &#8211; and might be enough to learn to reason.</p></li><li><p><strong>Next-token prediction is verifiable.</strong> This is perhaps the central argument I&#8217;m making. The current strategy of checking if a math problem has been answered correctly is spiritually no different than confirming whether a model has outputted the proper next tokens.</p></li></ol><p>Putting all this together, I&#8217;m betting that the &#8220;right way&#8221; to scale RL is by unifying it with next-token prediction. We should teach models to reason by practicing reasoning at scale on the vast diversity of data available on the Web.</p><p>Learning to reason via next-token prediction.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tZ5F!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tZ5F!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png 424w, https://substackcdn.com/image/fetch/$s_!tZ5F!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png 848w, https://substackcdn.com/image/fetch/$s_!tZ5F!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png 1272w, https://substackcdn.com/image/fetch/$s_!tZ5F!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tZ5F!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png" width="1456" height="1075" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1075,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:296620,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168023871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tZ5F!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png 424w, https://substackcdn.com/image/fetch/$s_!tZ5F!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png 848w, https://substackcdn.com/image/fetch/$s_!tZ5F!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png 1272w, https://substackcdn.com/image/fetch/$s_!tZ5F!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb827aab5-7961-4e9b-ab48-5b666a8553ae_1744x1288.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The proposed framework of learning to reason via next-token prediction.</figcaption></figure></div><p>This shows a comparison of the new paradigm demonstrated on a math problem. Normal next-token prediction is guessing which tokens come next. Typical RLVR allows the model to &#8216;think&#8217; for a few tokens and then rewards it for outputting the right thing. Our idea of reasoning with next-token prediction (RNTP) would allow the model to think and then reward it based on the next-token prediction loss of the outputs in the &lt;answer&gt; tag. It&#8217;s a hybrid between traditional language model pretraining and the new reasoning model RL training.</p><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hhru!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hhru!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png 424w, https://substackcdn.com/image/fetch/$s_!hhru!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png 848w, https://substackcdn.com/image/fetch/$s_!hhru!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png 1272w, https://substackcdn.com/image/fetch/$s_!hhru!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hhru!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png" width="800" height="404" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:404,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:66186,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168023871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hhru!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png 424w, https://substackcdn.com/image/fetch/$s_!hhru!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png 848w, https://substackcdn.com/image/fetch/$s_!hhru!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png 1272w, https://substackcdn.com/image/fetch/$s_!hhru!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56425c6f-5cab-4cad-ae74-4749561d227c_800x404.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Grok 2 was trained on more FLOPs than GPT-4. Apparently Grok 3, which came out later, was trained on around 10^26 FLOPs, which would put it above the top of this graph. But that was all supervised learning. How do we scale RL to use this much compute?</p><p></p><h3>What do we even need RL for?</h3><p>Now that we&#8217;ve stripped things down to their base components, it might not be obvious what benefit we get from doing reinforcement learning, if any.</p><p>The answer lies in the &lt;think&gt; tokens. In the picture above, we generated everything between &lt;thinks&gt; directly from the model. <em>There&#8217;s no supervision for this</em>.</p><p>In other words, we&#8217;re trying to get the model to learn to reason without knowing what reasoning should look like. We just sample lots of things from the model and encourage it to do the things that get rewards. If there was ground-truth reasoning, we could use the typical supervised training techniques to train the model to output the proper reasoning chains.</p><p>But in the real world, there&#8217;s no ground-truth for reasoning, so we can&#8217;t do supervised learning. And in fact we want it this way &#8211; this is the magic of reinforcement learning. We&#8217;re <em>hoping</em> that the model will discover reasoning chains that are more useful than anything we could ever write ourselves.</p><p></p><h3>Scaling reasoning via next-token prediction</h3><p>If you&#8217;ve read this far, and you agree this idea makes sense, you might be thinking about how it could be tricky to implement.</p><p>And in fact, you&#8217;re right. This is where the research comes in. Almost all research that matters comes from figuring out how to implement and scale ideas that make sense from first principles.</p><p>For example: what exactly is the reward? Do we give the model a point for guessing a token correctly? Should we reward it more for multiple tokens in a row? Perhaps we use a string-similarity reward like BLEU score, <a href="https://arxiv.org/abs/1808.08866">as was common in machine translation in 2018</a>. We could do some kind of <a href="https://arxiv.org/abs/2506.18254">self-evaluation</a>, where a decent model can look at its own outputs and decide which ones should get rewards. Perhaps we filter tokens by <a href="https://arxiv.org/pdf/2506.08007">entropy</a> and use that to determine which to reason about. Or maybe we want to <a href="https://arxiv.org/abs/2505.19590">account for confidence in the reward</a>, and give the model more points for being confidently correct.</p><p>Another question: how many times should you &#8220;reason&#8221; within a single text chunk? One approach is to insert reasoning tokens at a random position per-chunk. Or perhaps we allow models to reason multiple times throughout each chunk. But then we&#8217;d have to figure out whether how many times can learn from a given text chunk with different reasoning patterns before memorization starts to occur.</p><p>There are additional difficulties that arise when switching from math and code to <em>general</em> reasoning. One reason we like math and code is because they&#8217;re difficult for base models to do &#8220;from scratch&#8221; but often easy to learn via reasoning. This won&#8217;t be the case with general text: some tokens are already extremely low-entropy, and therefore easy to predict; other tokens are nearly impossible, and will never be guessed correctly with any amount of reasoning.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!AwyA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!AwyA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png 424w, https://substackcdn.com/image/fetch/$s_!AwyA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png 848w, https://substackcdn.com/image/fetch/$s_!AwyA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png 1272w, https://substackcdn.com/image/fetch/$s_!AwyA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!AwyA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png" width="594" height="492.68227424749165" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:992,&quot;width&quot;:1196,&quot;resizeWidth&quot;:594,&quot;bytes&quot;:249764,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168023871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!AwyA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png 424w, https://substackcdn.com/image/fetch/$s_!AwyA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png 848w, https://substackcdn.com/image/fetch/$s_!AwyA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png 1272w, https://substackcdn.com/image/fetch/$s_!AwyA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d693ef-e1b1-4a78-afde-1983105526ac_1196x992.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Even the best pretraining datasets are still a long way from perfect.</figcaption></figure></div><p>Andrej Karpathy recently noted that if you <a href="https://huggingface.co/datasets/HuggingFaceFW/fineweb">actually look at samples from a typical pretraining dataset</a>, they&#8217;re quite ugly, and there is a lot of obvious noise. One benefit of scale is that it irons out many of these low-level idiosyncrasies: after a lot of training, much of this noise gets averaged away. It&#8217;s possible that this would happen with my proposed RL training scheme. <strong>If we train for long enough, on enough tokens, we might not even care what the exact reward or reasoning schema looks like.</strong></p><h3>But wait, didn&#8217;t somebody try this already?</h3><p>Those among us who diligently troll ArXiv for the latest nuggets of progress might recognize that someone recently proposed something like this in a recently released preprint (<a href="https://arxiv.org/abs/2506.08007">*Reinforcement Pre-Training</a>).* This research was praised on twitter (the title sounds important, and the figure is funny!) but disappointed a lot of researchers:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!e2No!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!e2No!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png 424w, https://substackcdn.com/image/fetch/$s_!e2No!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png 848w, https://substackcdn.com/image/fetch/$s_!e2No!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png 1272w, https://substackcdn.com/image/fetch/$s_!e2No!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!e2No!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png" width="1126" height="726" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:726,&quot;width&quot;:1126,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:358847,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/168023871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!e2No!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png 424w, https://substackcdn.com/image/fetch/$s_!e2No!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png 848w, https://substackcdn.com/image/fetch/$s_!e2No!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png 1272w, https://substackcdn.com/image/fetch/$s_!e2No!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8f12234-e78c-47fd-b0cd-9afbff8145f0_1126x726.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Headlining figure from the recent &#8216;Reinforcement Pre-Training&#8217; paper, which also proposes the idea of pretraining for RL via next-token prediction.</figcaption></figure></div><p>Headliner figure from the recent &#8216;Reinforcement Pre-Training&#8217; paper, which also proposes the idea of pretraining for RL via next-token prediction.</p><p>To be more specific, this paper proposed something similar to what I&#8217;m advocating for: using large unlabeled text datasets and next-token prediction to scale RL! And it has pretraining in the name, just like I was describing.</p><p>Alas, it turns out to be a classic case of academic titlegrabbing. What the paper actually does is very specific: they finetune a single model with chain-of-thought to improve single-token outputs for some multiple-choice questions. They&#8217;re not actually doing pretraining&#8211;just finetuning!&#8211; and train on a small subset of questions from <a href="https://omni-math.github.io/">a single math dataset</a>. There aren&#8217;t a lot of comparisons to any of the other RLVR papers, so it&#8217;s hard to tell whether this thing even works, and if so, when and how well.</p><p>Normally I&#8217;d file this type of work away as a sort of negative result &#8211; if a simpler and more general setting worked, they surely would have tried it in this paper, right? But that&#8217;s exactly what I don&#8217;t think we should do. My overall point in this piece is that <strong>if something makes sense from first principles, we should keep working on it until we work out all the kinks.</strong></p><p>Making good ideas work often turns out to require significantly more labor than academic researchers expect from a single project. But this is is the price of progress.</p><p></p><h3>What&#8217;s next?</h3><p>It&#8217;s very exciting to me that (a) RL works and (b) no one knows the right way to do it. There is so much opportunity here. One way or another, we will have much better reasoning models in a year or two; it&#8217;s just that the path is unclear. Before we can see with clarity, we have a lot to learn. If reasoning next-token prediction turns out to really be the right way to scale RL, we&#8217;re going to need to answer all these questions and many more.</p><div><hr></div><p>Thanks to my friends Wenting Zhao, Will Brown, and Nishanth Kumar for reading this blog post early and providing helpful feedback.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Superintelligence, from First Principles]]></title><description><![CDATA[There are only two ways to learn: supervised learning and reinforcement learning. Which one will give us superintelligence?]]></description><link>https://blog.jxmo.io/p/superintelligence-from-first-principles</link><guid isPermaLink="false">https://blog.jxmo.io/p/superintelligence-from-first-principles</guid><dc:creator><![CDATA[Jack Morris]]></dc:creator><pubDate>Wed, 18 Jun 2025 14:37:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Lots of people have been talking about how we&#8217;ll reach AGI (artificial general intelligence) or ASI (artificial super intelligence) with current technology. Meta has recently announced that they&#8217;re building a top-secret &#8220;superintelligence&#8221; lab with billions of dollars of funding. OpenAI, Anthropic, and Google DeepMind have all said in one way or another that their goal is to build superintelligent machines.</p><p>Sam Altman in particular has stated that superintelligence is simply a problem of <em>engineering</em>:</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7LbV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7LbV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png 424w, https://substackcdn.com/image/fetch/$s_!7LbV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png 848w, https://substackcdn.com/image/fetch/$s_!7LbV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png 1272w, https://substackcdn.com/image/fetch/$s_!7LbV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7LbV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png" width="494" height="204.2420168067227" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:492,&quot;width&quot;:1190,&quot;resizeWidth&quot;:494,&quot;bytes&quot;:83977,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/166246775?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7LbV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png 424w, https://substackcdn.com/image/fetch/$s_!7LbV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png 848w, https://substackcdn.com/image/fetch/$s_!7LbV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png 1272w, https://substackcdn.com/image/fetch/$s_!7LbV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F435e5a13-9acc-4aa5-9ef5-83b869a2036a_1190x492.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>This implies that researchers at OpenAI know how to build superintelligence, and simply have to put in time and effort to build the systems required to get there.</p><p>Now, I&#8217;m an AI researcher, and it&#8217;s not clear to me how to build superintelligence &#8211; I&#8217;m not even sure it&#8217;s possible. So in this post I want to think through the details a little bit and speculate how someone might try to build superintelligence from first principles.</p><p>We&#8217;re going to assume that the fundamental building blocks of the technology is decided: we&#8217;re going to build superintelligence with neural networks and train it with backpropagation and some form of machine learning.</p><p><a href="https://blog.jxmo.io/p/there-are-no-new-ideas-in-ai-only">I don&#8217;t really believe that architecture (the shape of the neural network) matters much.</a> So we&#8217;ll gloss over that architectural detail and make a strong assumption: superintelligence will be built using Transformers, by far the most popular architecture for training these systems on large datasets.</p><p>So then we already know a lot: superintelligence will be a transformer neural network, it will be trained via some machine learning learning objective and gradient-based backpropagation. There are still two major open questions here. What learning algorithm do we use, and what data?</p><p>Let&#8217;s start with the data.</p><h2>The data: it&#8217;s gotta be text</h2><p>Most of the big breakthroughs that led to chatGPT inevitably came from leveraging the treasure trove of human knowledge inside The Internet to learn things. The true scale of this enterprise is mind-boggling and mostly hidden by modern engineering, but let&#8217;s spend a second trying to get to the bottom of it all.</p><p><strong>The best systems we have right now all learn from Internet text</strong> As of writing this (June 2025) I don&#8217;t think there has been any demonstrated overall improvement from integrating non-text data into a model. This includes images, video, audio, and extrasensory data from robotics &#8211; we don&#8217;t know how to use <em>any</em> of these modalities to make chatGPT smarter.</p><p>Why is this, by the way? It&#8217;s possible this is just a science or engineering challenge, and we&#8217;re not doing things the right way; but it&#8217;s certainly also possible <strong>that there&#8217;s just something special about text</strong>. After all, every bit of text on the internet (<a href="https://www.technologyreview.com/2022/12/20/1065667/how-ai-generated-text-is-poisoning-the-internet/">before LLMs</a>, <a href="https://www.forbes.com/sites/torconstantino/2024/08/26/is-ai-quietly-killing-itself-and-the-internet/">anyway</a>) is a reflection of a human&#8217;s thought process. In a sense, huma-written text is preprocessed to have very high information content.</p><p>Take this in contrast to images, for example, which are are raw views of the world around us, captured with no human intervention. It&#8217;s certainly possible that the fact text written by actual people carries some intrinsic value that pure sensory inputs from the world around us never will.</p><p>So until someone demonstrates otherwise, let&#8217;s operate under this assumption that only text data is important.</p><p><strong>Ok, so how much text do we have?</strong></p><p>A next question is how large this dataset might be.</p><p>Lots of people have written about what to do if we&#8217;re running out of text data. Dubbed the &#8220;<a href="https://www.lesswrong.com/posts/axjb7tN9X2Mx4HzPz/the-data-wall-is-important">data wall</a>&#8221; or &#8220;<a href="https://arxiv.org/abs/2305.13230">token crisis</a>&#8221;, people have written about what to do and <a href="https://arxiv.org/abs/2305.16264">how to scale our models</a> if we really run out of data.</p><p>And it seems that might really be happening. Many engineers at the big AI labs have spend countless hours dutifully scraping every last useful bit of text from the darkest corners of the Web, going so far as to <a href="https://www.theverge.com/2024/4/6/24122915/openai-youtube-transcripts-gpt-4-training-data-google">transcribe a million hours of YouTube videos</a> and <a href="https://apnews.com/article/openai-chatgpt-associated-press-ap-f86f84c5bcc2f3b98074b38521f5f75a">purchasing large troves of news stories</a> to train on.</p><p>Luckily there might be another data source available here (verifiable environments!) but we&#8217;ll get to hat later.</p><h2>The learning algorithm</h2><p>Above we discovered another important principle: the best path we have toward superintelligence lies in text data. In other words, <strong>AGI is probably just LLMs or nothing</strong>. Some other promising areas are learning from video and robotics, but neither of those seem nearly far along enough to produce independent intelligent systems by at least 2030. They&#8217;re also far more data-hungry; learning from text is naturally extremely efficient.</p><p>Now we have to confront the most important question. What&#8217;s the learning algorithm for superintelligence?</p><p>In the field of machine learning, there are basically two tried-and-true ways to learn from large datasets. One is <em>supervised learning</em>, training a model to increase the likelihood of some example data. The other is <em>reinforcement learning</em>, which involves generating from a model and rewarding it for taking &#8220;good&#8221; actions (for some user-written definition of &#8220;good&#8221;).</p><p>Now that we know this taxonomy, it becomes clear that any potential superintelligent system as written has to be trained using either supervised or reinforcement learning (or both).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WV6A!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WV6A!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png 424w, https://substackcdn.com/image/fetch/$s_!WV6A!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png 848w, https://substackcdn.com/image/fetch/$s_!WV6A!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png 1272w, https://substackcdn.com/image/fetch/$s_!WV6A!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WV6A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png" width="1456" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1077447,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/166246775?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WV6A!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png 424w, https://substackcdn.com/image/fetch/$s_!WV6A!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png 848w, https://substackcdn.com/image/fetch/$s_!WV6A!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png 1272w, https://substackcdn.com/image/fetch/$s_!WV6A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb9da5764-9e1f-4492-9f56-44ad2d7028ef_2208x1244.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Yann LeCun has infamously stated that he knows the recipe to intelligence. In fact, intelligence is a cake, and reinforcement learning is but a tiny cherry on top.</figcaption></figure></div><p></p><p>Let&#8217;s think through both options.</p><h3>Hypothesis 1: Superintelligence arises from supervised learning</h3><p>Remember 2023? That was around when people really became excited about scaling laws; GPT-4 came out, and people were worried that if models continued to scale, they would become dangerous.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!csYA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!csYA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png 424w, https://substackcdn.com/image/fetch/$s_!csYA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png 848w, https://substackcdn.com/image/fetch/$s_!csYA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png 1272w, https://substackcdn.com/image/fetch/$s_!csYA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!csYA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png" width="274" height="310.92198581560285" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/abcd084a-fa7a-429c-a149-0645802da5eb_564x640.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:640,&quot;width&quot;:564,&quot;resizeWidth&quot;:274,&quot;bytes&quot;:226471,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/166246775?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!csYA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png 424w, https://substackcdn.com/image/fetch/$s_!csYA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png 848w, https://substackcdn.com/image/fetch/$s_!csYA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png 1272w, https://substackcdn.com/image/fetch/$s_!csYA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabcd084a-fa7a-429c-a149-0645802da5eb_564x640.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Around 2023, a lot of people online became concerned that LLMs scaled up with simple supervised learning would soon become superintelligent.</figcaption></figure></div><p>There was a prevailing opinion for some time that lots of supervised learning, specifically in the form of &#8220;next-token prediction&#8221;, could lead to superintelligent AI. Notably Ilya Sutskever gave a talk about how <a href="https://www.youtube.com/watch?v=AKMuA_TVz3A&amp;t=1s">next-token prediction is simply learning to compress the Universe</a>, since doing this well requires simulating all possible programs (or something like that).</p><p>I think the argument went something like this:</p><ol><li><p>Accurate next-token prediction requires modeling what <em>any</em> human would write in <em>any</em> scenario</p></li><li><p>The better you can do this for a single person, the more closely you approximate the intelligence of that person</p></li><li><p>Since the internet contains text written by many people, to this well on e.g. a large text pretraining corpus requires accurately modeling the intelligence of <em>many</em> people</p></li><li><p>Accurately modeling the intelligence of many people is superintelligence</p></li></ol><p><strong>The vibes argument: can we even reach superintelligence by modeling humans?</strong></p><p>Personally I think there are a few flaws in this logic, starting with the fact that we seem to have already created systems that are far above human-level in next-token prediction but still fail to exhibit human-level general intelligence. Somehow we&#8217;ve built systems that learned what we told them to learn (to predict the next token) but still can&#8217;t do what we want them to do (answer questions without making things up, follow instructions perfectly, etc.).</p><p>This might be simply a failure of machine learning. We&#8217;re training a model to predict the average human outcome in every situation. The learning objective disincentives giving too-low a probability to any possible outcome. This paradigm often leads to something called <a href="https://en.wikipedia.org/wiki/Mode_collapse">mode collapse</a>, where a model gets very good at modeling average outcomes without learning the tails of distribution.</p><p>It&#8217;s possible that these issues would go away at scale. Billion-parameter models, like LLAMA, hallucinate. But this is only 10^9. What happens when we train models with 10^19 parameters? Perhaps that&#8217;s enough <a href="https://arxiv.org/abs/2505.24832">capacity</a> for a single LLM to independently model all eight billion humans and provide independent data-driven predictions for each.</p><p><strong>The infra argument: we won&#8217;t be able to scale up our models and/or our data</strong></p><p>But it turns out it&#8217;s a moot point, because we may never scale up to 10^19 parameters. This hypothesis basically arose from the 2022-or-so-era deep learning school of thought driven by the wild success of language model scaling laws that continually scaling model and data size would lead to perfect intelligence.</p><p>It&#8217;s 2025 now. The theoretical argument remains unchallenged and the scaling laws have held faithful. But it turns out that scaling models gets really hard above a certain size (and in 2022 we were already really close to the line of what we could do well). Companies are already far, far beyond what we can do with a single machine &#8211; all the latest models are trained on giant networks of hundreds of machines.</p><p>The continued effort at scaling model size toward trillions of parameters is causing a hardware shortage (see: NVIDIA stock) and actually an <em>electricity</em> shortage too. Bigger models would draw so much power that they can&#8217;t be located in a single place; companies are looking into research to distribute model-training over multiple far-flung datacenters and even buying old nuclear power plants in an effort to resuscitate them and use them to train the next generation of bigger AI models. We&#8217;re living in wild times.</p><p>In addition to model size, we are potentially running out of data. No one knows how much of the Internet was used to train each model, but it&#8217;s safe to say that it&#8217;s quite a lot. And great engineering effort over the last few years at big AI labs has gone into scraping the proverbial barrel of Internet text data: OpenAI has apparently transcribed all of Youtube, for example, and high-signal sites like Reddit have been scraped time and time again.</p><p>Scaling model size orders of magnitude beyond 100B parameters looks to be hard, as is scaling data size beyond 20T tokens or so. These factors seem to indicate that it will be hard to scale supervised learning much more than 10x further in the next three years or so &#8211; so attempts at superintelligence might have to come from somewhere else.</p><h3>Hypothesis 2: Superintelligence from a combination of supervised and reinforcement learning</h3><p>So maybe you bought one of the arguments above: either we won&#8217;t be able to scale up pre-training by orders of magnitude for a long time, or even if we did, doing really well at predicting human tokens won&#8217;t produce systems that are any smarter than humans.</p><p>Whatever your issue with supervised learning, there&#8217;s another way. The field of <em>reinforcement learning</em> offers a suite of methods to learn from <em>feedback</em> instead of demonstrations.</p><blockquote><p>&#9888;&#65039; <strong>Why do we need supervised learning?</strong> <br>Well, RL is hard. You might be wondering why we can&#8217;t just use reinforcement learning all the way down. From a practical perspective, there are a lot of drawbacks to RL. The short explanation is that SL is a lot more stable and efficient than RL. An easy-to-understand reason is that because RL works by letting the model generate &#8220;actions&#8221; and rating them, a randomly initialized model is basically terrible and all of its actions are useless, and it has to accidentally do something good to get any semblance of a reward. This is called the <em>cold start problem</em>, and it&#8217;s just one of many issues with reinforcement learning. Supervised learning from human data turns out to be a great way to get around the cold start problem.</p></blockquote><p>So let&#8217;s reiterate the RL paradigm: the model tries stuff, and we tell it whether it did well or not. This can come in two ways: either human raters tell the model it did well (this is more or less how typical RLHF works) or an automated system does that instead.</p><h3>Hypothesis 2A: RL from human verifiers</h3><p>Under this first paradigm, we provide the model with human-based rewards. We want our model to be superintelligent, so we want to reward it for producing text that looks closer to superintelligence (as judged by humans).</p><p>In practice, this data is really expensive to collect. A typical reinforcement learning from human feedback (RLHF) setting <a href="https://arxiv.org/abs/2203.02155">involves training a reward model</a> to emulate the human feedback signal here. Reward models are necessary because they allow us to give much more feedback than we can feasibly collect from actual humans. In other words, they&#8217;re a computational crutch. We&#8217;re going to treat reward models as an engineering detail and ignore them for now.</p><p>So we&#8217;re imagining a world where we have infinite humans available to label data for an LLM and provide arbitrary rewards, where higher rewards indicate getting the model closer to superintelligence.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!k6d6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!k6d6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png 424w, https://substackcdn.com/image/fetch/$s_!k6d6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png 848w, https://substackcdn.com/image/fetch/$s_!k6d6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png 1272w, https://substackcdn.com/image/fetch/$s_!k6d6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!k6d6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png" width="640" height="480" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e69a9a95-89a1-4674-b520-7c182f08751b_640x480.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:480,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:421174,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/166246775?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!k6d6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png 424w, https://substackcdn.com/image/fetch/$s_!k6d6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png 848w, https://substackcdn.com/image/fetch/$s_!k6d6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png 1272w, https://substackcdn.com/image/fetch/$s_!k6d6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe69a9a95-89a1-4674-b520-7c182f08751b_640x480.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">"This is a thousand monkeys working at a thousand typewriters. Soon, they'll have written the greatest novel known to man.&#8221; &#8211;Mr. Burns</figcaption></figure></div><p>Ignore all the logistic complexity. Assume this approach is possible to run at scale (as it could be some day, even if it&#8217;s not today). Would it work? Could a machine learning purely from human reward signals climb the intelligence ladder and pass humans?</p><p>Another way of phrasing this question is: <em>can we verify superintelligence</em> when seeing it, even if we can&#8217;t generate examples of it ourselves? Remember that humans <em>by definition</em> aren&#8217;t superintelligent. But can we recognize superintelligence when shown to us? And can we do so reliably enough to provide useful gradient signal to an LLM which could gather lots of feedback of this type and bootstrap its way to superintelligence?</p><p>Some people like to point out that <em>&#8220;generation is naturally harder than verification&#8221;</em>. you know a good movie when you watch one, but that doesn&#8217;t mean you&#8217;re able to go out there and produce one yourself. This dichotomy pops up a lot in machine learning. It&#8217;s computationally much easier to differentiate cat photos from dog photos than it is to generate entire cats.</p><p>Similarly, if humans can verify superintelligence, then it may be possible to train a superintelligent model based using reinforcement learning on human feedback. As a concrete example, you could have an LLM write many novels, reward it based on a human notion of the ones that are good, and repeat this many many times, until you have a superintelligent novel-writing machine.</p><p>Do you notice any problems with this logic?</p><h3>Hypothesis 2B: RL from automated verifiers</h3><p>More recently people have been excited about using similar approaches to train better language models.</p><p>And when we let the computer rate the intermediate performance of our RL algorithms, we can either do this with a <em>model</em> or an <em>automated verifier</em>. For automated verifiers, think of chess, or coding. We can write rules that check if the computer won a chess game and issue a reward if a checkmate is reached. In coding, we can run unit tests that reward the computer for writing code that correctly passes some specifications.</p><p>Using a verifier would be much more practical &#8211; it would allow us to remove humans from the loop here entirely (besides the fact that they were used to write the entire Internet). A recipe for superintelligence using verifiers would be something like:</p><ol><li><p>Pretrain an LLM using supervised learning on a large amount of Internet text</p></li><li><p>Plug it into some verification system that can provide rewards for good LLM outputs</p></li><li><p>Run for a long time</p></li><li><p>Achieve superintelligence</p></li></ol><p>Is this a good idea? Would it even work?</p><p>Famously, DeepMind&#8217;s AlphaGo achieved &#8220;Go supremacy&#8221; (read: beat all the humans, even the ones who trained for decades) by a combination of reinforcement and supervised learning. The second version of AlphaGo, known as AlphaGo Zero, learned by <a href="https://deepmind.google/discover/blog/alphago-zero-starting-from-scratch/">playing against itself for forty straight days</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!x7bo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!x7bo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png 424w, https://substackcdn.com/image/fetch/$s_!x7bo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png 848w, https://substackcdn.com/image/fetch/$s_!x7bo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png 1272w, https://substackcdn.com/image/fetch/$s_!x7bo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!x7bo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png" width="1200" height="675" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:675,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:564084,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/166246775?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!x7bo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png 424w, https://substackcdn.com/image/fetch/$s_!x7bo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png 848w, https://substackcdn.com/image/fetch/$s_!x7bo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png 1272w, https://substackcdn.com/image/fetch/$s_!x7bo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b48f536-7afb-4dfc-a11b-b9ab32503acc_1200x675.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">In 2016, <a href="https://www.cnet.com/tech/tech-industry/google-alphago-artificial-intelligence-victor-in-game-of-man-vs-machine/">AlphaGo won four out of five games</a> to defeat human Go champion Lee Sedol. The original AlphaGo was trained using supervised learning.  The next version of AlphaGo learned with reinforcement learning: by playing against itself for millions of games.</figcaption></figure></div><p>Note that Go has a very important property that many real-world tasks do not. Go is naturally <em>verifiable</em>. By this I mean that we can plug my game of Go into a rule-based computer program and receive a signal indicating whether I won or not. Extrapolating this over a long time horizon, you can tell whether an individual move is &#8220;good&#8221; or not based on how it affects the probability of the game ending with a win. This is more or less how RL works.</p><p>And using this verifiability, AlphaGo was able to achieve something very important, something that the AI labs have aspired to for a while: AlphaGo gets better when it thinks for longer. Language models, by default, cannot do this.</p><p>But this was essentially the breakthrough that OpenAI announced last fall. They used this paradigm of reinforcement learning from verifiable rewards to train O1, a model that, like AlphaGo, can think for longer and produce better outputs:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DvLK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DvLK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png 424w, https://substackcdn.com/image/fetch/$s_!DvLK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png 848w, https://substackcdn.com/image/fetch/$s_!DvLK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png 1272w, https://substackcdn.com/image/fetch/$s_!DvLK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DvLK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png" width="1456" height="818" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:818,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:94405,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/166246775?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DvLK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png 424w, https://substackcdn.com/image/fetch/$s_!DvLK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png 848w, https://substackcdn.com/image/fetch/$s_!DvLK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png 1272w, https://substackcdn.com/image/fetch/$s_!DvLK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F634ecb88-0f64-44a3-b593-3edb22885ccf_1980x1113.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">In the <a href="https://openai.com/index/learning-to-reason-with-llms/">o1 blog post</a>, OpenAI introduced a line of &#8220;reasoning models&#8221; that learn from reinforcement learning on verifiable rewards.</figcaption></figure></div><p>Gazing at the beautiful graphs above (and noting the log x-axis!) we observe that o1 is indeed getting better with more thinking time. But take a look at the title: this is on <a href="https://artofproblemsolving.com/wiki/index.php/2024_AIME_I_Answer_Key">AIME, a set of extremely difficult math problems </a><em><a href="https://artofproblemsolving.com/wiki/index.php/2024_AIME_I_Answer_Key">with integer solutions</a></em>. Read: it&#8217;s not an open-ended task. It&#8217;s a task with <em>verifiable</em> training data, because you can check if an LLM generated the proper answer or not and reward the model accordingly.</p><p>It turns out that with current LLMs, pretrained to do arbitrary tasks pretty well, they can make a decent guess at AIME problems, and we can use this to train them via RL to make better and better guesses with time. (And the coolest part, which we won&#8217;t talk about here, is that they generate more and more &#8220;thinking tokens&#8221; to do so, giving us the test-time compute graph from the o1 blog post shown above.)</p><h3>Is RLVR a path to superintelligence?</h3><p>It is clear that OpenAI, Google, and the other AI labs are very excited about this type of RL on LLMs and think it might just give them superintelligence. It also seems likely to me that this paradigm is what Sam A. was mentioning in his vague tweet above. The &#8220;engineering problem&#8221; of superintelligence is to build lots of RL environments for many different types of tasks and train an LLM to do them all at the same time.</p><p>Let&#8217;s think through the bull case here. The verifiable tasks we know of are coding (you can check if code is correct by running it) as well as math (not proofs, really, but stuff with numeric solutions). If we were able to gather all the verifiable things in the world and train on them all at the same time (or separately, and then do <a href="https://proceedings.neurips.cc/paper_files/paper/2022/hash/70c26937fbf3d4600b69a129031b66ec-Abstract-Conference.html">model merging</a>) &#8211; would this really produce <em>general</em> <em>superintelligence</em>?</p><p>There are a few logical leaps here. The most important is that we don&#8217;t know how well RL on verifiable task transfers out-of-domain. Does training a model to do math problems somehow inherently teach it how to book a flight? Or would training a model to code better in verifiable environments even make it a better software engineer overall?</p><p>Let&#8217;s pretend for a second that this <em>does</em> turn out to be true, and RL transfers neatly to all sorts of tasks. This would be huge. It would put the AI companies in an arms race to produce the most diverse, useful, and well-engineered set of tasks to RL their LLMs on. There would likely be multiple companies producing a &#8220;superintelligent LLM&#8221; in this fashion.</p><p>But this outcome seems unlikely to me. I&#8217;d guess that if RL did transfer extremely well to other domains then we&#8217;d know about it by now. My humble prediction is that LLMs will continue to get better at things <em>within their training distribution</em>. As we gather more and more types of tasks and train on them, this will produce more and more useful LLMs on a wide array of tasks. But it won&#8217;t give us a single superintelligent model.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Case for More Ambition]]></title><description><![CDATA[why AI researchers should dream bigger and publish less]]></description><link>https://blog.jxmo.io/p/the-case-for-more-ambition</link><guid isPermaLink="false">https://blog.jxmo.io/p/the-case-for-more-ambition</guid><dc:creator><![CDATA[Jack Morris]]></dc:creator><pubDate>Mon, 09 Jun 2025 14:47:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>What are the most important problems in your field? And why aren&#8217;t you working on them?</em></p><p>This is the pitch made by Richard Hamming in his famous talk <a href="https://www.cs.utexas.edu/~dahlin/bookshelf/hamming.html">You and Your Research.</a> It&#8217;s been nearly forty years since the talk was first given and yet it feels like more people are ignoring this advice than ever before.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>This may be especially true in AI.</p><p>Our field is full of brilliant people &#8211; researchers, engineers, scientists. Students are flocking to major in computer science and learn about the technology from an early age. More people are applying to graduate school than ever before. The AI labs are raising money and publishing all sorts of findings in preprints and blog posts.</p><p>And yet it doesn&#8217;t feel like the actual <em>science</em> is progressing faster than it was five years ago. We&#8217;re publishing more papers than ever, but discovering about the same amount. I don&#8217;t think it&#8217;s a talent issue, and it&#8217;s certainly not a funding issue. I think we have an <em>ambition</em> issue.</p><h3>More papers are being published than ever</h3><p>This year there were over <a href="https://cspaper.org/topic/76/submission-tsunami-at-neurips-2025-is-peer-review-about-to-collapse">25,000 papers submitted to NeurIPS</a>. This represents 30% annual growth since 2017, when around 3,000 papers were submitted. This growth comes from all sorts of places. More companies are doing AI research. More universities are doing AI research. More <em>people</em> are getting into AI research.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xoUu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xoUu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png 424w, https://substackcdn.com/image/fetch/$s_!xoUu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png 848w, https://substackcdn.com/image/fetch/$s_!xoUu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png 1272w, https://substackcdn.com/image/fetch/$s_!xoUu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xoUu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png" width="372" height="230.71153846153845" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:903,&quot;width&quot;:1456,&quot;resizeWidth&quot;:372,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!xoUu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png 424w, https://substackcdn.com/image/fetch/$s_!xoUu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png 848w, https://substackcdn.com/image/fetch/$s_!xoUu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png 1272w, https://substackcdn.com/image/fetch/$s_!xoUu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38bd1181-4e34-476e-838b-5a940c1905f7_1642x1018.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption">The number of submitted papers to NeurIPS, the most popular AI conference, over the last fifteen years. It&#8217;s nearing exponential growth territory.  (Data from papercopilot.com)</figcaption></figure></div><p>The number of submitted papers to NeurIPS, the most popular AI conference, over the last fifteen years. It&#8217;s nearing exponential growth territory. (Data from <a href="http://papercopilot.com">papercopilot.com</a>)</p><p>With more people writing more papers about more different topics we&#8217;d expect some <em>explosive</em> growth. And yet I&#8217;d argue that <strong>the pace of progress feels pretty constant</strong>, if not slowing down by a small margin. This is a difficult thing to measure, so I polled a bunch of researchers:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SFl8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SFl8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png 424w, https://substackcdn.com/image/fetch/$s_!SFl8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png 848w, https://substackcdn.com/image/fetch/$s_!SFl8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png 1272w, https://substackcdn.com/image/fetch/$s_!SFl8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SFl8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png" width="366" height="245.63758389261744" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:800,&quot;width&quot;:1192,&quot;resizeWidth&quot;:366,&quot;bytes&quot;:308694,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/165548690?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SFl8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png 424w, https://substackcdn.com/image/fetch/$s_!SFl8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png 848w, https://substackcdn.com/image/fetch/$s_!SFl8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png 1272w, https://substackcdn.com/image/fetch/$s_!SFl8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79053e31-32a1-4789-98d8-e02bfd765c51_1192x800.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If we take the mean response here we find that progress is moving faster &#8211; but only a bit. This should be surprising in an area that&#8217;s seen explosive growth along just about every dimension.</p><h2>Expanding the space of everything</h2><p>If we&#8217;re publishing exponentially more, why aren&#8217;t we learning exponentially more? One possible explanation is that the space of possible ideas we&#8217;re exploring has stayed a constant size, but the number of <em>explorers</em> has grown at a near-exponential pace. If this is an accurate model, then the world of research looks something like this:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8guq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8guq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png 424w, https://substackcdn.com/image/fetch/$s_!8guq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png 848w, https://substackcdn.com/image/fetch/$s_!8guq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png 1272w, https://substackcdn.com/image/fetch/$s_!8guq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8guq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png" width="550" height="356.21565934065933" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:943,&quot;width&quot;:1456,&quot;resizeWidth&quot;:550,&quot;bytes&quot;:589031,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/165548690?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8guq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png 424w, https://substackcdn.com/image/fetch/$s_!8guq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png 848w, https://substackcdn.com/image/fetch/$s_!8guq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png 1272w, https://substackcdn.com/image/fetch/$s_!8guq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50bfce50-1260-4174-ab1e-b72aaed879fa_1554x1006.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If the number of researchers has increased by 10x, but amount of space is only slightly larger, then the current landscape of research looks something like the bubble on the right &#8211; but at a larger scale. And maybe packed tighter.</p><p>This mental model explains why many AI researchers report experiencing <a href="https://kyunghyuncho.me/i-sensed-anxiety-and-frustration-at-neurips24/">anxiety and frustration</a> around the current research and publication process. There are so many people doing so many things that it&#8217;s simply hard to find any open space to occupy.</p><p>I think this anxiety also comes from a mindset that the amount of &#8216;research territory&#8217; is constant, and good research will come from interpolating between existing ideas, rather than pushing forward into the unknown. This is exactly what Hamming warns against.</p><p>And this also means that <strong>the best research expands the space</strong> of territory available to us to explore. And yet clearly most research is doing something different &#8211; most papers aren&#8217;t groundbreaking. They&#8217;re comfortably nestled somewhere in high-dimensional space, in some unoccupied pigeonhole between things that already exist.</p><p>Let&#8217;s think of research ideas in physical terms. New research can be territory-expanding, in that it opens up more space for other researchers than it consumes, or territory-consuming, in that it `eats' some of the available territory without giving anything back.</p><p>The best research is territory-expanding. It's generous. It leaves things to the imagination, sparks more ideas than it explores. It&#8217;s explorative rather than exploitative. The most important research <em>always</em> has this property, where it expands space more than occupies it.</p><p>So why aren&#8217;t more people doing this kind of research?</p><h2>Good research takes ambition</h2><p>My theory for why most research isn&#8217;t territory-expanding is because the odds of <strong>producing something groundbreaking with a project are directly proportional to its chance of failure</strong>. This means that if you set out to change the world, you will probably fail. The converse here is that if you set your sights low, and choose a topic that&#8217;s more likely to bear fruit in some way, then your chances of `success' (publishing <em>something</em>) are higher, but the chances of it expanding the space of knowledge are smaller.</p><p>And thus, we end up with 25,000 NeurIPS submissions. This is what happens when the community optimizes too strongly for the number of papers submitted and too loosely for solving big problems.</p><p>I also want to recognize that for some groups, external factors are at play. I&#8217;ve been told that you need to publish at least one paper to get into a CS graduate program these days &#8211; meaning there are many people who just need to publish <em>something</em>, even if it isn&#8217;t their best work. And some tech companies prevent employees from publish findings that feel too important, giving their workers have a perverse incentive to make their research sound small to give their work a chance to see the light of day. Let&#8217;s ignore both of these edge cases.</p><h2>Higher ambition means a higher rate of failure</h2><p>One question every researcher should ask: what are you optimizing for?</p><p>If your goal as a researcher is to publish as much as possible and reach the maximum h-index then logic says to lower your ambition to the minimum possible level such that you can still publish. After a bit of practice, a good researcher with the singular goal of publishing can often write papers every few months. You&#8217;ll see this happen. Some conferences even have PhD students publishing multiple papers in the same conference indicating that they worked on multiple projects at the same time (I&#8217;m not pointing any fingers, as I have also done this once. Also, it can happen for different timeline-related reasons.)</p><p>If your goal is to <em>move the field forward</em> then you should plan differently. Raise your level of ambition. Try  crazy things. </p><p>A consequence of this is that <em>you will publish less</em>. Deadlines will come around for which you have nothing to show. Perhaps you found out that your idea has already been done, or doesn&#8217;t work at scale, or is doing things that you just don&#8217;t understand. The most likely explanation is that most great things take more than three months to execute.</p><p>(As an aside, I have been doing research for six years or so and have already found that periods without publishing are really just fine. People care most about the one or two projects you&#8217;ve ever worked on that are most exciting. They care much less about the six month gaps where you were quietly trying to make something work. In my own career this has already happened several times: I spent my year in the Google AI Residency working on something that never quiet worked. I worked on a music model for six months that never went anywhere. In the third year of my PhD I wanted to build a <a href="https://arxiv.org/abs/2410.02525">contextual embedding model</a>, which was a neat idea but took over a year to get the details right. The list does not stop there&#8230;)</p><h2>Most research doesn&#8217;t matter much in the long run</h2><p>But if you look back at the conference proceedings from ten years ago, they don&#8217;t feel exceptionally relevant:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!J-xQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!J-xQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png 424w, https://substackcdn.com/image/fetch/$s_!J-xQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png 848w, https://substackcdn.com/image/fetch/$s_!J-xQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png 1272w, https://substackcdn.com/image/fetch/$s_!J-xQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!J-xQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png" width="630" height="412.3557692307692" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:953,&quot;width&quot;:1456,&quot;resizeWidth&quot;:630,&quot;bytes&quot;:1867854,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.jxmo.io/i/165548690?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!J-xQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png 424w, https://substackcdn.com/image/fetch/$s_!J-xQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png 848w, https://substackcdn.com/image/fetch/$s_!J-xQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png 1272w, https://substackcdn.com/image/fetch/$s_!J-xQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd876e04b-2cc5-41f8-92eb-5fea5820f2e8_2048x1341.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">A screenshot of the proceedings from <a href="https://papers.nips.cc/paper_files/paper/2015">NeurIPS 2015</a>.  Many of these papers did not stand the test of time.</figcaption></figure></div><p>These are interesting and important problems, but have almost nothing to do with the AI systems we use today. This makes one wonder how much of the content from this year&#8217;s 25,000 NeurIPS papers might still be relevant in 2035.</p><p>You might have your own hypotheses on why academic research tends toward small-but-eventually-forgotten problems. Predicting the future is hard. And most research moves in cycles that last less than a year, so it&#8217;s important to choose problems that are tractable within that timeframe.</p><h3>Research that stands the test of time</h3><p>For our purposes, we&#8217;re talking about research that affects the way we think about or design modern AI systems. This is an exceptionally small sliver of research papers. Under this definition, I&#8217;d guess <strong>less than 1% of published AI research &#8216;matters&#8217;.</strong></p><p>Why is this number so small? One explanation is that modern AI grew out of machine learning and probabilistic modeling, which is a much deeper subfield with many sub-areas of study that have nothing to do with modern AI. Here&#8217;s one example: if you took statistics in school, you know that <a href="https://en.wikipedia.org/wiki/Foundations_of_statistics">modern statistics is divided into two schools (Bayesian and frequentist)</a> but modern AI is totally frequentist. Even though <a href="https://proceedings.mlr.press/v37/hernandez-lobatoc15.html">lots</a> <a href="https://ieeexplore.ieee.org/abstract/document/9756596">of</a> <a href="https://proceedings.mlr.press/v139/izmailov21a.html">people</a> <a href="https://link.springer.com/chapter/10.1007/978-3-030-42553-1_3">have</a> <a href="https://proceedings.neurips.cc/paper/2016/hash/a96d3afec184766bfeca7a9f989fc7e7-Abstract.html">researched</a> <a href="https://arxiv.org/abs/1903.05779">Bayesian</a> <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=LHHrBwAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;dq=bayesian+neural+networks&amp;ots=K6xgRO4x-a&amp;sig=tRHTOY4SxZVKub-vuMO7_psMa_o#v=onepage&amp;q=bayesian%20neural%20networks&amp;f=false">neural</a> <a href="https://seunghan96.github.io/assets/pdf/BNN/paper/05.Ensemble%20Learning%20in%20Bayesian%20Neural%20Networks.pdf">networks</a>, they&#8217;ve basically been abandoned in favor of learning everything from data without a prior.</p><p>So what makes up this 1%? What research matters? To answer this question, let&#8217;s first gather a bit of data. NeurIPS (the biggest AI conference) gives &#8220;Test of Time&#8221; awards to ten-year-old papers that have made an outsized impact over the preceding decade. I asked Deep Research to gather me a list of the papers that have won the award:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2P0F!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2P0F!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png 424w, https://substackcdn.com/image/fetch/$s_!2P0F!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png 848w, https://substackcdn.com/image/fetch/$s_!2P0F!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png 1272w, https://substackcdn.com/image/fetch/$s_!2P0F!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2P0F!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png" width="506" height="695.0446096654275" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1478,&quot;width&quot;:1076,&quot;resizeWidth&quot;:506,&quot;bytes&quot;:355768,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!2P0F!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png 424w, https://substackcdn.com/image/fetch/$s_!2P0F!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png 848w, https://substackcdn.com/image/fetch/$s_!2P0F!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png 1272w, https://substackcdn.com/image/fetch/$s_!2P0F!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef9451b1-73c2-406a-a0d9-48f5019d3913_1076x1478.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Papers that have won the &#8220;Test of Time&#8221; award at NeurIPS, prepared by Deep Research.</figcaption></figure></div><p>Out of the ten papers that have won the Test-of-Time award, I notice the following themes:</p><ul><li><p>These are simpler than the average NeurIPS paper</p></li><li><p>Each of them seems to leverage more data and more compute than was common at the time</p></li></ul><p>And let&#8217;s remember that basically all deep learning research changed with the ImageNet paper in 2012, which only won its test-of-time award in 2022. So only the awards from 2022&#8211;2024 reflect research from the &#8220;modern&#8221; (post-AlexNet) era.</p><p>We can clearly see that in the long run, the AI research that matters is research that (a) is simple and (b) <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">scales well</a>. In fact, <strong>modern AI systems like Claude 4 and Veo 3 are so simple that they use almost none of the decades of published ML research</strong>. And yet this is likely what Hamming would have predicted &#8211; in the end, most science is forgotten.</p><p>This is probably a good thing. Downstream effects are that AI is easy for newcomers to learn, and relatively simple to implement. And it seems quite reasonable to bet that the next big breakthroughs in AI will be simple.</p><p>Remind yourself that simple questions are often the most ambitious. They often also take the longest to mentally grapple with, and are the hardest to engineer experiments for.</p><h3>Putting Hamming&#8217;s advice into practice</h3><p>I suspect that the problem here isn&#8217;t lack of ambitious ideas, it&#8217;s more related to a fear of failure. Most people that have done research for a long time confront big, messy, unanswered questions that continually resurface through day-to-day experimentation.</p><p>It seems the best strategy is to do whatever you can to can to <strong>confront those questions directly</strong>. There is no question too big to make into a research project. Einstein famously used the thought experiment of riding along on a beam of light to explore the ideas that ended up revolutionizing modern physics. What&#8217;s your question like this?</p><p>Perhaps you want to imagine your own personal data being stored inside an LLM, or what it would be like to slide down a gradient during training, or some new mechanism for visualizing and manipulating the internals of an LLM as it processes a sequence along during inference.</p><p>Hamming also notes that the best way to move forward is often by talking to capable people that you trust. He specifically mentions that solo daily brainstorming sessions haven&#8217;t proved fruitful in his career. The best way to make progress, he says, is to talk to someone and explain where you&#8217;re starting from all the way up to the point where you get stuck.</p><p>Maybe the specific question isn&#8217;t even that important, since so few people seem to be thinking like this. The bottom line is that we&#8217;d all be better off if most people worked on big ambitious projects for a longer time. We would see fewer papers, but each individual paper would have more depth. By publishing less, we would move faster.</p><div><hr></div><p>Thanks to Ege Erdil for Jeffrey Emanuel for feedback on an early draft of this post.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[There Are No New Ideas in AI… Only New Datasets]]></title><description><![CDATA[LLMs were invented in four major developments... all of which were datasets]]></description><link>https://blog.jxmo.io/p/there-are-no-new-ideas-in-ai-only</link><guid isPermaLink="false">https://blog.jxmo.io/p/there-are-no-new-ideas-in-ai-only</guid><dc:creator><![CDATA[Jack Morris]]></dc:creator><pubDate>Wed, 09 Apr 2025 21:32:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!56cS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Most people know that AI has made unbelievable progress over the last fifteen years&#8211; especially in the last five. It might feel like that progress is <em>*inevitable*</em> &#8211; although large paradigm-shift-level breakthroughs are uncommon, we march on anyway through a stream of slow &amp; steady progress. In fact, some researchers have recently declared a <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">&#8220;Moore&#8217;s Law for AI&#8221;</a> where the computer&#8217;s ability to do certain things (in this case, certain types of coding tasks) increases exponentially with time:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!56cS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!56cS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png 424w, https://substackcdn.com/image/fetch/$s_!56cS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png 848w, https://substackcdn.com/image/fetch/$s_!56cS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png 1272w, https://substackcdn.com/image/fetch/$s_!56cS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!56cS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png" width="720" height="430" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:430,&quot;width&quot;:720,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Length of asks AIs can do is doubling every 7 months&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Length of asks AIs can do is doubling every 7 months" title="Length of asks AIs can do is doubling every 7 months" srcset="https://substackcdn.com/image/fetch/$s_!56cS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png 424w, https://substackcdn.com/image/fetch/$s_!56cS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png 848w, https://substackcdn.com/image/fetch/$s_!56cS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png 1272w, https://substackcdn.com/image/fetch/$s_!56cS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6c8b571-bdbe-46cc-aa5c-8fd5e5555b01_720x430.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">the proposed &#8220;Moore&#8217;s Law for AI&#8221;. (by the way, anyone who thinks they can run an autonomous agent for an hour with no intervention as of April 2025 is fooling themselves)</figcaption></figure></div><p>Although I don&#8217;t really agree with this specific framing for a number of reasons, I can&#8217;t deny the trend of progress. Every year, our AIs get a little bit smarter, a little bit faster, and a little bit cheaper, with no end in sight.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Most people think that this continuous improvement comes from a steady supply of ideas from the research community across academia &#8211; mostly MIT, Stanford, CMU &#8211; and industry &#8211; mostly Meta, Google, and a handful of Chinese labs, with lots of research done at other places that we&#8217;ll never get to learn about.</p><p>And we certainly have made a lot of progress due to research, especially on the systems side of things. This is how we&#8217;ve made models cheaper in particular. Let me cherry-pick a few notable examples from the last couple years:</p><p>- in 2022 Stanford researchers gave us <a href="https://arxiv.org/abs/2205.14135)">FlashAttention</a>, a better way to utilize memory in language models that&#8217;s used literally everywhere;</p><p>- in 2023 Google researchers developed <a href="https://arxiv.org/abs/2211.17192">speculative decoding</a>, which all model providers use to speed up inference (also developed at <a href="https://arxiv.org/pdf/2302.01318">DeepMind</a>, I believe concurrently?)</p><p>- in 2024 a ragtag group of internet fanatics developed <a href="https://kellerjordan.github.io/posts/muon/">Muon</a>, which seems to be a better optimizer than SGD or Adam and may end up as the way we train language models in the future</p><p>- in 2025 DeepSeek released <a href="https://arxiv.org/abs/2501.12948">DeepSeek-R1</a>, an open-source model that has equivalent reasoning power to similar closed-source models from AI labs (specifically Google and OpenAI)</p><p>So we&#8217;re definitely figuring stuff out. And the reality is actually cooler than that: we&#8217;re engaged in a decentralized globalized exercise of Science, where findings are shared openly on <a href="https://arxiv.org/">ArXiv</a> and at conferences and on social media and every month we&#8217;re getting incrementally smarter.</p><p>If we&#8217;re doing so much important research, why do some argue that progress is slowing down? <a href="https://www.lesswrong.com/posts/4mvphwx5pdsZLMmpY/recent-ai-model-progress-feels-mostly-like-bullshit">People are still complaining</a>. The two most recent huge models, <a href="https://x.ai/news/grok-3">Grok 3</a> and <a href="https://openai.com/index/introducing-gpt-4-5/">GPT-4.5</a>, only obtained a marginal improvement on capabilities of their predecessors. In one particularly salient example, when <a href="https://arxiv.org/abs/2503.21934v1">language models were evaluated on the latest math olympiad exam</a>, they scored only 5%, indicating that <a href="https://cdn.openai.com/o1-system-card-20241205.pdf">recent announcements may have been overblown</a> when reporting system ability.</p><p>And if we try to chronicle the <em>*big*</em> breakthroughs, the real paradigm shifts, they seem to be happening at a different rate. Let me go through a few that come to mind:</p><h3>LLMs in four breakthroughs</h3><p>1. Deep neural networks: Deep neural networks first took off after the <a href="https://www.notion.so/There-Are-No-New-Ideas-in-AI-Only-New-Data-1cf5109a45d880e6b0d5d6e3a4ba2fdc?pvs=21">AlexNet model</a> won an image recognition competition in 2012</p><p>2. Transformers + LLMs: in 2017 Google proposed transformers in <a href="https://arxiv.org/abs/1706.03762">Attention Is All You Need</a>, which led to <a href="https://arxiv.org/abs/1810.04805">BERT</a> (Google, 2018) and the original <a href="https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf">GPT</a> (OpenAI, 2018)</p><p>3. RLHF: first proposed for LLMs in the <a href="https://arxiv.org/abs/2203.02155">InstructGPT paper</a> <s>from OpenAI in 2022</s>*<br><strong>* Correction: Paul Christiano <a href="https://arxiv.org/abs/1706.03741">wrote a paper on RLHF in 2017</a>; the concept of aligning models to human preferences is much older. Thanks to everyone who pointed this out!</strong></p><p>4. Reasoning: in 2024 OpenAI released O1, which led to DeepSeek R1</p><p>If you squint just a little, these four things (DNNs &#8594; Transformer LMs &#8594; RLHF &#8594; Reasoning) summarize everything that&#8217;s happened in AI. We had DNNs (mostly image recognition systems), then we had text classifiers, then we had chatbots, now we have reasoning models (whatever those are).</p><p>Say we want to make a fifth such breakthrough; it could help to study the four cases we have here. What new research ideas led to these groundbreaking events?</p><p>It&#8217;s not crazy to argue that <strong>all the underlying mechanisms of these breakthroughs existed in the 1990s,</strong> if not before. We&#8217;re applying relatively simple neural network architectures and doing either supervised learning (1 and 2) or reinforcement learning (3 and 4).</p><p>Supervised learning via cross-entropy, the main way we pre-train language models, emerged from Claude Shannon&#8217;s work in the 1940s.</p><p>Reinforcement learning, the main way we post-train language models via RLHF and reasoning training, is slightly newer. It can be traced to the <a href="https://people.cs.umass.edu/~barto/courses/cs687/williams92simple.pdf">introduction of policy-gradient methods in 1992</a> (and these ideas were certainly around for the first edition of the Sutton &amp; Barto &#8220;Reinforcement Learning&#8221; textbook in 1998).</p><h3>If our ideas aren&#8217;t new, then what is?</h3><p>Ok, let&#8217;s agree for now that these &#8220;major breakthroughs&#8221; were arguably fresh applications of things that we&#8217;d known for a while. First of all &#8211; this tells us something about the <em>*next*</em> major breakthrough (that &#8220;secret fifth thing&#8221; I mentioned above). Our breakthrough is probably not going to come from a completely new idea, rather it&#8217;ll be the resurfacing of something we&#8217;ve known for a while.</p><p>But there&#8217;s a missing piece here: each of these four breakthroughs <strong>enabled us to learn from a new data source:</strong></p><p>1. AlexNet and its follow-ups unlocked <a href="http://(https://www.image-net.org/">ImageNet</a>, a large database of class-labeled images that drove fifteen years of progress in computer vision</p><p>2. Transformers unlocked training on &#8220;The Internet&#8221; and a race to download, categorize, and parse all the text on <a href="https://arxiv.org/abs/2101.00027">The Web</a> (which <a href="https://www.lesswrong.com/posts/6Fpvch8RR29qLEWNH/chinchilla-s-wild-implications">it seems</a> <a href="https://arxiv.org/abs/2305.16264">we&#8217;ve mostly done</a> <a href="https://arxiv.org/abs/2305.13230">by now</a>)</p><p>3. RLHF allowed us to learn from human labels indicating what &#8220;good text&#8221; is (mostly a vibes thing)</p><p>4. Reasoning seems to let us learn from <a href="http://incompleteideas.net/IncIdeas/KeytoAI.html">&#8220;verifiers&#8221;</a>, things like calculators and compilers that can evaluate the outputs of language models</p><p>Remind yourself that each of these milestones marks the first time the respective data source (ImageNet, The Web, Humans, Verifiers) was used at scale. Each milestone was followed by a frenzy of activity: researchers compete to (a) siphon up the remaining useful data from any and all available sources and (b) make better use of the data we have through new tricks to make our systems more efficient and less data-hungry. (I expect we&#8217;ll see this trend in reasoning models throughout 2025 and 2026 as researchers compete to find, categorize, and verify everything that might be verified.)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!U_q8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!U_q8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg 424w, https://substackcdn.com/image/fetch/$s_!U_q8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg 848w, https://substackcdn.com/image/fetch/$s_!U_q8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!U_q8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!U_q8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg" width="750" height="300" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:300,&quot;width&quot;:750,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;How to train and validate on Imagenet&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="How to train and validate on Imagenet" title="How to train and validate on Imagenet" srcset="https://substackcdn.com/image/fetch/$s_!U_q8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg 424w, https://substackcdn.com/image/fetch/$s_!U_q8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg 848w, https://substackcdn.com/image/fetch/$s_!U_q8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!U_q8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ddd3045-3daa-4adf-8d06-c75b3ac6f436_750x300.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Progress in AI may have been inevitable once we gathered <a href="https://www.image-net.org/">ImageNet</a>, at the time the largest public collection of images from the Web</figcaption></figure></div><h3>How much do new ideas matter?</h3><p>There&#8217;s something to be said for the fact that our actual technical innovations may not make a huge difference in these cases. Examine the counterfactual. If we hadn&#8217;t invented AlexNet, maybe another architecture would have come along that could handle ImageNet. If we never discovered Transformers, perhaps we would&#8217;ve settled with LSTMs or SSMs or found something else entirely to learn from the mass of useful training data we have available on the Web.</p><p>This jibes with the theory some people have that nothing matters but data. Some researchers have observed that for all the training techniques, modeling tricks, and hyperparameter tweaks we make, the thing that makes the biggest difference by-and-large is changing the data.</p><p>As one salient example, some researchers worked on <a href="https://arxiv.org/abs/2212.10544">developing a new BERT-like model using an architecture other than transformers</a>. They spent a year or so tweaking the architecture in hundreds of different ways, and managed to produce a different type of model (this is a state-space model or &#8220;SSM&#8221;) that performed about equivalently to the original transformer when trained on the same data.</p><p>This discovered equivalence is really profound because it hints that <em>*there is an upper bound to what we might learn from a given dataset*</em>. All the training tricks and model upgrades in the world won&#8217;t get around the cold hard fact that there is only so much you can learn from a given dataset.</p><p>And maybe this apathy to new ideas is what we were supposed to take away from <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">The Bitter Lesson</a>. If data is the only thing that matters, why are 95% of people working on new methods?</p><h3>Where will our next paradigm shift come from? <em>*(YouTube&#8230;maybe?)</em></h3><p>The obvious takeaway is that our next paradigm shift isn&#8217;t going to come from an improvement to RL or a fancy new type of neural net. It&#8217;s going to come when we unlock a source of data that we haven&#8217;t accessed before, or haven&#8217;t properly harnessed yet.</p><p>One obvious source of information that a lot of people are working towards harnessing is video. According to <a href="https://www.dexerto.com/entertainment/how-many-videos-are-there-on-youtube-2197264/">a random site on the Web</a>, about 500 hours of video footage are uploaded to YouTube *per minute*. This is a ridiculous amount of data, much more than is available as text on the entire internet. It&#8217;s potentially a much richer source of information too as videos contain not just words but the inflection behind them as well as rich information about physics and culture that just can&#8217;t be gleaned from text.</p><p>It&#8217;s safe to say that as soon as our models get efficient enough, or our computers grow beefy enough, Google is going to start training models on YouTube. They own the thing, after all; it would be silly not to use the data to their advantage.</p><p>A final contender for the next &#8220;big paradigm&#8221; in AI is a data-gathering systems that some way <em>embodied</em>&#8211; or, in the words of a regular person, robots. We&#8217;re currently not able to gather and process information from cameras and sensors in a way that&#8217;s amenable to training large models on GPUs. If we could build smarter sensors or scale our computers up until they can handle the massive influx of data from a robot with ease, we might be able to use this data in a beneficial way.</p><p>It&#8217;s hard to say whether YouTube or robots or something else will be the Next Big Thing for AI. We seem pretty deeply entrenched in the camp of language models right now, but we also seem to be running out of language data pretty quickly. But if we want to make progress in AI, maybe we should stop looking for new ideas, and start looking for new data.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.jxmo.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[CoPilot for Everything]]></title><description><![CDATA[Our employers have all the data they need to train AI models to replace us. How long will it be until this actually happens?]]></description><link>https://blog.jxmo.io/p/copilot-for-everything</link><guid isPermaLink="false">https://blog.jxmo.io/p/copilot-for-everything</guid><dc:creator><![CDATA[Jack Morris]]></dc:creator><pubDate>Fri, 28 Feb 2025 13:45:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!37xC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Between 2020 and 2021 I worked full-time at Google. Although my original plan was to work from Google headquarters in Mountain View, California, due to an international pandemic, my year-long job ended up being fully remote from start to finish. I received my laptop in the mail before my first day and sent it back in a box when I left the company. </p><p>During this stint I never set foot on the Google campus. Every contribution I made was digital: a series of keypresses, touchpad movements, and mouse clicks made from my official Google laptop. I also provided audio and video inputs via my laptop&#8217;s camera and microphone during video meetings. But no one ever saw me in person. </p><p>It only recently dawned on me that my actions may have been recorded. Of course, certain things I already knew *were* recorded: all the lines of code I submitted to the Google monorepository, for example, are probably still there. I&#8217;d imagine my emails are stored somewhere too, as well as the notes I wrote via Google docs. But what about the rest of my actions on the company computer? </p><p>It&#8217;s entirely possible that the entire daily work process was documented. In theory, my employer had a right to collect every input I provided on my company computer. Every mouse click, every keypress. And they could have sent this all back to store in some data warehouse. This is a much richer data type than simply the lines of code I outputted: these are *behavioral traces* that define how I solve problems from start to finish. </p><p>As I&#8217;ll explain later in the post, the thing that scares me about the existence of this data is that it seems well within the capabilities of current technology to train a model that can replicate *me*, in some sense. I&#8217;m calling it as a *Copilot for Everything*: an assistant that can auto-complete entire tasks for me, based on the actions I&#8217;ve taken at work in the past. And this feels like such an economically useful tool that it would be crazy for it *not* to happen in the next few years. </p><p>I don&#8217;t intend to pick on Google specifically, and I don&#8217;t know whether they do this or even what their policies are; I&#8217;m just trying to use my own experience in the corporate world to speculate on what I imagine will be a much bigger issue in the future. <strong>With today&#8217;s technology, it is totally plausible that a corporation might train a large AI model to mimic the digital work of an employee, (</strong>even after that employee has left the company)<strong>.</strong> Is it ethical for my former employer to use my prior work output to create a digital version of me?  Is it ethical for someone&#8217;s *current* employer to do this?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!37xC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!37xC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png 424w, https://substackcdn.com/image/fetch/$s_!37xC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png 848w, https://substackcdn.com/image/fetch/$s_!37xC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png 1272w, https://substackcdn.com/image/fetch/$s_!37xC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!37xC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png" width="1250" height="833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:833,&quot;width&quot;:1250,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:706875,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://jxmnop.substack.com/i/158101095?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!37xC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png 424w, https://substackcdn.com/image/fetch/$s_!37xC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png 848w, https://substackcdn.com/image/fetch/$s_!37xC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png 1272w, https://substackcdn.com/image/fetch/$s_!37xC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88499ba2-78c2-450c-ae68-55608682e52c_1250x833.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Automating remote work with supervised learning </h3><p>Behavioral traces reduce the concept of automating work output to a well-defined learning problem for AI. The model inputs would be all the inputs the computer provides me: pixels on the screen, maybe audio. Outputs are the &#8220;actions&#8221; that I took on my computer: keystrokes, mouse movements, clicks. </p><p>At a large enough scale, it seems entirely feasible to train a large model that can predict actions from computer inputs. This is doable with the same current technology that powers large language models: supervised learning, which would allow learning to predict actions from computer inputs, and transformers, a type of neural network that excels at learning from large amounts of data. </p><p>From the company&#8217;s perspective, the process would be simple. Employer records you doing your work for some amount of time; employer trains model to replicate your output for your most monotonous and boring tasks. Or perhaps the company takes all the data they have and trains a model on the aggregate of *all* employees outputs. That would be closer to what we&#8217;ve seen work well in other domains, like vision and language, where <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">training on more data is usually the right answer</a>. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Krur!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Krur!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Krur!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Krur!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Krur!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Krur!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Krur!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Krur!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Krur!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Krur!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F958a286a-ad26-4e9e-9ad3-ab3d73b7475b_1800x1200.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Can imperfect models of us improve our productivity?</h3><p>So that tells us how a company could train a model to mimic employees&#8217; actions. But how do we use these models? If we think about the resulting artifact, it wouldn&#8217;t be a drop-in replacement for an employee: instead it&#8217;d be a probabilistic model that tells us the likelihood of a given action given some input. First of all, these models have issues. Neural networks are <a href="https://arxiv.org/abs/1312.6199">easily tricked by adversarial inputs</a> and generally don&#8217;t perform well if the inputs are too out-of-distribution. Neither of these are problems with humans. </p><p>Second, how do we do sampling? The obvious answer is to be &#8220;greedy&#8221; and take the most likely action at every step. This would work well in very low-entropy situations. For example, performing a task that&#8217;s identical to one that you&#8217;ve performed many times would be easy, and a good model could do it perfectly with no supervision. And in the case where the model has also learned from *other* employees&#8217; actions, if it were to perform a rote task that someone else had done many times, that would be easy too. </p><p>But what about higher-entropy situations? In the case of higher uncertainty, the model would output a less useful distribution over potential actions. In this case, greedy action-sampling could lead to &#8220;exposure bias&#8221; where the model does something weird, ends up in a situation it&#8217;s never seen before, and inevitably goes off the rails. </p><p>Even good models will sometimes end up in situations of significant uncertainty. What I&#8217;d imagine is that we&#8217;ll end up with a solution similar to what has worked in self-driving cars: the computer can &#8216;drive&#8217; by itself until it encounters something completely unexpected, at which point a human will intervene. Once the novel scenario is resolved and the task again resembles something that the computer knows how to do, the model can resume operation. </p><p>It&#8217;s hard to imagine what this might look from a user interface perspective, but I&#8217;ll suggest two basic options: &#8220;fast-forward&#8221; style and &#8220;foreman&#8221; style. </p><p>In a fast-forward setup, the assistant would perform certain tasks for you very quickly while you wait. In theory the model could still operate the computer at 100x speed; if input latency isn&#8217;t a limiting factor, the computer could complete tasks in the blink of an eye. Maybe a button on your screen will light up when the computer detects you&#8217;re about to perform a task that&#8217;s &#8220;understood&#8221; by the assistant; clicking the button would fast-forward through that task being accomplished at 100x speed. </p><p>In the foreman mode, one human would oversee *many* AIs working in parallel. This is only possible if the models can do most of their work without any interevention. In the foreman setup, one person would be responsible for the success of several AI assistants, and jump in to help them when the task reaches a certain level of uncertainty. This is similar to <a href="https://www.nytimes.com/interactive/2024/09/03/technology/zoox-self-driving-cars-remote-control.html">Zoox operating stations</a> where human operators remotely pilot cars out of unexpected situations from hundreds of miles away.</p><h3>What does this mean for us?</h3><p>Even if none of the exact futures I&#8217;ve imagined here play out, note the underlying theme: **progress increases the average entropy of digital work**. This principle has has nothing to do with neural networks. As we develop better tools, we&#8217;re able to do repetitive tasks more quickly. Better abstractions reduce the amount of input it takes for us to generate the same amount of output. One way to do this, the one that&#8217;s discussed here, is directly modeling employee behavior with big neural networks.</p><p>The reality is that every remote worker&#8217;s job tends to be repetitive sometimes. And your value comes from how you handle the *least* repetitive things you do: reacting to novel situations, adapting to change. New tools will reduce the amount of repetition in your work. Our day-to-day jobs will become *less* predictable as our most mundane and monotonous tasks are modeled away.</p><p>Given all this &#8212; how can you make yourself indispensable? The answer is by fighting entropy. If the most efficient company is the one where its employees are doing the least repetitive thing at all times, then the most productive employee is the one that&#8217;s the least modelable. Focus on the things that are hard to gather training data for.</p>]]></content:encoded></item><item><title><![CDATA[Please Stop Talking About AGI]]></title><description><![CDATA[Why I think Yann Lecun was right about LLMs (but perhaps only by accident)]]></description><link>https://blog.jxmo.io/p/we-should-stop-talking-about-agi</link><guid isPermaLink="false">https://blog.jxmo.io/p/we-should-stop-talking-about-agi</guid><dc:creator><![CDATA[Jack Morris]]></dc:creator><pubDate>Fri, 21 Feb 2025 18:03:35 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!N3r5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It&#8217;s become very popular over the last few years to speculate how close society might be to <em>Artificial General Intelligence </em>(AGI). What AGI actually means is murky, and often-debated, but mentioning AGI is usually a good jumping-off point for discussions of future artificial intelligences&#8217; capabilities. Many following the field maintain <em>AGI timelines</em>, rigorous guesses for the probability of this mythical intelligence to emerge at future points. Those in the know might ask you for your timelines over coffee, classifying them as &#8220;long&#8221; &#8211; that it might be a decade or two before AIs are smart enough to take all of our jobs &#8211; or &#8220;short&#8221; &#8211; that it could happen any day now.</p><p>This isn&#8217;t the most useful way of thinking about the progression of AI capabilities. The existence of a <em>timeline</em> implies AGI has a rigorous definition and can be measured. It also implies that AGI is inevitable, the only question being when it will arrive.</p><p>What I see is not a march towards complete general intelligence, but rather a trend of increasing AI productivity per unit of human input. This trend holds across many disparate applications. Our AIs can label more data, write more code, do more math, as well as drive cars and pilot planes for longer with less intervention from us. It may be possible that we&#8217;ll <em>never</em> reach a point where AIs can run forever, uninterrupted, without human guidance. Rather we&#8217;re pushing the boundary of how much we can get for what we give.</p><p>Instead of talking about the mythical final frontier of AGI, I think we should start thinking more realistically and measuring the ratio of human input per useful AI output.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!N3r5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!N3r5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png 424w, https://substackcdn.com/image/fetch/$s_!N3r5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png 848w, https://substackcdn.com/image/fetch/$s_!N3r5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png 1272w, https://substackcdn.com/image/fetch/$s_!N3r5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!N3r5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png" width="579" height="332.0735294117647" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:702,&quot;width&quot;:1224,&quot;resizeWidth&quot;:579,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!N3r5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png 424w, https://substackcdn.com/image/fetch/$s_!N3r5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png 848w, https://substackcdn.com/image/fetch/$s_!N3r5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png 1272w, https://substackcdn.com/image/fetch/$s_!N3r5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F295078c0-2a16-4d53-8ba7-0d9cb5ce4fb5_1224x702.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">What will the future trend of human input per AI output look like?</figcaption></figure></div><p>Imagine for a moment the curve of how much we input have to provide for a unit of economic value the computer produces, and how this has changed over time. A very rough estimate is pictured above; one important open question is whether we&#8217;re approaching some unknowable carrying capacity, or if this figure will eventually decay to zero. (If this happens, it means that computers will be able to produce economic value with <em>zero human input</em>. This would be a frightening outcome.)</p><p>To understand what I mean better, let&#8217;s take a trip back in time to 2017&#8230;</p><h4><strong>We&#8217;ve seen this before (in self-driving cars)</strong></h4><p>If you&#8217;re new to the AI field, you should know that before language models, there was a previous AI craze circa 2017: the rise (and fall?) of the self-driving car.</p><p>If you&#8217;re not new to AI, let me remind you.</p><p>Around that time, several companies declared that within a year they would have Fully Self-Driving cars. Billions of dollars were raised. Millions of miles were driven. Many companies were founded, some of which eventually went bankrupt.</p><p>And years later, we&#8217;re still not quite at FSD. Teslas certainly can&#8217;t drive themselves; Waymos mostly can, within a pre-mapped area, but still have issues and intermittently require human intervention.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SgrO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SgrO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png 424w, https://substackcdn.com/image/fetch/$s_!SgrO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png 848w, https://substackcdn.com/image/fetch/$s_!SgrO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png 1272w, https://substackcdn.com/image/fetch/$s_!SgrO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SgrO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png" width="606" height="346.79296875" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b175eec3-875e-4f90-869e-adcd945434d9_1024x586.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:586,&quot;width&quot;:1024,&quot;resizeWidth&quot;:606,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!SgrO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png 424w, https://substackcdn.com/image/fetch/$s_!SgrO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png 848w, https://substackcdn.com/image/fetch/$s_!SgrO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png 1272w, https://substackcdn.com/image/fetch/$s_!SgrO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb175eec3-875e-4f90-869e-adcd945434d9_1024x586.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">In 2016, Tesla CEO Elon Musk promised that a Tesla would drive itself fully autonomously from Los Angeles to New York City by the end of the year. That still hasn&#8217;t happened. (Teslas are still sold with an optional &#8220;<a href="https://www.tesla.com/support/full-self-driving-subscriptions">Full Self-Driving</a>&#8221; subscription)</figcaption></figure></div><p>In response, the field has moved on from speculating the exact point cars will be fully-self driving. People instead discuss miles-per-disengagement (or miles-per-human-intervention). How far can the car drive without a human getting involved? This new lens gives us something that we can measure and track over time. Better technology gives us more miles driven per necessary human action.</p><p>What does the future look like for FSD? A recent report said Teslas can drive<a href="https://arstechnica.com/cars/2024/09/tesla-full-self-driving-requires-human-intervention-every-13-miles/"> thirteen miles per human intervention</a>; this estimate feels a little low to me, but still seems pretty good. We can certainly drive this number up with bigger models, faster inference, more data, and improved overall engineering.</p><p>A crucial question is whether with current technology, the miles-per-intervention number is bounded by some theoretical limit we don&#8217;t understand. We don&#8217;t know whether our models will keep getting better forever (approaching infinite miles driven with <em>no</em> interventions) or if there really is some amount of human intervention that will always be necessary.</p><h4><strong>Why Yann Lecun was wrong (kind of)</strong></h4><p>Now let&#8217;s apply this idea to today&#8217;s AI craze: language models.</p><p>A few years ago, Meta&#8217;s Chief AI Scientist Yann Lecun gave a talk about how language models won&#8217;t give us a direct path to human-level intelligence. He argued that because language models generate outputs token-by-token, and each token introduces a new probability of error, if we generate outputs that are too long, this per-token error will compound to inevitable failure.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cdjK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cdjK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png 424w, https://substackcdn.com/image/fetch/$s_!cdjK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png 848w, https://substackcdn.com/image/fetch/$s_!cdjK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png 1272w, https://substackcdn.com/image/fetch/$s_!cdjK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cdjK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png" width="580" height="301.328125" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:532,&quot;width&quot;:1024,&quot;resizeWidth&quot;:580,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!cdjK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png 424w, https://substackcdn.com/image/fetch/$s_!cdjK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png 848w, https://substackcdn.com/image/fetch/$s_!cdjK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png 1272w, https://substackcdn.com/image/fetch/$s_!cdjK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F033f75cc-0873-4ab5-854d-22368a4713aa_1024x532.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Yann presenting his &#8220;unpopular opinion&#8221;: an argument for why researchers shouldn&#8217;t work on language models. </figcaption></figure></div><p></p><p>Yann has used this simple argument to explain to the masses why we shouldn&#8217;t work on language models if we care about achieving human-level AI. He presents this problem of compounding errors as a critical flaw in language models themselves, something that can&#8217;t be overcome without switching away from the current autoregressive paradigm.</p><p>But this has turned out to be wrong. A few new AI systems (notably OpenAI o1/o3 line and Deepseek R1) contradict this theory. They are autoregressive language models, but actually get better by generating longer outputs:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nNys!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nNys!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png 424w, https://substackcdn.com/image/fetch/$s_!nNys!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png 848w, https://substackcdn.com/image/fetch/$s_!nNys!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png 1272w, https://substackcdn.com/image/fetch/$s_!nNys!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nNys!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png" width="464" height="294" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:294,&quot;width&quot;:464,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!nNys!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png 424w, https://substackcdn.com/image/fetch/$s_!nNys!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png 848w, https://substackcdn.com/image/fetch/$s_!nNys!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png 1272w, https://substackcdn.com/image/fetch/$s_!nNys!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5980e6b-9bf0-49a6-9a68-bccd703ab694_464x294.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">A graph from the<a href="http://arxiv.org/abs/2501.12948"> DeepSeek R1 report</a> showing how their system generates longer outputs as it gets smarter. This directly contradicts the contention that language models will eventually fail if we let them &#8220;think&#8221; for too long.</figcaption></figure></div><p>The finding that language models can get better by generating longer outputs directly contradicts Yann&#8217;s hypothesis. I think the flaw in his logic comes from the idea that errors must compound per-token. Somehow, even if the model makes a mistake, it is able to correct itself and decrease the sequence-level error rate. This is an incredible development, and was not the case with prior generations of LLMs.</p><p>And it turns out that the models&#8217; mechanisms for correcting themselves are interesting and interpretable:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EPOU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EPOU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png 424w, https://substackcdn.com/image/fetch/$s_!EPOU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png 848w, https://substackcdn.com/image/fetch/$s_!EPOU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png 1272w, https://substackcdn.com/image/fetch/$s_!EPOU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EPOU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png" width="598" height="329" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bd180725-048f-4a74-84d8-472e46d420f9_598x329.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:329,&quot;width&quot;:598,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!EPOU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png 424w, https://substackcdn.com/image/fetch/$s_!EPOU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png 848w, https://substackcdn.com/image/fetch/$s_!EPOU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png 1272w, https://substackcdn.com/image/fetch/$s_!EPOU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd180725-048f-4a74-84d8-472e46d420f9_598x329.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">An example from the<a href="http://arxiv.org/abs/2501.12948"> DeepSeek R1 report</a> of a language model increasing its probability of success mid-sequence, which Yann Lecun has argued for several years is impossible.</figcaption></figure></div><p></p><p>As shown in the image, models can in fact increase their likelihood of success mid-sequence by generating specific strings of tokens. A cottage industry of research is emerging trying to characterize and induce these behaviors, such as &#8220;backtracking&#8221; to a better solution. (It&#8217;s worth noting here that we still don&#8217;t know how generalizable these techniques are outside of the types of problems these models were trained on, like coding and math problems.)</p><h4><strong>Why Yann Lecun was right (kind of)</strong></h4><p>Naturally, people have been upset about all this. One of the founding members of the field has been giving bad advice to early-stage researchers based on a busted intuition. It&#8217;s infuriating, right?</p><p>Well, not exactly. I think that people are taking Yann&#8217;s argument a little too literally. Yes, we&#8217;ve figured out a way to build language models that don&#8217;t strictly get worse as we use them to generate longer outputs. But the limiting behavior remains the same: <em>eventually, if we continue generating from a language model, the probability that we get the answer we want still goes to zero</em>.</p><p>The practical takeaway from this is that AIs can&#8217;t work on their own forever. Lots of people are working on building Agents, systems that use language models to accomplish tasks over long time horizons. But the quest for a fully autonomous agent feels similar to the quest for fully self-driving cars: it might <em>never</em> be possible to build this, at least with the current stack.</p><p>There may be a kind of <a href="https://en.wikipedia.org/wiki/Data_processing_inequality">data processing inequality</a> going on behind the scenes. In some sense, the highest-quality information inputted to language models comes from the human-written prompts (and potentially inputs read in via tool use, like checking flight times or the weather). When the language models are left on their own to generate infinitely-long chains of thought, that input &#8220;signal&#8221; attenuates to nothing; eventually, without further input from a human, those chains of thought lose all meaningful value. Improving our technology can delay this, and improve the quality and amount of work we can do with a single input prompt. But it doesn&#8217;t seem likely that I&#8217;ll wake up one day next year and this figure (work / prompt) will have spiked to infinity.</p><p>This is why measuring language models&#8217; progress in terms of AGI timelines is misguided. We should be thinking about language models the same way we think about cars: How long can a language model operate without needing human intervention to correct errors? Framing our inquiries into language models like this allows us to reconcile Yann&#8217;s valid concerns about the models with new advances from OpenAI and DeepSeek&#8212;and will also lead to more productive research and conversation in the language model field, as it has with cars.</p><p>Instead of waiting for FAA (fully-autonomous agents) we should understand that this is a continuum, and we&#8217;re consistently increasing the amount of useful work AIs can do without human intervention. Even if we never push this number to infinity, each increase represents a meaningful improvement in the amount of economic value that language models provide. It might not be AGI, but I&#8217;m happy with that.</p>]]></content:encoded></item></channel></rss>