<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[ModelAnalysis.ai]]></title><description><![CDATA[An in-depth technical analysis of different AI models (including LLMs) and their applications in finance.]]></description><link>https://www.modelanalysis.ai</link><image><url>https://substackcdn.com/image/fetch/$s_!WXJI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe437b9f-353e-4fee-8821-958ab73f24de_1024x1024.png</url><title>ModelAnalysis.ai</title><link>https://www.modelanalysis.ai</link></image><generator>Substack</generator><lastBuildDate>Mon, 27 Apr 2026 01:02:13 GMT</lastBuildDate><atom:link href="https://www.modelanalysis.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Ram  Komarraju]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[ramkomarraju@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[ramkomarraju@substack.com]]></itunes:email><itunes:name><![CDATA[Ram  Komarraju]]></itunes:name></itunes:owner><itunes:author><![CDATA[Ram  Komarraju]]></itunes:author><googleplay:owner><![CDATA[ramkomarraju@substack.com]]></googleplay:owner><googleplay:email><![CDATA[ramkomarraju@substack.com]]></googleplay:email><googleplay:author><![CDATA[Ram  Komarraju]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Auto_Draw: Frontier Models Can See What's Wrong. Fixing it is a Different Story]]></title><description><![CDATA[What a Weekend Project Revealed About Self Improvement]]></description><link>https://www.modelanalysis.ai/p/auto_draw-frontier-models-can-see</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/auto_draw-frontier-models-can-see</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Sun, 19 Apr 2026 01:48:01 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3fde7540-703d-4c6c-87f0-0364f2fde5b6_3000x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This project was inspired by Andrej Karpathy&#8217;s <a href="https://github.com/jajko1/autoresearch">autoresearch</a> concept, but instead of an AI research task, I wanted something visual where the gap between what the model thinks it produced and what it actually produced would be immediately obvious to any human observer. </p><p>So I created a simple closed-loop experiment around programmatically generated images. A model had to write code that rendered an image, inspect the output, critique what was wrong, edit the code to fix those problems, and then repeat. I used coding agents including Codex, Claude Code, and Antigravity to run the loop, and the underlying models were GPT-5.4, Claude Opus 4.6, and Gemini 3.1 Pro. To keep the setup simple and measure each model&#8217;s end-to-end capabilities, I used self-evaluation throughout: the same model that generated the image also scored and critiqued it. The prompts ranged from easy to hard, starting with a wall clock and a world map, then moving to a bicycle and finally an elephant, all to be rendered as realistically as possible.</p><h2><strong>What the Models Did Well</strong></h2><p>The models followed the overall instructions quite well which is not surprising. They created the run folders, made per-iteration subfolders, wrote draw.py, rendered images, evaluated the outputs, edited the code, and stopped when they hit either the iteration limit or their quality threshold. They created some images quite realistically - maybe because they were trained on them specifically. More interestingly, especially in the early iterations of a run, they were often pretty good at seeing what was wrong with the image they had just produced. Even when their scores were a bit too optimistic, their written critiques were often specific, sensible, and directionally correct. They could tell when an elephant was missing a trunk, when a bicycle cockpit looked synthetic, when labels on a map overlapped, or when a clock face was missing an hour hand. The evaluation step, at least at first in each run, often looked stronger than the overall result.</p><h2><strong>Where the Self-Improvement Broke Down</strong></h2><p>The problem was that good critique did not reliably translate into good code edits. Again and again, the models identified real flaws, described them clearly, and then failed to make code changes that actually addressed those flaws. Sometimes the next iteration barely changed the image at all. Sometimes it made the result worse. A typical iteration might claim something like: &#8220;Added directional contour wrinkles to the ear and trunk, improved the shoulder-to-head transition with graduated shading, and introduced subtle skin texture variation across the flank.&#8221; But when I compared the before-and-after images, they often looked nearly identical. All three models often produced a plausible narrative of improvement while producing zero tangible change.</p><p>The elephant runs made this especially obvious. In one Opus run, iterations 12 through 20 contain detailed descriptions of anatomical corrections, texture refinements, and shading improvements. The self-scores keep inching upward, and the written evaluations remain articulate and specific. But when I compared iteration 12 and iteration 20 side by side, the visible improvement was minimal.</p><h2><strong>What This Suggests About Self-Correction</strong></h2><p>I do not think this experiment proves anything dramatic about self-improvement in the broadest sense. But I do think it says something useful about task-level self-correction. If self-improvement in practice means producing artefacts, evaluating the result, identifying what is wrong, and then making changes that lead to meaningful gains over repeated iterations, these experiments suggest that the loop is still much weaker than the hype implies, at least outside the domains where models are most heavily RL optimized. The agents did not fail completely. All three followed instructions well. All three identified real flaws. All three made some improvements on easier attempts (perhaps because they were trained explicitly on those images), and all three sometimes recovered from regressions.</p><h2><strong>My Current Take</strong></h2><p>What I saw most consistently was a disconnect between identifying what&#8217;s wrong, and then taking the necessary steps to fix it. Reinforcement learning is highly domain-specific, and coding gets a huge amount of frontier attention because it is measurable, economically valuable, and relatively easy to score, which makes it a natural target for RL training. But once you move into narrower or less directly rewarded domains like auto_draw, the loop seems much patchier. </p><h2>Now the Results</h2><p><strong>The wall clock</strong> was the easiest subject, and it shows. Opus moved from 5/10 to 8/10 in 7 iterations. Codex went from 3.8/10 to 8.1/10 in three iterations, and Gemini in just two! A wall clock is geometrically simple and the failures are concrete. So the models seem to be able to detect the problems but also make correlated code changes quite easily.  This is the auto_draw loop at its most effective and efficient.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LoL2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LoL2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png 424w, https://substackcdn.com/image/fetch/$s_!LoL2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png 848w, https://substackcdn.com/image/fetch/$s_!LoL2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png 1272w, https://substackcdn.com/image/fetch/$s_!LoL2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LoL2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png" width="1456" height="1057" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1057,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1203549,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LoL2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png 424w, https://substackcdn.com/image/fetch/$s_!LoL2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png 848w, https://substackcdn.com/image/fetch/$s_!LoL2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png 1272w, https://substackcdn.com/image/fetch/$s_!LoL2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc691e951-aafe-4c7b-ab28-8dd5c6266655_2400x1742.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NgIM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NgIM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png 424w, https://substackcdn.com/image/fetch/$s_!NgIM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png 848w, https://substackcdn.com/image/fetch/$s_!NgIM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png 1272w, https://substackcdn.com/image/fetch/$s_!NgIM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NgIM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png" width="1456" height="611" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:611,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:648725,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NgIM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png 424w, https://substackcdn.com/image/fetch/$s_!NgIM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png 848w, https://substackcdn.com/image/fetch/$s_!NgIM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png 1272w, https://substackcdn.com/image/fetch/$s_!NgIM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a22ad5b-f36e-4605-ac04-2d1b4dbf32a0_2400x1007.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tFFq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tFFq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png 424w, https://substackcdn.com/image/fetch/$s_!tFFq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png 848w, https://substackcdn.com/image/fetch/$s_!tFFq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png 1272w, https://substackcdn.com/image/fetch/$s_!tFFq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tFFq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png" width="1456" height="611" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:611,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:287111,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tFFq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png 424w, https://substackcdn.com/image/fetch/$s_!tFFq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png 848w, https://substackcdn.com/image/fetch/$s_!tFFq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png 1272w, https://substackcdn.com/image/fetch/$s_!tFFq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15f42a9c-84de-4cfe-b56f-1e82dc3d3d06_2400x1007.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>The world map</strong> drawings turned out to be better than expected -  I suspect it&#8217;s because the models must&#8217;ve been explicitly trained on getting the world map correct given some hoopla around this a couple of years ago.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!z8iN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!z8iN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png 424w, https://substackcdn.com/image/fetch/$s_!z8iN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png 848w, https://substackcdn.com/image/fetch/$s_!z8iN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png 1272w, https://substackcdn.com/image/fetch/$s_!z8iN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!z8iN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png" width="1456" height="1262" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1262,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1458311,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!z8iN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png 424w, https://substackcdn.com/image/fetch/$s_!z8iN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png 848w, https://substackcdn.com/image/fetch/$s_!z8iN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png 1272w, https://substackcdn.com/image/fetch/$s_!z8iN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e458bb0-5685-4cc3-ba79-3008804ecbfe_2400x2081.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wZPn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wZPn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png 424w, https://substackcdn.com/image/fetch/$s_!wZPn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png 848w, https://substackcdn.com/image/fetch/$s_!wZPn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png 1272w, https://substackcdn.com/image/fetch/$s_!wZPn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wZPn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png" width="1456" height="554" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:554,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:429600,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wZPn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png 424w, https://substackcdn.com/image/fetch/$s_!wZPn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png 848w, https://substackcdn.com/image/fetch/$s_!wZPn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png 1272w, https://substackcdn.com/image/fetch/$s_!wZPn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41fce3b0-a087-4564-a124-6d86d7d0a330_2400x913.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!O6Ly!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!O6Ly!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png 424w, https://substackcdn.com/image/fetch/$s_!O6Ly!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png 848w, https://substackcdn.com/image/fetch/$s_!O6Ly!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png 1272w, https://substackcdn.com/image/fetch/$s_!O6Ly!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!O6Ly!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png" width="1456" height="611" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:611,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:580115,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!O6Ly!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png 424w, https://substackcdn.com/image/fetch/$s_!O6Ly!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png 848w, https://substackcdn.com/image/fetch/$s_!O6Ly!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png 1272w, https://substackcdn.com/image/fetch/$s_!O6Ly!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21096faa-d267-4333-bd3f-1a8f1afb0e65_2400x1007.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Drawing the bicycle </strong>turned out to be more challenging for the models. There were multiple iterations where Opus and GPT-5.4 kept talking about making massive changes without any discernible change in the images generated. Gemini was just super positive in its own self assessment though :)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zjSV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zjSV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png 424w, https://substackcdn.com/image/fetch/$s_!zjSV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png 848w, https://substackcdn.com/image/fetch/$s_!zjSV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png 1272w, https://substackcdn.com/image/fetch/$s_!zjSV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zjSV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png" width="1456" height="1312" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1312,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:801070,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zjSV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png 424w, https://substackcdn.com/image/fetch/$s_!zjSV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png 848w, https://substackcdn.com/image/fetch/$s_!zjSV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png 1272w, https://substackcdn.com/image/fetch/$s_!zjSV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a8f2828-59b6-4f0e-b41b-1eec69d01b70_2400x2162.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wTTo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wTTo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png 424w, https://substackcdn.com/image/fetch/$s_!wTTo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png 848w, https://substackcdn.com/image/fetch/$s_!wTTo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png 1272w, https://substackcdn.com/image/fetch/$s_!wTTo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wTTo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png" width="1456" height="3125" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:3125,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4681355,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wTTo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png 424w, https://substackcdn.com/image/fetch/$s_!wTTo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png 848w, https://substackcdn.com/image/fetch/$s_!wTTo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png 1272w, https://substackcdn.com/image/fetch/$s_!wTTo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd74d9ad7-91ce-47e2-b2d1-bac6e053521a_2400x5151.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!r0nD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!r0nD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png 424w, https://substackcdn.com/image/fetch/$s_!r0nD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png 848w, https://substackcdn.com/image/fetch/$s_!r0nD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png 1272w, https://substackcdn.com/image/fetch/$s_!r0nD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!r0nD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png" width="1456" height="572" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:572,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:728235,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!r0nD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png 424w, https://substackcdn.com/image/fetch/$s_!r0nD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png 848w, https://substackcdn.com/image/fetch/$s_!r0nD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png 1272w, https://substackcdn.com/image/fetch/$s_!r0nD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c62e0aa-9ffc-46a3-9af0-8f856bc52009_2400x943.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>The elephant</strong> was where the models struggled the most. All three models could often accurately and specifically describe what was wrong. But they couldn&#8217;t convert those diagnoses into code changes that moved the image closer to photorealism. Opus was probably the worst performing amongst the three models, and Gemini&#8217;s iterations kept on gradually getting better even if its own self assessment was way too optimistic.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3reO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3reO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png 424w, https://substackcdn.com/image/fetch/$s_!3reO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png 848w, https://substackcdn.com/image/fetch/$s_!3reO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png 1272w, https://substackcdn.com/image/fetch/$s_!3reO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3reO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png" width="1456" height="2076" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2076,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4123785,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3reO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png 424w, https://substackcdn.com/image/fetch/$s_!3reO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png 848w, https://substackcdn.com/image/fetch/$s_!3reO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png 1272w, https://substackcdn.com/image/fetch/$s_!3reO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ef7c19e-daee-41e8-ba04-e055b66cfc69_2400x3422.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Bj32!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Bj32!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png 424w, https://substackcdn.com/image/fetch/$s_!Bj32!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png 848w, https://substackcdn.com/image/fetch/$s_!Bj32!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png 1272w, https://substackcdn.com/image/fetch/$s_!Bj32!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Bj32!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png" width="1456" height="1943" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1943,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3187477,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Bj32!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png 424w, https://substackcdn.com/image/fetch/$s_!Bj32!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png 848w, https://substackcdn.com/image/fetch/$s_!Bj32!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png 1272w, https://substackcdn.com/image/fetch/$s_!Bj32!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456d394-bd4e-4a10-8b56-f4117d39dd16_2400x3202.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NaO_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NaO_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png 424w, https://substackcdn.com/image/fetch/$s_!NaO_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png 848w, https://substackcdn.com/image/fetch/$s_!NaO_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png 1272w, https://substackcdn.com/image/fetch/$s_!NaO_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NaO_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png" width="1456" height="2076" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2076,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1675993,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/194008046?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NaO_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png 424w, https://substackcdn.com/image/fetch/$s_!NaO_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png 848w, https://substackcdn.com/image/fetch/$s_!NaO_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png 1272w, https://substackcdn.com/image/fetch/$s_!NaO_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d98b75-bdc8-4e92-b81f-9fd6c5ce5a7a_2400x3422.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p>]]></content:encoded></item><item><title><![CDATA[The Case for AI-Powered User Interfaces]]></title><description><![CDATA[How LLMs turn UI from a build-time artifact into a runtime capability.]]></description><link>https://www.modelanalysis.ai/p/ai-powered-user-interfaces-are-the</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/ai-powered-user-interfaces-are-the</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Sat, 14 Feb 2026 13:56:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!OjQH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#171;Back in April 2025, in my <a href="https://www.modelanalysis.ai/p/the-end-of-the-app-store">End of the App Store?</a> essay, I speculated about a future where a single AI-powered &#8220;super app&#8221; would replace all the apps. Models and agents have improved a lot since then at tool use and code generation, turning that &#8220;super app&#8221; concept into a clear and present capability whereby user interfaces can be generated on the fly to address a user&#8217;s need. This article is my February 2026 update, backed by a demo from a weekend project.&#187;</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OjQH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OjQH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!OjQH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!OjQH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!OjQH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OjQH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png" width="1024" height="559" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:559,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:754662,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/187949816?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OjQH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!OjQH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!OjQH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!OjQH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5856128a-8f1a-4663-b47f-1c4099d6267b_1024x559.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>The key takeaways up front (sort of like a very long TL;DR)</h3><p>The static app model has worked for decades, but it forces every user to learn how to navigate many apps, each with its own fixed set of predetermined screens. The apps in turn need armies of product managers, business analysts, architects, developers, testers, and SREs to build and maintain them.</p><p>In this article, I propose that AI-powered Dynamic Interface Generation (DIG) exchanges rigid consistency, long build times, and high build costs for high flexibility, lower predictability, just-in-time builds, and higher runtime costs. The LLM generates exactly the interface the moment requires. No more, no less.</p><p>The range of tasks where a generated interface is good enough is growing as models improve. Functions like task tracking, workflows, email triage, calendar management, and integration of different UI components such as maps, tables, charts, and data dashboards are all within reach today. Stateful, high-frequency workflows and complex functionality will likely need a purpose-built UI for a while. And the backend services that implement core business logic will remain (of course, AI coding agents will be used to generate them as well!).</p><p>Given these rapidly increasing capabilities, this is the future I envision for phones, laptops/desktops, and enterprises:</p><ul><li><p>On phones, you stop hopping between apps and instead ask a smart assistant for outcomes, and it dynamically assembles UI and flow to achieve that outcome for you.</p></li><li><p>On laptops/desktops, there&#8217;s more room for specialized tools but the main interface for many users will likely be a single adaptive workspace that pulls the right panes and actions across email, docs, tickets, dashboards, and more.</p></li><li><p>In enterprises, multiple front-end apps converge into one governed interface sitting above many systems and services, making most internal CRUD and workflows generated, and pushing SaaS value downward from vendor screens toward APIs offering high-value capabilities.</p></li></ul><p>In the short to medium term, browsers are a natural choice for implementing DIG for enterprises because of their secure-by-default runtimes, the JS/HTML/CSS combination that lends itself to dynamic interface generation, and centralized deployment capability. On the other hand, mobile operating systems will likely develop OS-native DIG capabilities for tighter device integration and lower latency.</p><p>All leading LLMs already exhibit some dynamic interface generation capabilities (when they generate deep research artefacts, for example). Google&#8217;s experimental browser, Disco, also takes a stab at this with its gentabs functionality. But it&#8217;s not quite what I thought would illustrate the vision I had in mind. So, I built a prototype to show how the dynamic interface generation loop might work. You can watch the video here: </p><div id="youtube2-qng2o79W5ck" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;qng2o79W5ck&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/qng2o79W5ck?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>For those of you who are interested in reading through, here we go&#8230;</p><h3>1. Introduction</h3><p>We are so used to doing things a certain way that we assume it is the only sensible way to get things done. But every so often, a real shift in technology lets us question our most basic assumptions. The idea that we need a separate app for each kind of use, with a bunch of prebuilt screens and rigid flows, is something we&#8217;ve all come to expect. But it doesn&#8217;t have to be that way!</p><p>Many of us literally have hundreds of apps, each created for its own special use. Wouldn&#8217;t it be nice if your &#8220;smart&#8221; phone was actually smart, and gave you a single place to start, whether that is a screen or voice, and then quietly invoked whatever functions and services were needed to get stuff done?</p><p>The same problem shows up even more clearly in enterprises. Every corporate worker needs to learn dozens of applications to do their job. Sure, there are massive systems like ERP software that try to pull it all together. But customizing these systems is a gargantuan task, and the result is often a complex jumble of screens that reflects what was feasible to build, and users still have to learn how to go through them to get their job done. Wouldn&#8217;t it be better to have a single screen that adapts itself to the job you are trying to do? It might still interact with dozens of systems using all kinds of APIs behind the scenes, but you would not have to care.</p><p>That&#8217;s what I think is possible with LLM powered Dynamic Interface Generation (DIG). In the rest of this essay, I&#8217;ll talk about the pain points faced by today&#8217;s app-centric model and static user interfaces, make the case for DIG, share a video of an app that I created to demonstrate the art of the possible, and conclude with current limitations and why I believe many of them are likely to be overcome within the next few years.</p><h3>2. The Problem with Thousands of Apps and Static User Interfaces</h3><p>Almost every UI we use today was designed by a team that guessed what screens we&#8217;d need. They built navigation hierarchies, screen layouts, and interaction patterns long before we ever opened the app. When our needs don&#8217;t fit their assumptions, we submit a feature request (if we can) and hope it gets fixed in the next version.</p><p>These feature requests go into backlogs. Product teams prioritize based on aggregate demand. The interface we actually want, a specific combination of data, layout, and actions, may never ship because it&#8217;s too niche to justify engineering time.</p><p>The result is a world of rigid, prebuilt interfaces that users learn to navigate. Instead of software adapting to them, they adapt to software. </p><p>Thousands of special purpose apps also create very concrete operational drag. Users lose time context switching as they move between tools. Work gets stuck in the seams between apps (copy-paste!) and brittle handoffs that rely on people remembering &#8220;the next step.&#8221;</p><h3>3. What Dynamic Interface Generation Changes</h3><p>DIG replaces the static design-build-deploy model with a fundamentally different approach. The user describes what they want to accomplish, and an LLM generates the interface on demand, connected to real data, capable of real actions to support the user&#8217;s intent.</p><p>It also changes how &#8220;UI requirements&#8221; work, pushing them into runtime via the intent expressed in the user&#8217;s prompt. The &#8220;requirement&#8221; becomes &#8220;help me accomplish X,&#8221; and the system is free to render the best UI it can for the moment, given the tools and the data available. That is a different way of building software, and it shifts effort away from endless UI iterations.</p><p>There will be fewer and fewer apps with prebuilt screens. No navigation menus and screens designed by committee. The LLM is provided with a list of tools and services, and it decides which tools to invoke, what data to fetch, what interactions to offer, and how to render the user interface. The things steering it are the user&#8217;s intent, the state provided in context, and what it knows of the user&#8217;s preferences and history.</p><h3>4. The Opportunities</h3><h4>Smartphones</h4><p>We will see a gradual transition toward phones where the primary interface is no longer a grid of apps. The OS comes with one &#8220;smart&#8221; app, and &#8220;apps&#8221; become API based service providers. For example, in the case of the iPhone, Siri will not just be a voice assistant you occasionally invoke. We will move toward a future where Siri is the main interface you use to get things done. I know, I know. Siri gets a lot of flak, and deservedly so. But I&#8217;m talking about a future where Siri is powered by an AI that actually works. </p><p>There might remain some special cases where precise control and efficiency of a static UI would be needed, but for the vast majority of cases an LLM generated dynamic UI is likely to be the better choice (especially with LLMs specially trained with UI generation in mind). Consumer devices will likely push more of the DIG experience into OS-native implementations for tighter integration with device capabilities and better latency.</p><h4>Laptops and Desktops</h4><p>Laptops and desktops are where DIG can get interesting because the surface area for work is larger, and some specialist interfaces like the unix terminal are likely to remain in use for a while. But most casual users will likely work inside a single adaptive workspace that generates the right mix of panes: email, calendar, documents, spreadsheets, and chat.</p><h4>The Enterprise</h4><p>Similar to the mobile and desktop experience, there will be a convergence toward a single interface that sits on top of a complex web of AI agents and traditional deterministic services.  And the UI layer orchestrates them based on the user&#8217;s intent.</p><p>A huge amount of enterprise software consists of CRUD screens, workflow enabling screens (approvals, routing, etc.), reconciliation views, status dashboards etc. And these are ripe for replacement by a single dynamically generated interface.</p><p>Modern web browsers have intrinsic advantages that make them a strong starting point for DIG as a universal integration shell in enterprises. They provide mature isolation and security primitives, and the HTML/CSS/JS stack is a powerful, widely supported substrate for dynamic rendering.</p><p>SaaS might still matter because the cost of building and operating a complex business domain can be shared among many customers, but the vendors&#8217; primary interface becomes less defensible. It&#8217;s likely to become an API and capability layer, while the UI becomes generated and owned by the enterprise rather than by each vendor. </p><p>This also reframes build-vs-buy more broadly. Enterprises will still buy capabilities, but end up with one adaptive interface that can unify workflows across many systems, while integrating with corporate data, and enforcing consistent governance and permissioning.</p><h3>5. So, what changes?</h3><h4>Every user gets a tailored experience</h4><p>Two users asking &#8220;show me my schedule&#8221; might get completely different layouts depending on how many events they have, what data is available, and how they phrased the request. There is no single &#8220;calendar view&#8221; that has to work for everyone. The LLM tailors every render to the specific context.</p><h4>The long tail of internal tools disappears</h4><p>Enterprises maintain hundreds or thousands of internal dashboards, admin panels, and CRUD interfaces, each with its own deployment pipeline, bug backlog, and maintenance burden. Most of these are simple combinations of data fetching and form submission. A DIG system replaces them with a single deployment that generates each interface on demand. The integration surface shrinks to a set of tool definitions and API credentials.</p><h4>Product iteration method changes substantially</h4><p>The traditional cycle (design a screen, ship it, measure engagement, redesign) takes weeks or months. With DIG, this cycle is broken, and traditional methods like A/B testing become a lot less relevant as every interaction is a unique rendering. We&#8217;ll need to find new ways to evaluate the effectiveness of the screens generated, and how to feed what we&#8217;ve learned back into the next version of the agent or the LLM.</p><h3>6. What Doesn&#8217;t Work Yet</h3><p>Many of the limitations described below can (and will) be addressed as models are trained with more examples of DIG (how to build better interfaces dynamically to meet user intent while using context, data, and available tools and services effectively).</p><p><strong>Consistency.</strong> The same prompt produces different layouts on different runs. For workflows where muscle memory matters, this could be a real problem. The system needs layout pinning, preference learning, and perhaps &#8220;UI contracts&#8221; for certain workflows.</p><p><strong>Latency.</strong> Generating a UI render takes many seconds compared to the milliseconds taken by traditional prebuilt screens. Caching common patterns and using partial renders would help, but we will likely rely on special purpose SLMs or on LLM performance improved further as it&#8217;s been over the past few years.</p><p><strong>Complexity ceiling.</strong> Leading LLMs can generate most of the UI functionality dynamically. But, there is a threshold beyond which it needs to rely on prebuilt software (e.g., a spreadsheet component). But, LLMs can be trained to embed and use prebuilt components where needed. </p><p><strong>Cost.</strong> Every interaction is an LLM call. At current token pricing, a power user using a DIG based app all day will accumulate significant API costs. This will improve as inference gets cheaper, but it&#8217;s a real constraint today.</p><p><strong>Security.</strong> This isn&#8217;t a new issue with DIG per se, but I wanted to mention it here before someone calls me out on this :) The traditional security best practices will still matter. Backend services will still need to authenticate requestors, check if they have the right permissions, validate all input coming from the UI, and enforce all business rules, etc.</p><h3>7. Conclusion</h3><p>The static app model is optimized for a world where UI is expensive to build and where the safest way to ship software is to freeze screens into predictable, testable flows. But this world is changing.</p><p>With DIG, we give up some predictability in exchange for a UI that can match intent, context, and workflow in real time. But the &#8220;good enough&#8221; frontier is expanding fast. As models and the agentic solutions around them improve, the number of tasks that can be handled by generated interfaces will keep expanding. If that happens, single-purpose apps will reduce in number, and more of what we think of as &#8220;apps&#8221; will collapse into services and tools behind generated interfaces.</p>]]></content:encoded></item><item><title><![CDATA[Model Anatomy]]></title><description><![CDATA[A visual way to understand and compare LLM architectures]]></description><link>https://www.modelanalysis.ai/p/model-anatomy</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/model-anatomy</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Sat, 17 Jan 2026 18:51:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!0phx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I built <a href="https://model-anatomy-ram-komarraju.netlify.app/">Model Anatomy</a> to help visualize and compare the internal structures of large language models. Here&#8217;s what it does.</p><h2><strong>Inspect a Single Model or Compare Models Side-by-Side</strong></h2><p>Select any two models from the dropdown menus and their architectures appear side by side. You can filter by model family (Llama, Gemma, Mistral, etc.) to narrow down your selection. You can see in the diagram below that Gemma 3 uses a hybrid normalization (combination of pre and post-norms) and OLMo 3 uses post-norm, and both of these are different from the more common pre-norm approach. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0phx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0phx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png 424w, https://substackcdn.com/image/fetch/$s_!0phx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png 848w, https://substackcdn.com/image/fetch/$s_!0phx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png 1272w, https://substackcdn.com/image/fetch/$s_!0phx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0phx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png" width="1456" height="1123" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1123,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:276197,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/184874241?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0phx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png 424w, https://substackcdn.com/image/fetch/$s_!0phx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png 848w, https://substackcdn.com/image/fetch/$s_!0phx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png 1272w, https://substackcdn.com/image/fetch/$s_!0phx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a3d8435-93f4-426c-b837-d26ef0091ee5_1470x1134.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The LLM architecture doesn&#8217;t seem to have changed that much since GPT-2. But a careful comparison reveals interesting architectural differences. Most architectures use post-norm instead of pre-norm. Position embeddings gave way to RoPE. The original multi-head attention mechanism saw a few innovations like GQA. Mixtral stands out with its Mixture-of-Experts (MoE) layers visible in the diagram.</p><h2><strong>Understand Parameter Counts</strong></h2><p>Toggle &#8220;Show parameter counts&#8221; to see where parameters actually live in each model. Every number is interactive. Hover over any parameter count to see exactly how it&#8217;s calculated.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HDp6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HDp6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png 424w, https://substackcdn.com/image/fetch/$s_!HDp6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png 848w, https://substackcdn.com/image/fetch/$s_!HDp6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png 1272w, https://substackcdn.com/image/fetch/$s_!HDp6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HDp6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png" width="1456" height="1196" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1196,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:383915,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/184874241?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HDp6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png 424w, https://substackcdn.com/image/fetch/$s_!HDp6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png 848w, https://substackcdn.com/image/fetch/$s_!HDp6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png 1272w, https://substackcdn.com/image/fetch/$s_!HDp6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e35dab-1db7-4baa-93c2-cdf2a76b64e2_1498x1230.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>For example, hover over the attention layer&#8217;s parameter count and you&#8217;ll see the formula break down: Q/K/V projection weights, output projection, the dimensions multiplied out. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ilSm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ilSm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png 424w, https://substackcdn.com/image/fetch/$s_!ilSm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png 848w, https://substackcdn.com/image/fetch/$s_!ilSm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png 1272w, https://substackcdn.com/image/fetch/$s_!ilSm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ilSm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png" width="1042" height="700" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:700,&quot;width&quot;:1042,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172168,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/184874241?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ilSm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png 424w, https://substackcdn.com/image/fetch/$s_!ilSm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png 848w, https://substackcdn.com/image/fetch/$s_!ilSm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png 1272w, https://substackcdn.com/image/fetch/$s_!ilSm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec2342f-dd99-4333-b70f-d0393d567f59_1042x700.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This reveals some interesting patterns:</p><ul><li><p><strong>Embedding layers are expensive.</strong> In smaller models, the embedding table can account for a significant fraction of total parameters. A 50k vocabulary &#215; 4096 dimensions = 200M parameters before you even get to the transformer blocks.</p></li><li><p><strong>FFN layers dominate.</strong> In most architectures, the feed-forward network within each transformer block contains roughly 2/3 of that block&#8217;s parameters. The attention mechanism everyone talks about is actually the minority.</p></li><li><p><strong>Scaling isn&#8217;t uniform.</strong> Models don&#8217;t just multiply everything by the same factor when going from 8B to 70B. Hidden dimensions, number of layers, and attention heads all scale at different rates.</p></li></ul><h2><strong>Inspect Individual Modules</strong></h2><p>Click any module in the diagram to see its details: the mathematical formula, the tensor shapes flowing through it, and what each component does. The side panel shows a graphical breakdown, and if you enable &#8220;Show implementation code,&#8221; the bottom panel shows PyTorch-style pseudocode.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ocyv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ocyv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png 424w, https://substackcdn.com/image/fetch/$s_!ocyv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png 848w, https://substackcdn.com/image/fetch/$s_!ocyv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png 1272w, https://substackcdn.com/image/fetch/$s_!ocyv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ocyv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png" width="778" height="1352" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1352,&quot;width&quot;:778,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:113688,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/184874241?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ocyv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png 424w, https://substackcdn.com/image/fetch/$s_!ocyv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png 848w, https://substackcdn.com/image/fetch/$s_!ocyv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png 1272w, https://substackcdn.com/image/fetch/$s_!ocyv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d26fd58-3a58-4f22-b86d-b90ac362d6eb_778x1352.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Implementation details with pseudocode:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ky9q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ky9q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png 424w, https://substackcdn.com/image/fetch/$s_!Ky9q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png 848w, https://substackcdn.com/image/fetch/$s_!Ky9q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png 1272w, https://substackcdn.com/image/fetch/$s_!Ky9q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ky9q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png" width="1456" height="410" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:410,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:195502,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/184874241?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ky9q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png 424w, https://substackcdn.com/image/fetch/$s_!Ky9q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png 848w, https://substackcdn.com/image/fetch/$s_!Ky9q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png 1272w, https://substackcdn.com/image/fetch/$s_!Ky9q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb160339e-7c82-42cc-8f7d-db54ad8b1190_2472x696.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>What I Learned Building This</strong></h2><p>I first sketched this idea over a year ago but never built it. The implementation seemed tedious, involving parsing model configs, building SVG diagrams, handling all the edge cases between different architectures.</p><p>Last week I sat down with a coding agent and had a working version in an evening. The agent handled the boilerplate, I directed the design decisions. Things that would have taken me hours of documentation-reading happened in seconds. The productivity multiplier was real.</p><p>This tool is far from perfect, and I selected only some model variants, missing edge cases in the visualization, or features that would be useful. But, hopefully you&#8217;ll still find it useful whether you&#8217;re new to the field or just interested in comparing two models.</p><h2><strong>Learn More</strong></h2><p>If you want to go deeper into model architectures, Sebastian Raschka has an excellent breakdown of modern LLM designs at <em><a href="https://substack.com/home/post/p-168650848">Understanding Large Language Models</a></em>. His blog covers some interesting insights and implementation details that this tool visualizes. Shout out to Jay Alammar whose <em><a href="https://jalammar.github.io/illustrated-gpt2/">Illustrated GPT-2</a></em> explained GPT-2 pretty comprehensively and intuitively when I started out years ago.  </p><h2><strong>Try It / Contribute</strong></h2><p>The tool works best on desktop but should be usable on mobile browsers (the layout stacks vertically on phones).</p><p>What models are missing? What features would be useful? What&#8217;s confusing, incorrect, or broken? Leave a comment or reach out. I&#8217;d like to make this more useful for anyone trying to understand what&#8217;s actually inside these models. I will also give you credit for all your findings and contributions :)</p>]]></content:encoded></item><item><title><![CDATA[The Great AI Code Transition]]></title><description><![CDATA[A Vision of What the Future of AI-Written Software Could Look Like]]></description><link>https://www.modelanalysis.ai/p/the-great-code-transition</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/the-great-code-transition</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Thu, 08 Jan 2026 07:38:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Io51!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#171;Disclaimer and Notes: as is usually the case, there are many ways to present a narrative like this. I could&#8217;ve sliced the software into a completely different set of categories but I chose the set of categories that I thought would effectively illustrate the points I wanted to make. It should then be obvious that the percentages presented in the chart for various categories of code are meant to indicate a rough guess of what their relative sizes might be like. A more rigorous, data-based approach is needed to better ground this chart. It is often said that prediction, especially about the future, is hard. The biggest of my speculations is the extent to which AI generated disposable code would take over our interactions on edge devices. Finally,</em> <em>each of the points I wrote below can be expanded into an essay, and I might do that if there&#8217;s enough interest and response to this article.&#187;</em></p><p>It is very clear as we enter 2026 that LLM-based AI coding agents have become powerful enough that more and more of code will be written using them. In fact, it is inevitable that in the not-so-distant future, almost all of the code will transition from being hand-crafted by humans to being written by these AI coding agents. This article looks at what a transition to such a future could look like. </p><p>Let&#8217;s start off with a chart that compares at a high-level the present state of affairs with this future. The left-hand side of the chart presents the current state where almost all the code is written by humans (I know, I know, some of the leading tech companies and start ups might have been aggressively switching to coding agents already, but by and large the vast majority of places are still using humans to write the code). As we shift right along the X-axis, you can see more and more of the code being written by AI. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Io51!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Io51!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!Io51!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!Io51!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!Io51!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Io51!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png" width="728" height="397.4140625" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:559,&quot;width&quot;:1024,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:719844,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/183879093?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Io51!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!Io51!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!Io51!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!Io51!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F997fb7d7-fe6b-4d14-a9f8-00f5cb56670f_1024x559.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Here are some key observations about each of the four layers that will comprise the AI-written software:</p><ol><li><p>How fast <strong>disposable software</strong> emerges on edge devices depends not only on model capability but also on privacy, latency, cost, power constraints, and offline reliability. Many consumer experiences may route through cloud inference to access stronger models, while privacy needs and regulatory requirements will push more on-device inference even if those small models are weaker. This will likely create a practical split in user experience, where the best functionality would be available to users that can afford to pay for better connectivity, newer hardware, and premium inference.</p></li><li><p>Platforms that enable sophisticated <strong>&#8220;vibe-coded&#8221;</strong> small business apps and sites will expand quickly. Users will focus on the business value proposition and user experience, while relying on the platform for scaffolding that provides security, payments, observability, data storage and compliance. This reduces the need for deep technical knowledge and extensive testing at small scale while increasing dependence on platform guardrails and concentrates risk in platform providers.</p></li><li><p><strong>Large enterprises</strong> outside big tech will take longer to reach a fully AI-written software stack, maybe a decade or more, because the bottleneck is not capability but liability, controls, and regulations. They will need new SDLC and operating models that make AI-assisted changes auditable, reproducible, testable, and attributable, with clear responsibility when something breaks. The transition will be uneven: experiments and greenfield products will move quickly, while core systems of record and regulated workflows will lag. Organizationally, some middle layers of coordination shrink or change shape, and effort shifts toward governance, verification, security, reliability engineering, and incident response.</p></li><li><p>For <strong>systems and platforms</strong>, AI will write much of the code churn, but expert steering remains essential because correctness, performance, and security constraints are unforgiving. Specialists will focus less on manual implementation and more on precise specifications, choosing/designing the correct algorithms, and rigorous verification.</p></li></ol><p>The next four points will focus on what this means from <strong>socio-economic and legacy system impact</strong> perspectives:</p><ol><li><p>Many software builders who take pride and identity from craftsmanship will find the transition emotionally and professionally difficult as the work shifts from writing code to defining intent, setting constraints, verifying behavior, and owning outcomes. The people most likely to prosper are those with strong product judgment, clear vision, enough technical literacy to reason about tradeoffs, and the ability to communicate precise techno-functional requirements to coding agents. Deep specialists will still matter, especially in systems with high-performance, industrial-grade security, and safety-critical needs, either hand-crafting core components or steering agents with high precision. Status and fulfillment will increasingly come from accountability, reliability, and impact rather than elegance of implementation.</p></li><li><p>Entry-level coding jobs and many mid-skill implementation roles may shrink as AI can easily write simple code that would historically have been written by junior coders. It&#8217;d be interesting to see how the junior coders will be able to learn to become good coders in the age of AI coding assistants. At the same time, more people will be able to build sophisticated functionality without formal training broadening participation.</p></li><li><p>As to how AI functionality itself will be integrated into software, I&#8217;m expecting that two competing product visions will coexist for a while. One is special-purpose applications with AI embedded, optimized for functionality, precision, and performance. The other is a general-purpose AI layer that generates tailored interfaces and orchestrates specialized modules and systems on demand. If the general-purpose layer becomes the dominant interface, it challenges traditional suites like Microsoft Office by demoting Word and Excel to the back-end while elevating agent-driven interfaces to the top.</p></li><li><p>Finally, if something like AGI emerges, the technical need for human steering could drop further, but human oversight and accountability may still be required for legitimacy. I&#8217;m assuming (for humanity&#8217;s sake!) that ownership and liability will ultimately remain in human hands.</p></li></ol><p>That concludes this short essay. There&#8217;s more that I can write on this topic, but this is all I could write in the time I&#8217;d allotted to myself. Let me know your thoughts, and thank you for reading all the way to the end :)</p>]]></content:encoded></item><item><title><![CDATA[Software Development in the Age of AI Coding Agents]]></title><description><![CDATA[When cost of code approaches zero, implementing intent and managing complexity become the job]]></description><link>https://www.modelanalysis.ai/p/software-development-in-the-age-of</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/software-development-in-the-age-of</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Fri, 26 Dec 2025 21:35:57 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!nwOj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#171;No AI was used in writing this article. Other than the obvious use of Gemini Nano Banana Pro for the images.&#187;</em></p><p>It&#8217;s hard to make predictions, especially about a future that&#8217;s rapidly unfolding like science fiction in real life and feels surreal and palpable at the same time. But, that&#8217;s exactly what I&#8217;ll try to do in this article: predict the future of software development across multiple dimensions.</p><div><hr></div><h2>AI Coding Agents Will Continue to Get Better</h2><p>As we stand at the cusp of 2026, I think we can confidently envision a future where almost all of the code will be written by AI coding agents. Here are a few supporting factors that will turn the vision into reality:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nwOj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nwOj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png 424w, https://substackcdn.com/image/fetch/$s_!nwOj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png 848w, https://substackcdn.com/image/fetch/$s_!nwOj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png 1272w, https://substackcdn.com/image/fetch/$s_!nwOj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nwOj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png" width="1024" height="565" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:565,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:823204,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/182237010?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nwOj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png 424w, https://substackcdn.com/image/fetch/$s_!nwOj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png 848w, https://substackcdn.com/image/fetch/$s_!nwOj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png 1272w, https://substackcdn.com/image/fetch/$s_!nwOj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8668c904-f4dd-45b2-8f7f-015ee8228ca4_1024x565.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ul><li><p>Cost of producing code will approach zero for all practical purposes. This is pure tokenomics at play. There&#8217;s little to distinguish between different leading models, and when coupled with the inevitable algorithmic and infra efficiencies, this can only lead to the eventual commoditization of the coding LLMs (and the agents on top of that).</p></li><li><p>The three major limitations that currently plague LLM-based coding agents will become only minor limitations over time:</p><ul><li><p>The amount of memory (context) that LLMs can hold at a given time will increase over the next couple of years. So much so that it won&#8217;t be the major limitation that it is today. So, context engineering, much like prompt engineering from &#8216;23-24, will become a sidebar conversation within the next couple of years.</p></li><li><p>Context rot and fragility to prompting will be addressed at the underlying technology and surrounding tooling levels, as well as by users learning to use the coding agents more effectively. For example, we will learn to discard a poisoned context and move to a new context quickly, and configure system and tool prompts to extract the best performance.</p></li><li><p>LLMs will get better at following best practices and architectural principles. In my opinion, this is a matter of RL training on more and more examples of high quality code annotated with comments highlighting best practices. In other words, creating high quality code will get better assuming that correct context and prompts with architectural guidance are used. </p></li></ul></li><li><p>We will develop new techniques and methodologies to use the AI coding agents effectively (e.g., spec driven implementation and plan driven programming). These practices will continue to evolve at a rapid pace as long as the coding agents continue to evolve in their capabilities, and will reach their maturity as coding agents will reach a mature state. For example, the importance of context engineering will diminish as context length increases, and coding agent developers would have learned how and what to include in the context efficiently. Similarly existing tooling for coding (e.g., IDEs or CLIs) will continue morph and adapt to this new paradigm. We see this transition already whereby IDEs and CLIs adapting their interfaces and tooling around agent-centric development. I predict that we will see integrated tooling that will enable users to manage the entire lifecycle of the application (from testing to deployment) from within a single application like Cursor.</p></li><li><p>Note that we don&#8217;t need coding agents to work completely independently for hours and days in order for these predictions to come true. </p></li></ul><div><hr></div><h2><strong>What are the likely broad implications for software development?</strong> </h2><p><br>We&#8217;ll see Jevon&#8217;s Paradox in action. When cost of writing code approaches zero, there will be more demand than ever to write new software, and the bottlenecks move elsewhere:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IKgb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IKgb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png 424w, https://substackcdn.com/image/fetch/$s_!IKgb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png 848w, https://substackcdn.com/image/fetch/$s_!IKgb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png 1272w, https://substackcdn.com/image/fetch/$s_!IKgb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IKgb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png" width="1024" height="565" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:565,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:827931,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/182237010?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IKgb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png 424w, https://substackcdn.com/image/fetch/$s_!IKgb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png 848w, https://substackcdn.com/image/fetch/$s_!IKgb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png 1272w, https://substackcdn.com/image/fetch/$s_!IKgb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cb2b6ba-5ad9-4c8f-8acf-0e85b6bd3c3e_1024x565.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ol><li><p>Deciding what software to build or what changes to make to the existing software. There&#8217;s a tremendous amount of work done outside of coding. You need to understand the business problem. You need to understand the trade offs between scalability, maintainability, and latency vs throughput, and choose the options suitable for the context at hand. You need to agree on the interfaces with other teams. And then you find the way to implement it in an existing system in as elegant a way possible while keeping in mind the impact it might have on end users.</p></li><li><p>Reviewing and testing what&#8217;s been written. In &#8216;26 it might still be about ensuring that the correct architectural choices are made, but increasingly it would be about ensuring the intent is captured appropriately. Because language is an imperfect communication mechanism and even the best coding agent cannot divine the user&#8217;s intentions correctly all the time.</p></li><li><p>Iterate through what&#8217;s been written. Just because software can be produced easily doesn&#8217;t mean that can we can get to what we want very quickly. For one thing, it&#8217;s often very difficult to know what one wants, and getting things right requires iterations. So, we&#8217;ll likely try options A, B and C because of the cost of trying things is trivial, while the cost of getting it wrong remains high.</p></li></ol><div><hr></div><h2><strong>What does this mean for people/jobs?</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pkxY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pkxY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png 424w, https://substackcdn.com/image/fetch/$s_!pkxY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png 848w, https://substackcdn.com/image/fetch/$s_!pkxY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png 1272w, https://substackcdn.com/image/fetch/$s_!pkxY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pkxY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png" width="1024" height="565" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:565,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:866677,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/182237010?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pkxY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png 424w, https://substackcdn.com/image/fetch/$s_!pkxY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png 848w, https://substackcdn.com/image/fetch/$s_!pkxY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png 1272w, https://substackcdn.com/image/fetch/$s_!pkxY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11f2dbe8-e71c-4d94-ab4a-23ef76d9b6d9_1024x565.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ul><li><p>There would be many hardcore vibe coders who just want to develop applications to solve the business problem at hand. These individuals will gravitate towards platforms like Lovable, Replit (and other offerings that hyperscalers like AWS are likely to provide). These are the Squarespace/Shopify-like platforms of the future.</p></li><li><p>Software engineers that understand code better, know architectural best practices and can wield the agents/tools to operate reliable, scalable, and maintainable enterprise software of the future. This forms the bulk of the current SWE space, and I predict that Jevon's paradox will hold here. The skills needed in this space will change, but the number of jobs won't reduce as demand grows in response even as the cost of code approaches zero. </p></li><li><p>A few specialists will remain writing core engines of the software world - those that develop technical libraries, frameworks, and operating systems, etc. There have been very few individuals of this calibre and the revolution in coding agents isn't going to change that.</p></li><li><p>New graduates will end up falling in one of the above three buckets. Those that have more of a product bent might end up landing in the first bucket. Those that have the discipline and aptitude to learn through actual coding (when an all-powerful coding agent is a tantalizing keystroke away) and reading technical books will be valuable and employable, joining the latter two categories.</p></li></ul><p>A 10x individual of the future would be a combination of a top product manager and a top programmer of today. And just as today, they will be as rare to find. They will know how to:</p><ol><li><p>identify the right problem to solve</p></li><li><p>frame the problem so that finding the solution becomes possible</p></li><li><p>shape that solution into an elegant, maintainable, and scalable design</p></li></ol><div><hr></div><h2><strong>What does this mean for the enterprises?</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XNPf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XNPf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!XNPf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!XNPf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!XNPf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XNPf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png" width="1024" height="559" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:559,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:756643,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/182237010?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XNPf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!XNPf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!XNPf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!XNPf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61c7efba-72d3-4bc0-b81d-37d07c0757c1_1024x559.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ul><li><p>Adoption would be gradual but unrelenting. I think by this point it must be clear to even the staunchest of the skeptics that coding agents aren&#8217;t just a part of the AI hype. Some of the enterprises have already started allowing coding agents for in-house application development, and this number will only increase in 2026 and beyond.</p></li><li><p>Managing the complexity of AI-generated code will be challenging. To borrow from Fred Brooks, software systems always had two different types of complexity - essential complexity (the intrinsic complexity of the problem we&#8217;re actually solving) and accidental complexity (all implementation details we&#8217;ve accumulated to &#8216;manage&#8217; the application). Coding agents make the addition of accidental complexity incredibly easy. Imagine being able to write thousands of lines of code in a few hours where each thing &#8216;just about&#8217; works. Modularization and compositionality along with detailed specs will become ever more important to address this problem.</p></li><li><p>New SDLC methodologies for this new paradigm will spring into place as organizations employ coding agents. One can imagine a method whereby each module&#8217;s scope is kept narrow and specified to great precision, and code is produced every time any part of the spec is changed. This is CI/CD on steroids, and wouldn&#8217;t be possible without implementing the necessary controls (see next paragraph).</p></li><li><p>It is not enough for agent-written code to compile and pass unit tests. Enterprises will likely need clear IP provenance for generated code and secure-by-default pipelines that produce SBOMs and enforce supply-chain controls. I expect new controls like specs treated as versioned artifacts, automated functional and security testing, and policy checks as merge gates. Of course, AI will play a key role throughout this pipeline.</p></li><li><p>Organizational structures will have to change to reflect the new capabilities and best practices. I can see a large number of small and independent teams with each of them owning a significant chunk of the functionality, but centralized teams for ensuring architectural cohesiveness and managing operational data. </p></li><li><p>It is a mistake for organizations to stop investing in junior talent. While senior engineers bring experience, new graduates are natives of the AI era. They adapt to new workflows more naturally and will ultimately become the backbone of the industry.</p></li></ul><div><hr></div><p>In conclusion, I wouldn&#8217;t worry if I were a software engineer today or aspiring to become one. As the barrier between intent and code thins, software engineers will be valued for steering AI with precision and keeping complexity under control.</p><p></p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[December 2024 vs December 2025: What I Got Right (and Wrong) About AI]]></title><description><![CDATA[Here&#8217;s my evaluation of the predictions I made back in December 2024 with the current state of affairs vis a vis AI.]]></description><link>https://www.modelanalysis.ai/p/december-2024-vs-december-2025-what</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/december-2024-vs-december-2025-what</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Tue, 23 Dec 2025 19:20:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!eI_P!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Here&#8217;s my evaluation of the predictions I made back in December 2024 with the current state of affairs vis a vis AI. Hope you find this interesting. What were your predictions, and how did they pan out in 2025? Don&#8217;t forget to share your thoughts in the comments section!</p><p>What I wrote in 2024 is in <em>italics</em>. You can see that I got most of the predictions correctly but there were a couple of surprises as well.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eI_P!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eI_P!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!eI_P!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!eI_P!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!eI_P!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eI_P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png" width="724" height="394.81868131868134" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:724,&quot;bytes&quot;:5918353,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/182172310?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eI_P!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!eI_P!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!eI_P!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!eI_P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8423f595-7e6a-40bb-bd0e-ec93eb6eb56c_2816x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p><em>As 2024 concludes, I've been thinking about what's next for AI. Here are ten observations and predictions for 2025 - would love to hear your perspectives and what I might have missed!<br><br>1. Given the huge R&amp;D investment going into frontier AI, we&#8217;ll continue to see tremendous amounts of innovation and progress along multiple fronts.</em></p><p>This turned out to be very true. Hype and expectations continued to drive investment into AI research, engineering, infrastructure, and applications. LLMs dominated research, suffocating most other areas due to a lack of attention and funding. If LLMs don&#8217;t lead us to AGI, then consider this time and effort spent as a journey &#8220;off-ramp on the way to AGI,&#8221; as Yann LeCun put it.<br><br><em>2. Unlike much of traditional computer science where algorithms can be analyzed on paper, deep learning remains highly empirical - ideas must be tested at scale to confirm their viability. With more known ideas than can be developed and tested in 2025, expect progress to continue at pace for several years before any slowdown.</em></p><p>Deep learning remains very empirical though valiant efforts are being made by a few die-hard mathematicians to put the field on a solid theoretical footing.  </p><p>This <a href="https://ma-lab-berkeley.github.io/deep-representation-learning-book/Chx1.html">book</a> aims &#8220;to establish a principled and rigorous approach to understand deep neural networks and, more generally, intelligence itself.&#8221; Here&#8217;s a <a href="https://www.youtube.com/watch?v=QWidx8cYVRs">MLST video interview with Prof. Yi Ma</a>, a co-author of the book. Here&#8217;s <a href="https://www.youtube.com/watch?v=AWqvBdqCAAE">another video on MLST</a> that talks about applying category theory to deep learning.</p><p><em>3. Still, progress will continue to appear gradual to those directly involved, and will appear to take place in occasional big leaps to the general public.</em></p><p>This unfolded differently than I&#8217;d envisioned. Because of the intense competition between the frontier labs, new models and features were being released to the public without a lot of holding time. New models were announced regularly with reports of  better benchmark scores. Most casual users of the chatbots stopped noticing any discernible improvements, and in some cases wanted their favorite old model (OpenAI&#8217;s 4o) to be kept available. Image and video generation started to improve gradually (people consistently have 5 fingers per hand and text could finally be displayed without problems), reaching its pinnacle with Google Gemini&#8217;s Nano Banana!<br><br><em>4. LLM Maximalism (the idea that foundation models such as ChatGPT will lead to AGI with just more data and compute - say with version 6!) is dead. But transformer based LLMs are the closest thing to AGI we still have.</em></p><p>I was wrong here. LLM Maximalism got a second wind with reasoning models. Test-time scaling with Reinforcement Learning with Verifiable Rewards (RLVR) was exactly what was needed to unlock the intelligence hidden within the pre-trained model (which was created from training for next-token prediction on web-scale data). RLVR enabled code assistants to become the second killer app based on LLMs (the first was chatbots - in case you&#8217;re wondering).  <br><br><em>5. AI/ML techniques such as Reinforcement learning, search/planning, bayesian networks and logic will claim their rightful place next to pure deep learning based architectures. The path to AGI is pursued along one of these paths:<br>A. A reasoning system with LLM at its core complemented by RL and Search (e.g., OpenAI&#8217;s o3). Here the LLM sits sort of separately and search happens in token space.</em></p><p>The &#8216;reasoning system&#8217; implemented by o3, and then by all the frontier labs, turned out to be RLVR (see the previous point). A particular version of RLVR was revealed to the general public through this <a href="https://arxiv.org/abs/2501.12948">paper</a> by DeepSeek, a Chinese lab. <br><em><br>B. An enhanced LLM with reasoning taking place within the models&#8217; latent space (e.g., Meta&#8217;s recent COCONUT paper).</em></p><p>Latent reasoning is still being worked on. Perhaps we&#8217;ll see it in 2026.</p><p><em>C. Some other technique that doesn&#8217;t involve an LLM at all (e.g., JEPA).</em><br>JEPA is still being worked on, as are a few other architectures. With LLMs getting all attention and the investment dollars, this will likely remain this way in 2026.<br><br><em>6. AGI benchmarking remains challenging, as existing metrics either can be gamed or measure only necessary but insufficient conditions. While Fran&#231;ois Chollet is developing ARC-AGI-2, more effort is needed in this area.</em></p><p>AI continues to have a benchmark problem. I think many people, even those that have been closely following AI progress, underestimate the power of the incentive for LLM makers to target benchmark maxxing. </p><p>It's classic game theory - even though each of the frontier labs is aware that targeting benchmarks isn't very helpful for the model performance in the real-world, they have no choice but to invest in it because every model release is prominently accompanied by a benchmark scorecard. <br><br>Labs target benchmarks by training on specially created data matching test data (even if the exact benchmark tests are private, labs have access to examples and they can create more of them). RLVR makes benchmark maxxing even easier, btw.<br><br><em>7. Evaluators/verifiers will be developing rapidly as they are needed for reasoning and agentic systems.</em></p><p>Turns out this is a hard problem to solve outside of special areas like coding. LLMs (and even humans) are being used as judges in areas without definite answers. Here&#8217;s a <a href="https://eugeneyan.com/writing/llm-evaluators/">fantastic article on the use of LLMs as evaluators</a>.<br><br><em>8. Ever-smaller models will achieve capabilities currently limited to much larger models, bringing costs down and pushing more AI capabilities to edge devices.</em></p><p>This prediction turned out to be accurate. 2025 saw many small(er) models that can run on laptops/phones with surprisingly strong capabilities. This trend will continue into 2026 and beyond, and this promise is what supports Apple&#8217;s stock price :)<br><br><em>9. Research efforts (such as mechanistic interpretation) towards understanding and explaining foundational models will progress slowly given the relatively low investment. But this is an essential ingredient in addressing the alignment problem.</em></p><p>Mechanistic interpretability continues to progress, albeit at a very small scale. Anthropic continues to be the only major frontier lab that seems to focus a bit on the alignment problem (at least publicly), and the focus of the industry seems to be on progressing AI as quickly as possible.<br><br><em>10. I agree with the idea that Embodied AI needs its own foundational model, similar to how LLMs unified different natural language processing capabilities (such as translation, sentiment classification, conversation, etc.). Such a model, incorporating LLM-like knowledge and reasoning, would be crucial even for specific applications like self-driving. However, this breakthrough is unlikely in 2025.</em></p><p>This prediction turned out to be true. This isn&#8217;t an easy problem to solve. Here&#8217;s an article from Physical Intelligence that explains <a href="https://www.physicalintelligence.company/blog/olympics">why embodied AI is such a tough problem to crack</a>. It also gives you a glimpse of the current state of the art for robotics.</p><p></p>]]></content:encoded></item><item><title><![CDATA[An AI-themed crossword to kickstart your holidays!]]></title><description><![CDATA[UPDATE: Here&#8217;s the crossword web page I vibe-coded specifically for this puzzle.]]></description><link>https://www.modelanalysis.ai/p/an-ai-themed-crossword-to-kickstart</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/an-ai-themed-crossword-to-kickstart</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Fri, 28 Nov 2025 20:14:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!7JY1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>UPDATE: Here&#8217;s the <a href="https://vibe-crossword-ram-komarraju.netlify.app/">crossword web page</a> I vibe-coded specifically for this puzzle. You can interactively fill-in the puzzle, and of course reveal the solution.</strong></p><p>I like AI and I like crosswords, and I hope you share at least one of those two interests. So, dear reader, here&#8217;s an AI-themed crossword to kickstart your holiday season! I tried to make it relatively easy, but couldn&#8217;t quite avoid using a few technical acronyms and bits of jargon. Give it a try and let me know what you think. I&#8217;ll share the solution tomorrow in a follow-up post.</p><p>This is my first time constructing a crossword, so consider this an experiment. If there&#8217;s enough interest, I&#8217;ll put more time into crafting a Christmas edition with the style and polish of a Monday New York Times puzzle.</p><p><em>Note: I used crosserville.com to build this crossword. It&#8217;s free, and I found it intuitive and easy to use.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7JY1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7JY1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png 424w, https://substackcdn.com/image/fetch/$s_!7JY1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png 848w, https://substackcdn.com/image/fetch/$s_!7JY1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png 1272w, https://substackcdn.com/image/fetch/$s_!7JY1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7JY1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png" width="1350" height="1648" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1648,&quot;width&quot;:1350,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:454192,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/180200438?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7JY1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png 424w, https://substackcdn.com/image/fetch/$s_!7JY1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png 848w, https://substackcdn.com/image/fetch/$s_!7JY1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png 1272w, https://substackcdn.com/image/fetch/$s_!7JY1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43aa9024-b322-4248-888c-cbd4915d523b_1350x1648.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><br></p>]]></content:encoded></item><item><title><![CDATA[Why Is It So Hard to Get AI Agents to Work in the Enterprise?]]></title><description><![CDATA[Hint: it&#8217;s not just picking the wrong business case, bad data, or MCP or any other framework.]]></description><link>https://www.modelanalysis.ai/p/why-is-it-so-hard-to-get-ai-agents</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/why-is-it-so-hard-to-get-ai-agents</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Sat, 22 Nov 2025 18:23:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!I0rI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I0rI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I0rI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!I0rI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!I0rI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!I0rI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I0rI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png" width="1024" height="559" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:559,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1013034,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/179055343?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I0rI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!I0rI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!I0rI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!I0rI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86bb65c6-f8aa-460a-a384-3fe1d668da5a_1024x559.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The year 2025 was supposed to be the &#8220;Year of the Agent,&#8221; but every week seems to bring another study from a major university or a leading consulting firm explaining why agent deployment in the enterprise has been far slower than expected. Many resort to platitudes like &#8220;put the business case front and center,&#8221; or &#8220;focus on getting the data right,&#8221; as if not following those tenets is all that&#8217;s holding agents back. Others blame MCP, LangChain, or some other framework. Yet others put the blame on software engineers for not knowing how to build agents properly.</p><p>In other words, many people are prone to blame anything but the LLM, the brain of the agent. In the rest of this article, I will share the current limitations of LLMs that make it extremely hard to implement agents in the enterprise. I&#8217;ll then cover potential ways in which these limitations can be addressed.</p><p><em>&#171;Note that I don&#8217;t include &#8220;lack of continual learning&#8221; as a major limitation preventing us from deploying AI agents successfully in the enterprise. I&#8217;m assuming that addressing the below limitations is sufficient for this purpose, but might have to think more deeply about lack of continual learning at some point.&#187;</em></p><h3><strong>Hallucinations</strong></h3><p>A key limitation driving even state-of-the-art models to be unreliable agents is hallucination, whereby a model generates plausible but wrong responses (i.e., responses inconsistent with either the input context, the model&#8217;s previous outputs, or the knowledge contained in its training data).&#8221;</p><p>Models have always suffered from hallucinations, but the impact is far greater when they are employed as autonomous agents:</p><ul><li><p>There&#8217;s simply a <strong>lot more opportunity for something to go wrong</strong> with all the stuff that happens while an agent is executing a task: reasoning/planning, tool calling, long-term memory/context, and multi-turn execution. Models tend to hallucinate more when context is embedded with incorrect or conflicting information (which is common and often unavoidable in real-world agent deployments). Where an average human might recognize the incorrect information and choose to ignore it or correct it, a model may instead rationalize it and expand upon it, compounding the error. Once a hallucination makes it into the context window of a model, there is no easy way to identify and remove it.</p></li><li><p>Model hallucination rates remain far higher than many imagine (ranging anywhere from 3% to 30+%, depending on the benchmark and the type of hallucination being measured). What&#8217;s more, <strong>reasoning models</strong> used in agentic solutions <strong>hallucinate more than smaller/&#8217;dumber&#8217; models</strong>. </p></li></ul><p>Of course, frontier labs are aware of this and are actively working to improve the situation. For example, OpenAI&#8217;s recent work proposes shifting the reward structure of benchmarks so that confident errors are penalized and recognizing uncertainty is credited. However, this does not fully solve the core problem, and OpenAI admits that even with improvements, hallucinations would still occur. BTW, I was able to elicit hallucinations from GPT-5.1, released two week ago, using the same techniques that I used in the past.</p><h3><strong>Limited Memory and Context Rot</strong></h3><p>An agent&#8217;s &#8220;memory&#8221; is the set of all tokens fed into the context window with every turn. Consequently, an agent&#8217;s performance ultimately depends on what fits into that context window. But LLMs have only a limited context window on which the entire agentic solution ends up being built. This has led to the rise of &#8220;context engineering&#8221; as a profession.<sup> </sup>A context engineer has to decide what parts of the overall context (e.g., previous user instructions, any supporting data, LLM&#8217;s responses, and history of tool calls/responses) to embed within the context window at any time. </p><p>Context engineers also have to deal with &#8220;context rot.&#8221; You see, while a model should in theory be able to pay attention to everything within its context window, that&#8217;s rarely the case in practice. An LLM&#8217;s performance becomes increasingly unreliable as the context window grows. A user might state a critical fact (the &#8220;needle&#8221;) at the beginning of a conversation. By the time we reach the 15-minute mark, that critical fact might be buried in a hundred-thousand-token &#8220;haystack&#8221; of task execution history. Because of context rot, the LLM&#8217;s ability to &#8220;find&#8221; that needle is now <em>lower</em> than it was at the beginning.</p><p>Of course, frontier labs are working on increasing the context window length. This is a challenging task because attention computations increase quadratically in relation to the context window length. Any attention optimization techniques to control computational costs and speed need to avoid worsening the context rot problem.</p><h3><strong>Fragility to prompt phrasing</strong></h3><p>Another fundamental limitation is the sensitivity of LLMs to the exact phrasing and wording of inputs. Hence the explosion of prompt engineering as a discipline, complete with dozens of guides from<a href="https://docs.anthropic.com/claude/docs/prompt-engineering"> Anthropic</a>,<a href="https://platform.openai.com/docs/guides/prompt-engineering"> OpenAI</a>, and<a href="https://ai.google.dev/docs/prompt_best_practices"> Google</a>, and even &#8220;prompt optimizers&#8221; that algorithmically search for better prompts. Even if we follow all the guidelines, it is impossible in practice to control everything that makes it into the context window of an LLM, and a single word, a clause, or emotional language can produce dramatically different outputs/behaviors.</p><p>In messy real-world environments, inputs often come in varying formats and tones from users, tool/LLM outputs, and logs. Since the LLM can misinterpret any of these uncontrolled inputs (e.g., literally interpret and act on an emotional outburst from the user), the system stays inherently unpredictable. And if inputs truly <em>could</em> be controlled precisely, a deterministic workflow would often suffice without needing an LLM in the first place.</p><p>To sum it up, the burden should be on the LLM to behave robustly, not on developers to discover ritual incantations to coax the system into functioning correctly or to hope and pray that the context stays clean.</p><h3><strong>The Ultimate Dilemma: Deterministic Automation vs Intelligent Agents</strong></h3><p>Finally, and most importantly, there&#8217;s perhaps a central, unsolvable contradiction in current LLM-based agents:</p><ol><li><p>The <em>promise</em> of an agent is that it can reason, plan, and think creatively to solve a complex problem to perform tasks that could only be performed by humans until now.</p></li><li><p>The <em>requirement</em> of the &#8220;enterprise&#8221; is reliability, predictability, and auditable outcomes. </p></li></ol><p>And the contradiction between promise and requirement is that LLMs are not reliable, not predictable, and (perhaps) not auditable. </p><p>Large enterprises employ thousands, if not hundreds of thousands, of HGIs (human general intelligences). Wherever possible, these organizations continue to replace manual activities with deterministic automation because it&#8217;s more reliable and less expensive. LLMs are AGIs (<em>LLMs are artificial, and they have general intelligence, so I&#8217;ll call them AGIs, thank you</em>), but even SOTA versions are less reliable, less predictable, and less capable than HGIs across multiple dimensions.</p><h3><strong>The Way Forward</strong></h3><ul><li><p><strong>Build robust scaffolding around LLMs</strong></p><p>Coding agents are a great example of this. Of course, the nature of the domain lends itself to such scaffolding. It&#8217;s not that LLMs don&#8217;t hallucinate or make mistakes when coding. It&#8217;s just that there are often verifiable and deterministic ways to check for problems and feed the results back to the LLM to fix them. For example, a coding assistant can send lint results, compiler error/warning messages, runtime logs, unit test failures, etc. to the LLM to fix any issues. </p><p></p><p>But building such scaffolding for other use cases may not be possible or may prove to be very expensive. The scaffolding approach is where I think we are headed in the short to medium term, because I can&#8217;t see the fundamental limitations I mentioned above going away anytime soon, and organizations will find the promise and allure of LLM agents too irresistible not to try this approach.</p><p></p></li><li><p><strong>Limit LLM/agent usage to small/narrow steps within deterministic workflows</strong></p><p>The approach requires using LLMs in very specific steps within a larger workflow, controlling tightly what&#8217;s fed into their context window, thus increasing their reliability. While this might work, the benefit to be gained in this approach is likely to be very limited, and a traditional, fully deterministic solution would probably be cheaper to implement.</p><p></p></li><li><p><strong>Limit their usage to areas where failure isn&#8217;t critical or keep a human in the loop</strong></p><p>In this scenario, we recognize the limitations of LLMs and use them in areas where failure isn&#8217;t critical. Deep Research is a major example of this approach. The reports generated by deep research implementations contain multiple hallucinations, but the reports are usually meant for some human to use as part of their work rather than publishing them as end products. Of course, having humans in the loop doesn&#8217;t mean much if they lack the expertise or the patience to carefully read through the reports, use only what&#8217;s correct, and discard or fix inaccurate information.</p><p></p></li><li><p><strong>Improving the LLMs by improving the benchmarks and algorithms</strong></p><p>This one is for the model labs, and each of these is a challenging problem with no clear/known solution.<br></p><p>Change LLM training objectives and benchmarks to punish guessing and reward acknowledgment of a lack of knowledge. There&#8217;s probably a theoretical limit to how much hallucinations can be reduced, however.</p><p></p><p>Find ways to increase context window length and fix context rot. <br><br>Make the LLM more robust to minor changes/tweaks in prompts. This change will likely make the LLMs seem boring/non-creative as chatbots, and therefore we are likely to have dedicated LLMs trained specifically to be effective in agentic settings.</p></li></ul>]]></content:encoded></item><item><title><![CDATA[Anatomy of Frontier AI Systems in 2025]]></title><description><![CDATA[When you interact with ChatGPT, Claude, or Gemini, it may feel like a simple exchange with a single model.]]></description><link>https://www.modelanalysis.ai/p/anatomy-of-frontier-ai-systems-in</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/anatomy-of-frontier-ai-systems-in</guid><pubDate>Fri, 07 Nov 2025 06:25:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vwg9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>When you interact with ChatGPT, Claude, or Gemini, it may feel like a simple exchange with a single model. In reality, you&#8217;re interacting with a large distributed system whose complexity has been growing dramatically as the leading LLM providers continue to add new capabilities to their offerings. </p><p>Between the prompt (which can be multimodal with text, voice, images, and video) you enter and what you see in response sits a layered architecture of user interfaces, routers, orchestrators, memory stores, specialized models, tools, agents, code executors and content renderers. These systems have evolved from basic chatbots into modular platforms where choices about context, privacy, latency, and cost are made continuously and automatically. Understanding how the pieces fit together makes you a better user.</p><p>For developers and architects, this understanding explains why the same LLM apps can behave differently on different machines/users. The set of available tools, their configurations, the memory retrieved and embedded in the context, the routing, pre/post processing choices would ultimately determine the outcomes. It also shows them where to look for when trouble shooting or when looking to design an innovative agentic solution.</p><p>The diagram below shows one way in which such a system could be designed, followed by a description of each of the numbered parts. The actual implementation details would vary from provider to provider, of course.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vwg9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vwg9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png 424w, https://substackcdn.com/image/fetch/$s_!vwg9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png 848w, https://substackcdn.com/image/fetch/$s_!vwg9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png 1272w, https://substackcdn.com/image/fetch/$s_!vwg9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vwg9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png" width="1456" height="1089" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1089,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:255231,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/178242834?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vwg9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png 424w, https://substackcdn.com/image/fetch/$s_!vwg9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png 848w, https://substackcdn.com/image/fetch/$s_!vwg9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png 1272w, https://substackcdn.com/image/fetch/$s_!vwg9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F401a1048-df45-41d5-af6c-9d292210de28_1618x1210.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>1. The User</strong></h2><p>Every session starts with the user&#8217;s intent. You enter text, speak a command, or attach a file, and that input becomes the first step in a long pipeline. Even at this starting point, the system makes decisions on your behalf: which hidden instructions to include, what metadata to attach, and whether to collect additional local context such as timezone or project settings. Those invisible choices shape how the downstream components interpret your request and decide what to do next.</p><h2><strong>2. The Native App</strong></h2><p>The front-end is no longer a thin veneer that just supports text-in and text-out. Instead, today&#8217;s clients play an active role in capturing the intent, managing local state, and integrating with your device. A desktop assistant can listen for a wake word, read files you&#8217;ve explicitly shared, or summarize notifications. An IDE assistant can index repositories and extract build errors to included in the context. A browser client can extract useful context from the web page along with the user&#8217;s prompt to pass on to the server.</p><h2><strong>3. Client-Side Tools</strong></h2><p>Client-side tools ground the system in your environment. With your permission, the Native App may use your mic as input, scan selected documents, assemble a code context from your workspace, or even use MCP to make calls to external services (e.g., shopping carts, travel argents, calendars). Many clients parse chosen files, embed their contents, and even build an index for fast search and retrieval so that when you ask a question, the relevant snippets can be included in the prompt sent to the server. Done well, this creates personalization without overflowing the context. Sensitive material should stay on-device unless you explicitly opt in, and the client should only be sending the minimum information required to answer your question accurately.</p><h2><strong>4. Router &amp; Pre-Processor</strong></h2><p>Once the request leaves the device, a server-side router becomes the air traffic controller. It classifies intent, inspects metadata, applies safety and normalization, and selects the appropriate path: language model, coding model, image generator, video generator or an agent (e.g., Deep Research). The router also enriches your input with server-side context (conversation summaries, long-term memory, or provider policies/guard-rails) and templates along with the prompt.</p><p>Apart from providing access to different models providing a rich and diverse set of functionality, this routing capability allows the LLM provider to save money by running simple(r) queries on smaller/faster LLMs.</p><h2><strong>5. Image/Video Generation</strong></h2><p>If the user&#8217;s request needs generating images or video, the orchestrator invokes a generative model (distinct from LLMs) specialized for this purpose.</p><h2><strong>6. LLM Orchestrator</strong></h2><p>The Orchestrator calls the LLMs as identified by the Router. It may end up calling just one LLM, or a sequence of them (e.g., start with a smaller, fast model to parse tasks and draft a plan, escalate to a larger reasoning model for multi-step logic, and switch to a code-tuned model to generate or execute programs). Intermediate results are stitched together, cached where useful, and reintroduced into the context for subsequent steps. If the intermediate responses from the LLMs include a request to make a tool call on the server-side, the orchestrator makes those tool calls, adds the resulting content to the context and submits them back to the LLM for processing. See section 9 for more details.</p><h2><strong>7. LLMs</strong></h2><p>This is very simple really, a given LLM takes the context submitted by the Orchestrator as input and provides its response (one token at a time).</p><h2><strong>8. Agent Layer</strong></h2><p>For complex workflows, an agent such as Deep Research may sit between the orchestrator and tools. The agent is responsible for executing the &#8220;agentic&#8221; loop by leveraging all the tools and a suitable LLM.</p><h2><strong>9. Tools including Memory System &amp; Execution Environment</strong></h2><p>In order to respond to the user&#8217;s prompt, the LLM might choose to make a tool call. The Orchestrator interprets these calls, executes the tool calls and returns the results to the LLM. Web search grounds answers in fresh information. A calculator guarantees numerical precision. A code runner executes snippets and returns outputs. API connectors let the system perform external actions such as querying a travel database, reading a calendar, or placing an order on a popular shopping site. Each tool&#8217;s output is appended to the context and the model continues, forming a tight loop of &#8220;plan &#8594; act &#8594; observe &#8594; refine.&#8221;</p><p>Because base models are stateless and context-limited, the system relies on memory to provide continuity. Short-term memory supplies the immediate chat history. Long-term memory retrieves relevant facts from historical chats and any additional documents and artefacts (e.g., supplied as part of a project conversation). The orchestrator then augments the current prompt with only those pieces of information required for the task at hand. The result is an experience that can pick up threads from prior conversations or recall user preferences.</p><h2><strong>10) Post-Processor</strong></h2><p>The text that streams back to you is rarely the raw model output. A post-processor formats the response and applies safety checks. It may redact sensitive material, downgrade uncertain claims, or request another pass from the model with tighter constraints. </p><h2><strong>11) Response Delivery and Feedback</strong></h2><p>The response returns through the client, which can take local actions by invoking tools such as saving a file, updating a dashboard, or reading the output aloud. From your perspective, this might like a single conversation but under the hood, it&#8217;s a carefully managed pipeline balancing speed, accuracy, privacy, and price.</p><p><strong>In conclusion,</strong> when using LLMs from leading providers, we don&#8217;t interact directly with a model but with complex systems built around models. The LLM remains the centerpiece, but it is part of a broader operating stack where memory and tools define the experience as much as the LLM does.</p><p></p>]]></content:encoded></item><item><title><![CDATA[LLMs and Human Brains]]></title><description><![CDATA[Eight analogies exploring how they each learn]]></description><link>https://www.modelanalysis.ai/p/llms-and-human-brains</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/llms-and-human-brains</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Sat, 25 Oct 2025 22:00:14 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!O7wG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#171;Disclaimer: these analogies, like all analogies, aren&#8217;t perfect. There&#8217;s a big difference between humans and LLMs - we operate in a very messy physical world; LLMs operate in a digital world with less messiness but also don&#8217;t have our rich sensory experience. A full book, not a short article like this, would be needed to list out all the commonalities, exceptions, and nuances. Finally, to reach the maximum audience I&#8217;ve eschewed using technical jargon and details as much as possible. &#187;</em></p><p>LLMs are created from human knowledge, culture, and values expressed in text. So, it&#8217;s not a surprise when they act and behave like humans. At the same time, they are different from us in many important ways given that they are created using a completely different method compared to evolution (think of birds vs airplane). </p><p>In this article, I&#8217;ll develop this idea more fully, comparing some key LLM concepts to their human analogues. For those that aren&#8217;t familiar with the technical details of how LLMs are built and work this comparison should give a better and intuitive feel about LLMs. Even those that are intimately familiar with LLMs&#8217; training and inner workings may gain some additional insights. Or perhaps more importantly, they can point out in the comments section where I got things horribly wrong :)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!O7wG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!O7wG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!O7wG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!O7wG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!O7wG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!O7wG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1333480,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/177089928?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!O7wG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!O7wG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!O7wG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!O7wG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F719e3a48-68a2-44a4-9d0d-902cd240d9a1_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ol><li><p><strong>Evolution vs. Pre-training</strong></p><p>Evolution spent billions of years &#8216;searching&#8217; for brain designs and encoding the instructions in our DNA, governing not only our brains but our entire bodies.  A newborn baby&#8217;s brain doesn&#8217;t come with pre-built factual knowledge, but it has innate capabilities to learn from and operate in its environment. For example, it will soon be able to grasp things by hand and learn the particular language(s) it&#8217;s exposed to. </p><p><br>In comparison, LLMs can be created in a few weeks&#8217; training on trillions of words of text with the objective of guessing the next token (word) given a sequence of tokens (words).  The resulting neural network comes with vast pre-built knowledge and a limited world model. It doesn&#8217;t, however, know quite how to engage in conversations, be helpful, reason through problems or how to use tools.</p><p></p></li><li><p><strong>Childhood immersion/imitation learning vs. Supervised Fine-Tuning (SFT)</strong></p><p>Toddlers learn how to communicate through immersion and imitation. LLMs learn how to converse after being trained with thousands human-generated examples of prompts (&#8220;How are you today?&#8221;) and responses (&#8220;I am fine. Thank you. How are you?&#8221;). Using these examples, SFT teaches LLMs the basic structure of conversations and task following behavior. </p><p></p></li><li><p><strong>Parental guidance &amp; social conditioning vs. Reinforcement Learning through Human Feedback (RLHF)</strong></p><p>Feedback from parents and the broader social circle teaches children to behave in ways the children find rewarding, thus sculpting their communication and behavior. RLHF fine-tunes an already pre-trained, instruction-following LLM by incorporating human preferences as a reward signal. The model learns to generate responses that humans prefer. For example, responses that are polite, helpful, and honest.</p><p></p></li><li><p><strong>School or Structured Learning vs. Reinforcement Learning with Verifiable Rewards (RLVR)</strong></p><p>IMO, this is a tricky analogy to make, but I&#8217;ll give it a go. Humans learn how to apply logic, solve math problems and write programs through structured teaching, and are tested and corrected with objective tests. To achieve a similar outcome in the case of LLMs, we take problems with verifiable answers and feed them to LLM - we reward them when they produce correct answers. Note that we don&#8217;t really tell them how to solve problem the way we teach humans. We just give them the problem and just reward them when they get it right, and somehow LLMs learn how to allocate more effort and &#8216;reason&#8217; to solve the problems. </p><p></p></li><li><p><strong>Lifelong learning and long term memory vs. In-context learning (ICL) &amp; short term memory</strong></p><p>Humans continue to learn throughout their lives. New experiences, skills, and knowledge gained are compressed and stored into their long term memory (sleep perhaps plays a role here), enabling later recall and reuse. Once training is complete, LLMs&#8217; knowledge and skills are frozen when they&#8217;re released to production. They have a very limited amount of working memory (called the context), and can learn new skills if we provide them with a description and/or a few examples within that context. This is called in-context learning, and it&#8217;s lost when the short term memory overflows or when a new context is created.</p><p></p></li><li><p><strong>Rules and Laws vs. System Prompts</strong></p><p>Humans are governed by rules and laws that they are meant to follow. LLMs are given a similar set of guidelines and rules via system prompts (which reside at the beginning LLM&#8217;s working memory/context). Of course, just like humans, LLMs may or may not follow those rules :)</p><p></p></li><li><p><strong>Reference materials vs. Placing information in working memory</strong></p><p>Humans aren&#8217;t expected to memorize every fact. They can look them up in books or online. A prompt to an LLM can also embed any additional information that the LLM can use to better answer the prompt.</p><p></p></li><li><p><strong>Human Tool use vs. LLM Tool use</strong></p><p>Humans have an innate ability to use tools in general, and during their lives they learn how to use specific tools through practice and demonstration. LLMs see code, APIs, and tool-related text during pre-training, which give them a rough sense of what &#8220;tool use&#8221; is. They are then given specific tool use examples during supervised fine-tuning and are rewarded for correct tool use during the reinforcement learning (RLHF/RLVR) phases. As mentioned above, LLMs can also learn how to use specific tools in-context when provided with a description and/or a few examples. This capability lasts while that context is active and is not carried forward into future sessions.</p></li></ol><p>In conclusion, the comparison isn&#8217;t perfect, but it reveals to us a little bit about how the two intelligences are similar and different. We are the result of a billion years of evolution, while LLMs are imperfect (or perhaps simply different) digital reflections of ourselves.</p>]]></content:encoded></item><item><title><![CDATA[Determinism, Truth, Creativity and Hallucinations: Untangling Often-Confused Ideas About LLMs]]></title><description><![CDATA[Why the word &#8220;determinism&#8221; keeps tripping people up (and why it matters)]]></description><link>https://www.modelanalysis.ai/p/determinism-truth-creativity-and</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/determinism-truth-creativity-and</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Sun, 21 Sep 2025 13:43:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!lSli!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Public discussions about LLMs often blur different ideas: determinism, truth, creativity and hallucinations. A recent wave of discussion starting with the publication of <em>Thinking Machine&#8217;s <strong>Defeating Nondeterminism in LLM Inference</strong></em><strong> </strong>surfaced this confusion even more. Determinism gets casually equated with truth and non-determinism with creativity and hallucinations. It&#8217;s therefore understandable that many casual followers of AI took this to mean that Thinking Machine figured out how to eliminating hallucinations and get to truth in LLMs. That&#8217;s understandable: stable outputs <em>feel</em> more reliable. But it&#8217;s wrong in two ways. First, a system can be perfectly deterministic and still be confidently, consistently wrong. Second, you don&#8217;t need non-determinism for a model to be creative or hallucinate. </p><p>This confusion isn't just academic; when these wires cross, people misread articles, over- or under-react to AI press releases, and make poor executive decisions. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>In this article, we&#8217;ll define in plain English what each of these concepts mean in the context of LLMs. Then we&#8217;ll attempt to correct common misconceptions, each with concrete examples. As a bonus for the technically minded among you, we&#8217;ll walk through how an LLM chooses the next token, highlighting intentional variability from unintentional variability. By dissecting the journey from raw scores to the final token, we&#8217;ll pinpoint exactly where true randomness is introduced and where unintentional system quirks can arise.</p><div><hr></div><h2><strong>Definitions</strong></h2><p><em>&#8220;When I use a word,&#8221; Humpty Dumpty said, in rather a scornful tone, &#8220;it means just what I choose it to mean&#8212;neither more nor less.&#8221; <br></em>&#8212; Humpty Dumpty, in Lewis Carroll's <em>Through the Looking-Glass</em></p><p><strong>Determinism.</strong> This refers for the ability to produce the same output from the same input (assuming the same model, software, drivers, and hardware). In practice, this refers to whether the model produces the same next-token prediction scores (logits) given the same input. It&#8217;s about <strong>reproducibility</strong> of the system&#8212;not about being right or creative.</p><p><strong>Truth.</strong> Whether a claim matches the real world. It's useful to consider this in two categories: objective and subjective truth.</p><ul><li><p><strong>Objective Truth or Factuality:</strong> For verifiable facts like, "What is the boiling point of water?" the model's output is "true" if it aligns with established, real-world knowledge. Its ability to produce objective truth is driven by the accuracy of its training and in-context data, and its own capability to use this data correctly.</p></li><li><p><strong>Subjective Truth with Representativeness:</strong> For matters of opinion, interpretation, and values like, "Is abstract art meaningful?" the model essentially synthesizes the vast spectrum of human perspectives from its training data to generate a plausible or representative viewpoint. This means the default pathways in the model reflect a majority or dominant cultural opinion, but specific directions in the prompt can steer the model to reflect a specific viewpoint.</p></li></ul><p><strong>Creativity.</strong> From a user's perspective, creativity in an LLM is the model's ability to produce text that feels novel, interesting, and non-obvious, while still being coherent and relevant to the prompt. It's the capacity to generate a new poem or a unique marketing slogan. This user experience of creativity is technically derived from <strong>intentional and</strong> <strong>controlled randomness</strong> in the token selection process (covered later in the article). </p><p><strong>Hallucinations.</strong> A hallucination is said to occur when an LLM generates text that is plausible and confident, but is factually incorrect. At its core, it is a failure of the model to <strong>distinguish between its internal, statistical representation of the real world (facts) and plausible-sounding fiction</strong>. Because the model's primary goal is to generate a fluent sequence of words, it can sometimes blend the <em>style</em> of a factual statement with the <em>content</em> of a fictional one. </p><p><strong>Active research is taking place to eliminate or reduce hallucinations</strong>: The fact that AI can be confidently wrong is one of the biggest hurdles to its reliable use. And of course there are major ongoing efforts focused on building in models the crucial ability of knowing when they don't know something. Here&#8217;s a recent <a href="https://openai.com/index/why-language-models-hallucinate/">post</a> from OpenAI, and another very recent <a href="https://research.google/blog/making-llms-more-accurate-by-using-all-of-their-layers/">one</a> from Google.</p><div><hr></div><h2><strong>Common Misconceptions</strong></h2><h3>Misconception #1: Deterministic responses are truthful responses.</h3><p><strong>Reality:</strong> Determinism guarantees repeatability, not truthfulness. In the context of LLMs, determinism can guarantee the same next-token probability distribution, but that doesn&#8217;t make it more true. </p><p><strong>Example:</strong> To the question &#8220;What is the capital of Australia?,&#8221; the LLM can answer every time with &#8220;Sydney&#8221; but that doesn&#8217;t make it right.</p><p><strong>Why it matters:</strong> Deterministic outputs lead to consistent model behavior, and help measurement and testing; they don&#8217;t guarantee factuality.</p><h3>Misconception #2: Non-determinism is what leads to Hallucinations </h3><p><strong>Reality:</strong> Hallucinations are the result of the model's inability to distinguish between when it knows something and when it doesn&#8217;t know something. A model can produce a fabricated "fact" via a perfectly deterministic path. Making the output deterministic doesn't make it true.</p><p><strong>Example:</strong> To the question &#8220;Who won the gold in 100m sprint in 2028 Oympics?,&#8221; the LLM can deterministically answer every-time with &#8220;Dean Jones&#8221; but that doesn&#8217;t mean it has basis in reality. Even if the event hasn&#8217;t happened yet, the model may  fabricate a plausible name.</p><h3>Misconception #3: Non-determinism is Necessary for Creativity</h3><p><strong>Reality:</strong> Creativity in LLMs comes from <strong>intentional randomness</strong> introduced through decoding strategies like temperature and top-p sampling. Non-determinism can also arise from <strong>unintentional system variability</strong> (e.g., batching effects), which produces random drift without creative intent.</p><p><strong>Example:</strong> A prompt run twice produces slightly different outputs due to batch variance, but both continuations are bland and repetitive. This is non-deterministic but not creative.</p><p><strong>Why it matters:</strong> Conflating the two leads to frustration; true creativity requires deliberately tuning the right knobs (like temperature), not just accepting random system quirks.</p><h3>Misconception #4: Creativity leads to falsehoods</h3><p><strong>Reality:</strong> Creativity is the ability to generate diverse and novel&#8212;but still plausible&#8212;continuations. A model can be creative while remaining entirely factual, or it can be creative in a fictional context. The output's truthfulness is a separate, unrelated dimension.</p><p><strong>Example:</strong> Asking an LLM to "explain photosynthesis in the style of a pirate" can produce a creative and engaging answer that is still factually correct.</p><p><strong>Why it matters:</strong> Dismissing all creative outputs as false means missing out on the LLM's power to brainstorm, rephrase, and explain concepts in more engaging ways.</p><div><hr></div><h2><strong>A Few Common Technical Misconceptions</strong></h2><h3><strong>Technical Misconception #1: Temperature=0 makes answers factual.</strong></h3><p><strong>Reality:</strong> Setting temperature to zero path removes decoding policy randomness. It doesn&#8217;t verify facts.</p><p><strong>Example:</strong> The model repeats an outdated statistic confidently; turning down temperature only makes the mistake consistent.</p><p><strong>Why it matters:</strong> Claims that &#8220;lower temperature improves accuracy&#8221; are over-reading a stability knob.</p><h3><strong>Technical Misconception #2: Setting a seed makes everything deterministic.</strong></h3><p><strong>Reality:</strong> The seed is only used in the Token Sampling step. It ensures that the <em>random choice</em> from the final candidate pool is repeatable. It has no effect whatsoever on the <em>decoding policy</em> introduced randomness or <em>unintentional variability</em> that can occur back in the inference phase, which can change the logits themselves and lead to a different output even with the same seed.</p><p><strong>Example.</strong> Two runs with the same seed diverge because one was processed in a larger batch, nudging the logits.</p><p><strong>Why it matters.</strong> &#8220;Seeded for determinism&#8221; means &#8220;sampling is fixed,&#8221; not &#8220;the entire stack is invariant.&#8221;</p><h3><strong>Technical Misconception #3: Batch-size nondeterminism is just floating-point voodoo.</strong></h3><p><strong>Reality.</strong> Floating-point quirks are part of it, but the another root cause is <strong>order of operations</strong> changing with batch shape or scheduling.</p><p><strong>Example.</strong> The same prompt answered alone vs. grouped with seven others takes a slightly different compute path causing tiny logit shifts flip a softmax later.</p><p><strong>Why it matters.</strong> When you see &#8220;same prompt, different output,&#8221; ask whether the inference setup changed. The concept to look for is batch invariance. Thinking Machine&#8217;s <a href="https://thinkingmachines.ai/blog/defeating-nondeterminism-in-llm-inference/">paper</a> proposes a way to address this.</p><h3><strong>Technical Misconception #4: &#8220;Deterministic&#8221; means identical across all machines and versions.</strong></h3><p><strong>Reality.</strong> Most real guarantees are scoped per same weights, tokenizer, libraries, drivers, and GPU family. Change any of those and tiny numeric changes can appear.</p><p><strong>Example.</strong> Bit-for-bit identical on an A100/CUDA stack does not imply bit-for-bit identical on an H100/CUDA.</p><p><strong>Why it matters.</strong> When reading papers or provider claims, look for the <strong>determinism envelope</strong> to understand what exactly the promise covers.</p><div><hr></div><h2>Anatomy of Token Generation: A Step-by-Step Guide</h2><p>At its core, an LLM is a sophisticated prediction engine. Its main job is to answer the question: "Given this sequence of words (tokens), what word (token) should come next?" The diagram below illustrates this process focusing on the token selection, which we can break down into seven key steps.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lSli!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lSli!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png 424w, https://substackcdn.com/image/fetch/$s_!lSli!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png 848w, https://substackcdn.com/image/fetch/$s_!lSli!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png 1272w, https://substackcdn.com/image/fetch/$s_!lSli!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lSli!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png" width="1456" height="663" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:663,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:226912,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/173483948?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lSli!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png 424w, https://substackcdn.com/image/fetch/$s_!lSli!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png 848w, https://substackcdn.com/image/fetch/$s_!lSli!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png 1272w, https://substackcdn.com/image/fetch/$s_!lSli!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470faa8f-edde-4849-8dbe-9fe0d0dfd032_1762x802.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ol><li><p><strong>The Model Generates Raw Scores (Logits).</strong> The model processes the prompt + prior tokens to produce logits (raw, unnormalized scores) for every token in the model&#8217;s vocabulary. This is where unintentional variability resulting from the numerical drift from batching, kernel order, or hardware tactics can creep in.</p></li><li><p><strong>Scores are Converted to Probabilities (Softmax).</strong> The raw logits are then passed through a <strong>Softmax</strong> function. This is a crucial mathematical step that transforms the long list of scores into a proper probability distribution, where the probabilities of all possible tokens add up to 1. This is also where the most important control for creativity, temperature, is applied. As the diagram shows, temperature controls the shape of the probability distribution:</p><ul><li><p><strong>High Temperature:</strong> Flattens the distribution. This boosts the probability of less likely tokens, making the model's output more diverse, creative, and sometimes surprising.</p></li><li><p><strong>Medium Temperature:</strong> Creates a "spiky" distribution that still strongly favors the most probable tokens but allows for some variation. This is often a good balance between creativity and coherence.</p></li><li><p><strong>Temperature = 0:</strong> This is a special case. It creates a distribution with a single peak, giving a 100% probability to the single most likely token and a 0% probability to everything else.</p></li></ul></li><li><p><strong>The Greedy Decoding Path</strong>. When temperature=0, one token gets the maximum probability of 1, and the model selects it directly. This is called greedy decoding. It's the most direct and predictable path, completely bypassing the need for any further decoding or sampling.</p></li><li><p><strong>Decoding policy.</strong> If you want diversity, you deliberately introduce randomness by using a decoding policy to create the final candidate pool of tokens you&#8217;ll sample from. </p></li><li><p><strong>Creating the candidate pool of tokens.</strong> For any temperature above zero, the model has a distribution of options. The two most common methods are:</p><ul><li><p><strong>Top-K:</strong> A simple rule that tells the model to only consider the 'k' tokens with the highest probabilities.</p></li><li><p><strong>Top-P:</strong> A more dynamic rule that tells the model to consider the smallest group of top tokens whose probabilities add up to at least a certain percentage 'p'.</p></li></ul></li><li><p><strong>Token Sampling (seed=&#8230;). </strong>From this final, filtered group, one token is chosen through a weighted random selection. A token with a higher probability has a higher chance of being picked, but other tokens may be selected as well with their corresponding probability. This is the step where a seed is used. By providing the same seed, you ensure that the same <em>pseudo-random</em> choice made at this specific step is the same every time. This makes the sampling process repeatable.</p></li><li><p><strong>The loop repeats (autoregression).</strong> The chosen token (from greedy or sampling) is appended to the input and the loop repeats. The entire process, from Step 1 to Step 6, then repeats to generate the token that comes after it. This cyclical, one-word-at-a-time process is called autoregression.</p></li></ol><p>Two purple callouts in the diagram reinforce the distinction:</p><ul><li><p><strong>Unintentional variability</strong> sits within model inference</p></li><li><p><strong>Intentional randomness</strong> sits with your decoding policy&#8212;what you dial to get creativity or stability.</p></li></ul><div><hr></div><h3>Conclusion: Mastery Begins with Clarity</h3><p>The language we use to describe LLMs is a minefield of misunderstood concepts, where technical terms are often used as proxies for desired outcomes. As we've seen, determinism is not a synonym for truth, creativity does not have to imply hallucinations, and random token selection doesn&#8217;t negate determinism. Instead, these are distinct dimensions of a model's behavior, each controlled by different mechanisms.</p><p>The path to building and deploying AI responsibly doesn't just depend on bigger models; it depends on the clarity of our thinking. </p><p></p><p></p><p></p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Holy Grail of AI in Trading: Can We Finally Forecast the Markets?]]></title><description><![CDATA[Challenges, past attempts, and a blueprint for building a multimodal fusion model for financial prediction.]]></description><link>https://www.modelanalysis.ai/p/the-holy-grail-of-ai-in-trading-can</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/the-holy-grail-of-ai-in-trading-can</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Sun, 07 Sep 2025 03:03:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Zs07!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2><strong>Preamble</strong></h2><p>Inspired by Asimov&#8217;s Hari Seldon and psychohistory, I&#8217;ve carried around an idea for a novel for years. The novel&#8217;s protagonist, a brilliant but flawed mathematician, finds the holy grail of applied AI in trading: an app that forecasts financial markets with uncanny precision. The app would take in torrents of data in real-time and would make probabilistic predictions minutes, hours, and days into the future. His challenge, of course, is to keep it a secret while profiting from it because making it public would stop it from working. </p><p>I never wrote that novel. Mostly because life got in the way, but I never stopped thinking about how to create that market forecaster. All this time a market forecasting app that can work with the accuracy of a weather forecasting app seemed more like a piece of science fiction than an engineering project. With all the recent advances in deep learning, however, creating such an app feels a lot more feasible now. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>In this article, I&#8217;ll talk about why market forecasting is such a hard problem to solve, look at some of the notable attempts at solving it, and propose an architectural vision for a model that might finally get us there.</p><div><hr></div><h2>The Herculean Challenges of Market Prediction</h2><p>At first glance, market prediction might not look daunting for the uninitiated. After all, if we can create planet-scale weather forecasters and LLMs that write poems, produce code, and even win gold medals at the International Math Olympiad, surely building a narrower, special-purpose system ought to be easier? A market forecaster doesn&#8217;t need to understand Tolstoy <em>and</em> TikTok memes (or does it?); it only needs to make sense of prices and events to make decent probabilistic predictions of future prices.</p><p>But forecasting markets isn't like predicting the weather, where the underlying physics is mostly understood. Neither is it like creating a general purpose large language model. Markets are a different beast entirely, plagued by a few fundamental problems.</p><p>The first major hurdle is the problem of <strong>non-stationarity</strong>. In simple terms, a stationary process is one whose statistical properties (e.g., mean, variance and correlations) don't change over time. And financial markets are <strong>maddeningly non-stationary</strong>. The rules of the game are constantly shifting. A pattern that worked beautifully for a year can suddenly vanish because of a viral outbreak, a regulatory change, the outbreak of a war, a technological disruption, or a shift in the geo-political environment. This is why simple models like ARIMA (Autoregressive Integrated Moving Average), which rely on past patterns repeating, often fail spectacularly. Any successful solution, therefore needs a sophisticated, <strong>adaptive</strong> <strong>world model</strong> that truly understands and adapts to the underlying regimes and dynamics.</p><p>The second major hurdle is <strong>data</strong>: </p><ul><li><p>Building a true world model requires an <strong>unimaginable amount of diverse, high-quality, and perfectly time-stamped data</strong>. We need everything: real-time news feeds, social media feeds from platforms like X, geopolitical event data, central bank announcements, earnings call transcripts, corporate actions, earnings reports, weather forecasts, satellite imagery of oil tankers, and of course, all the financial time-series data including stock prices, trading volumes, FX rates, interest rates, and bond yields. </p></li><li><p>Also, this data is <strong>multimodal</strong>, coming in the form of free text, structured data and maybe even images and video. A successful model must be able to fuse these different streams together to form a holistic picture. </p></li><li><p>Acquiring the data is also extremely challenging as much of the high-quality <strong>data is</strong> <strong>fragmented and locked behind expensive paywalls</strong>. </p></li><li><p>This also requires careful engineering to avoid <strong>data leakage</strong>, ensuring that our model isn't accidentally "seeing" information from the future during training, which would make it look deceptively brilliant in backtesting but useless in the real world. </p></li><li><p>Of course, the world of finance has adversarial players, and data of all varieties and especially those from social platforms <strong>can be poisoned</strong>. Defensive mechanisms need to be in place to identify and remove such data. Perhaps a sophisticated model would learn to ignore such data just as a expert human would!</p></li></ul><p>Finally, there&#8217;s the problem of <strong>reflexivity</strong>. If a forecasting engine gets widely adopted, the very action of trading off its predictions can change market behavior, weakening its predictive power. This might seem academic, but it can disincentivize firms from investing heavily in building such models as keeping them proprietary can be very difficult (their employees who know how to build the model can be poached by competition).</p><div><hr></div><h2>A Brief History of Market Forecasting Attempts</h2><p>Early attempts relied on statistical methods like <strong>ARIMA and GARCH</strong>. While foundational for time-series analysis, they don&#8217;t take into account all the non-time-series information available, and are ill-suited for the complex, multivariate, and ever-changing nature of financial markets. The machine learning revolution brought more powerful techniques, but many early models were still limited, often focusing on a narrow set of technical indicators.</p><p>The success of deep learning and transformers attracted researchers&#8217; attention, and one of the first notable efforts was <strong>FinBERT</strong>, a model based on Google's BERT architecture but fine-tuned on a large corpus of financial text to understand the sentiment of financial news. However, FinBERT is quite dated at this point, and is not a forecasting engine. <strong>BloombergGPT</strong> made headlines in 2023 as a massive language model trained from scratch on Bloomberg's vast, proprietary dataset of financial information spanning decades. Like FinBERT, its primary goal wasn't to be an end-to-end forecasting system. It's a financial knowledge base, not a predictive machine that integrates real-time data to make predictions.</p><p>Since 2024, there&#8217;s been a new wave of research projects that are moving closer to what I think is the right approach. With efforts like <strong>MSMF (Multi-Scale Multi-Modal Fusion)</strong> researchers have started moving in the right direction, attempting to fuse textual data like news with price data. Similarly, agent-based simulators like <strong>FinArena</strong> aimed to create realistic environments to test trading strategies. While these are important steps in the right direction, they are  academic exercises trained on limited public data and haven't yet been deployed at the scale or with the breadth of data needed to build a true market-spanning forecasting engine.</p><div><hr></div><h2>The Proposed Solution: A Multimodal Fusion based Market Forecaster </h2><p>So, here&#8217;s my vision: a <strong>multimodal fusion forecasting architecture</strong> that ingests all kinds and modalities of financial data simultaneously to build a unified understanding of market&#8217;s state, dynamics and trajectory to provide probabilistic forecasts across varying time horizons.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Zs07!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Zs07!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png 424w, https://substackcdn.com/image/fetch/$s_!Zs07!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png 848w, https://substackcdn.com/image/fetch/$s_!Zs07!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png 1272w, https://substackcdn.com/image/fetch/$s_!Zs07!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Zs07!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png" width="1456" height="872" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:872,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:321327,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/172984503?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Zs07!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png 424w, https://substackcdn.com/image/fetch/$s_!Zs07!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png 848w, https://substackcdn.com/image/fetch/$s_!Zs07!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png 1272w, https://substackcdn.com/image/fetch/$s_!Zs07!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3c9865c-b5bf-4327-af98-127ab475e19d_2000x1198.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nLXu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nLXu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png 424w, https://substackcdn.com/image/fetch/$s_!nLXu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png 848w, https://substackcdn.com/image/fetch/$s_!nLXu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png 1272w, https://substackcdn.com/image/fetch/$s_!nLXu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nLXu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png" width="1456" height="786" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:786,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:224669,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/172984503?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nLXu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png 424w, https://substackcdn.com/image/fetch/$s_!nLXu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png 848w, https://substackcdn.com/image/fetch/$s_!nLXu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png 1272w, https://substackcdn.com/image/fetch/$s_!nLXu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2ea691d-09cb-4f86-82c0-74c61052c495_1994x1076.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jAuT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jAuT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png 424w, https://substackcdn.com/image/fetch/$s_!jAuT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png 848w, https://substackcdn.com/image/fetch/$s_!jAuT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png 1272w, https://substackcdn.com/image/fetch/$s_!jAuT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jAuT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png" width="1456" height="948" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:948,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:292335,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/172984503?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jAuT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png 424w, https://substackcdn.com/image/fetch/$s_!jAuT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png 848w, https://substackcdn.com/image/fetch/$s_!jAuT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png 1272w, https://substackcdn.com/image/fetch/$s_!jAuT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8a524e2-6572-4d51-b627-8159bf18331b_2016x1312.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!c-Wf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!c-Wf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png 424w, https://substackcdn.com/image/fetch/$s_!c-Wf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png 848w, https://substackcdn.com/image/fetch/$s_!c-Wf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png 1272w, https://substackcdn.com/image/fetch/$s_!c-Wf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!c-Wf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png" width="1456" height="1114" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c2153583-d1c2-4441-8088-490de2921353_2024x1548.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1114,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:226725,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/172984503?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!c-Wf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png 424w, https://substackcdn.com/image/fetch/$s_!c-Wf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png 848w, https://substackcdn.com/image/fetch/$s_!c-Wf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png 1272w, https://substackcdn.com/image/fetch/$s_!c-Wf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2153583-d1c2-4441-8088-490de2921353_2024x1548.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!glTe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!glTe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png 424w, https://substackcdn.com/image/fetch/$s_!glTe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png 848w, https://substackcdn.com/image/fetch/$s_!glTe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!glTe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!glTe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png" width="1456" height="770" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:770,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:292931,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/172984503?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!glTe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png 424w, https://substackcdn.com/image/fetch/$s_!glTe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png 848w, https://substackcdn.com/image/fetch/$s_!glTe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!glTe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b79e3f5-f4db-4b96-9aca-ea5211e38493_1966x1040.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>Conclusion</h2><p>I'm confident that a model built on this architecture would outperform a human analyst or a specialized, single-purpose algorithm. It would make decisions based on a more complete and holistically processed set of information than anything that has come before. I also admit that it&#8217;s extremely difficult and expensive to build with massive data, compute and engineering demands, and could potentially face regulatory obstacles. Then there&#8217;s the challenge of reflexivity: models that forecast may, if broadly deployed, invalidate themselves.</p><p>With so much of the AI world currently focused on the grand challenge of Artificial General/Super Intelligence (AGI/ASI), it's unlikely that major labs like Google or OpenAI would pivot to such a specialized, high-risk financial project. This leaves the field open to a few specific players. A data behemoth like Bloomberg, which sits on a large proprietary dataset, is one obvious candidate. Another is a well-funded, AI-native quantitative hedge fund like High Flyer (which owns DeepSeek). They might have the resources, the incentive, and the right kind of DNA.</p><p>Smaller, more constrained versions could probably be built today to prove the concept. One could start by focusing on a single asset class, like FX or a specific sector of the stock market. But ultimately, I believe a comprehensive, all-encompassing fusion model is the only path to obtain the holy grail of market forecasting.</p><p>Psychohistory may still be science fiction, but with multimodal fusion, a market forecaster is closer to become real.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[When LLMs Try to Draw: A Weird Zoo of AI Line Art]]></title><description><![CDATA[Or: Vibe-coding my way through Cursor, I got GPT, Gemini, and Claude to sketch animals with JSON polylines &#8212; no diffusion, no code, just vibes.]]></description><link>https://www.modelanalysis.ai/p/when-llms-try-to-draw-a-weird-zoo</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/when-llms-try-to-draw-a-weird-zoo</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Mon, 25 Aug 2025 03:05:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!AzPu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A few days ago, I stumbled across an intriguing <a href="https://www.lesswrong.com/posts/xwdRzJxyqFqgXTWbH/how-does-a-blind-model-see-the-earth">LessWrong post</a> where someone used different LLMs to describe Earth's geography in the simplest way possible - just by asking it to tag particular latitude and longitude into land from water. And the <a href="https://www.lesswrong.com/posts/xwdRzJxyqFqgXTWbH/how-does-a-blind-model-see-the-earth">results</a> were fascinating. </p><p>This got me thinking: if language models can conceptualize geography, what happens when you ask them to create visual art? Not generate images through diffusion models, but actually <em>construct</em> line drawings using pure coordinate geometry. How well can an LLM, not explicitly trained, on this task can do this? </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Thus began my weekend project: getting 9 different language models to draw animals using nothing but JSON polylines, then rendering those drawings into actual images. The twist? I wouldn't write a single line of code myself.</p><h2>The Method: "Vibe Coding" with Cursor</h2><p>Instead of painstakingly coding a pipeline, I &#8220;vibe coded&#8221; it in <a href="https://cursor.sh/">Cursor</a>. My process was beautifully simple:</p><ol><li><p>Open Cursor and describe my vision and let Cursor write the code. Nine models to start with (the list can be customized) where each model received identical prompts requesting JSON output with normalized coordinates (0-1 space) and named polyline strokes. The five animals I picked were: African elephant, giraffe, domestic dog, domestic cat, and the bald eagle.</p></li><li><p>More often than not Cursor is able to detect when things went wrong and fixed them itself, which is amazingly helpful! It handled: API integration quirks (GPT-5 models don't support temperature=0!?), error handling for malformed responses, batch processing logic, and even helpful emoji-filled progress indicators. The code quality is genuinely impressive, better than what many humans would write. Occasionally I have to explicitly describe a problem to Cursor to get it fixed, or instruct it with particular tweaks.</p></li><li><p>Repeat until everything worked</p></li></ol><p>The result? An evaluation pipeline consisting of:</p><ul><li><p>A JSON generation orchestrator that prompts different models and collect their JSON-based &#8220;line drawings&#8221;</p></li><li><p>A matplotlib-based renderer that converts JSON polylines into PNG images</p></li><li><p>A config loader for API key management and other configuration</p></li></ul><p>Here&#8217;s the lineup of nine contestants: gpt-4o, gpt-5-nano, gpt-5-mini, gpt-5, gemini-1.5-pro, gemini-2.5-flash, gemini-2.5-pro, claude-sonnet-3.7, claude-sonnet-4</p><div><hr></div><h2><strong>The Results: A Weird Zoo of Wobbly Outlines</strong></h2><ol><li><p><strong>Within the same family of models, bigger variants usually did better than the smaller ones.</strong> Smaller models often devolved into geometric blobs. As you see below GPT-4o did absolute the worst. Gemini-2.5-flash produced what I think are the best drawings. <strong>Claude Sonnet 4</strong>&#8217;s dog looked like an alien donkey with antennae. <strong>GPT-5-nano</strong> went for something rodent-like. Definitely a weird kind of zoo. <strong>Gemini-2.5-pro</strong> even added hexagonal &#8220;spots&#8221; - an imaginative touch! </p></li><li><p>Broadly speaking, <strong>Google&#8217;s Gemini generated better drawings than OpenAI&#8217;s GPT models</strong>. Anthropic&#8217;s Claude series had issues generating valid JSONs. A different prompt with more detailed instructions helped Claude prduce more workable drawings, but I haven&#8217;t included them to allow for apples to apples comparisons.</p></li><li><p><strong>Latency and API woes.</strong> All GPT-5 variants were quite slow: mini and nano averaged just over 67 seconds per prompt, GPT-5 averaged over 137 seconds per prompt! API access felt throttled compared to simply pasting the same prompt into the chat interface, which responded faster. <strong>GPT-4o was the fastest at 5 seconds per prompt and performed worst!</strong> Gemini 2.5 models took under 45 seconds on average, with 1.5 Pro taking an average of 11 seconds.</p></li></ol><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!AzPu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!AzPu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png 424w, https://substackcdn.com/image/fetch/$s_!AzPu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png 848w, https://substackcdn.com/image/fetch/$s_!AzPu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png 1272w, https://substackcdn.com/image/fetch/$s_!AzPu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!AzPu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png" width="1456" height="874" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:874,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:442736,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/171746222?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!AzPu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png 424w, https://substackcdn.com/image/fetch/$s_!AzPu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png 848w, https://substackcdn.com/image/fetch/$s_!AzPu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png 1272w, https://substackcdn.com/image/fetch/$s_!AzPu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1cc92681-7900-4806-ac6e-4e0a38e3811b_2262x1358.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3></h3><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xjoP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xjoP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png 424w, https://substackcdn.com/image/fetch/$s_!xjoP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png 848w, https://substackcdn.com/image/fetch/$s_!xjoP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png 1272w, https://substackcdn.com/image/fetch/$s_!xjoP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xjoP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png" width="1456" height="814" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:814,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:442301,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/171746222?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!xjoP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png 424w, https://substackcdn.com/image/fetch/$s_!xjoP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png 848w, https://substackcdn.com/image/fetch/$s_!xjoP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png 1272w, https://substackcdn.com/image/fetch/$s_!xjoP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F799f3a92-6b4b-451a-b1ff-3a7321978b79_2378x1330.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!09Hx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!09Hx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png 424w, https://substackcdn.com/image/fetch/$s_!09Hx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png 848w, https://substackcdn.com/image/fetch/$s_!09Hx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png 1272w, https://substackcdn.com/image/fetch/$s_!09Hx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!09Hx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png" width="1456" height="904" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:904,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:328963,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/171746222?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!09Hx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png 424w, https://substackcdn.com/image/fetch/$s_!09Hx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png 848w, https://substackcdn.com/image/fetch/$s_!09Hx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png 1272w, https://substackcdn.com/image/fetch/$s_!09Hx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab72be5b-4614-4152-8236-c70e6cb6d4de_2178x1352.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Iedn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Iedn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png 424w, https://substackcdn.com/image/fetch/$s_!Iedn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png 848w, https://substackcdn.com/image/fetch/$s_!Iedn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png 1272w, https://substackcdn.com/image/fetch/$s_!Iedn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Iedn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png" width="1456" height="796" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:796,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:442471,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/171746222?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Iedn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png 424w, https://substackcdn.com/image/fetch/$s_!Iedn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png 848w, https://substackcdn.com/image/fetch/$s_!Iedn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png 1272w, https://substackcdn.com/image/fetch/$s_!Iedn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ebc81c2-5486-43d2-8ec8-f52347068816_2390x1306.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SXbG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SXbG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png 424w, https://substackcdn.com/image/fetch/$s_!SXbG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png 848w, https://substackcdn.com/image/fetch/$s_!SXbG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png 1272w, https://substackcdn.com/image/fetch/$s_!SXbG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SXbG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png" width="1456" height="933" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:933,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:374883,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/171746222?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SXbG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png 424w, https://substackcdn.com/image/fetch/$s_!SXbG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png 848w, https://substackcdn.com/image/fetch/$s_!SXbG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png 1272w, https://substackcdn.com/image/fetch/$s_!SXbG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F608f9695-2c3e-4101-a27f-9fb33c33e894_2078x1332.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2><strong>Closing Thoughts</strong></h2><p>I started this experiment as a fun curiosity, vibe-coding my way through Cursor. <strong>Language models do seem to have a kind of mental map of animals</strong>. That map, however, is <strong>fuzzy, distorted, and full of amusing failures</strong>: diagonal rectangles for cats, alien-donkeys for dogs, and cartoonish polygons for bald eagles.</p><p>I did try providing more detailed instructions (e.g., &#8220;make sure the elephant has tusks and four legs&#8221;), but that didn&#8217;t really help much. Perhaps there exist better prompts that can help here!</p><p>What do you think - which of these sketches is your favorite? Any ideas worth pursuing in this general area?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Can LLMs Be Used for Financial Forecasting?]]></title><description><![CDATA[How We Used Fine-Tuning with Synthetic Reasoning Traces and RLVR/GRPO to Train a Financial Reasoning Model]]></description><link>https://www.modelanalysis.ai/p/can-llms-predict-a-stocks-performance</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/can-llms-predict-a-stocks-performance</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Fri, 20 Jun 2025 01:21:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bZRc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>&#171;note: This article provides a high-level overview of our project. For more details, <a href="https://drive.google.com/file/d/1G6xT4LZg8qfVDE9kmeSAS5-RrHqVh6ie/view?usp=sharing">here's the full project report</a>.&#187;</p><h3>1. Introduction</h3><p>As part of  the <a href="https://cs224r.stanford.edu/">CS224R </a>(Deep Reinforcement Learning) course that recently completed, I partnered with Jonathan Larkin and Tamika Bassman to work on a project where we dove headfirst into one of the most common yet difficult problems in finance: predicting stock performance. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>More specifically, we explored whether state-of-the-art deep reinforcement learning could help a large language model (LLM) reason more effectively about future stock performance. Central to our approach is the use of reinforcement learning with verifiable rewards (RLVR) following supervised finetuning (SFT). We ran a baseline on Qwen3-1.7B, and then successively improved this baseline by (a) fine-tuning the model on synthetic reasoning traces, and (b) RLVR training using Group Relative Policy Optimization (GRPO) algorithm with binary correctness and format correctness rewards. Our goal was to measure the performance of this approach against the base model, a traditional linear model, and the predictions of human analysts.</p><h3>2. The Architecture: A Pipeline for Financial Reasoning</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bZRc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bZRc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png 424w, https://substackcdn.com/image/fetch/$s_!bZRc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png 848w, https://substackcdn.com/image/fetch/$s_!bZRc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png 1272w, https://substackcdn.com/image/fetch/$s_!bZRc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bZRc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png" width="1439" height="630" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:630,&quot;width&quot;:1439,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:149967,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/166332320?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bZRc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png 424w, https://substackcdn.com/image/fetch/$s_!bZRc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png 848w, https://substackcdn.com/image/fetch/$s_!bZRc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png 1272w, https://substackcdn.com/image/fetch/$s_!bZRc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b81f593-5af2-433b-a076-83f3dbfaa756_1439x630.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We designed a multi-stage pipeline aimed at systematically building financial reasoning ability into a base LLM.</p><p><strong>A. Data Preparation</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zi9u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zi9u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png 424w, https://substackcdn.com/image/fetch/$s_!zi9u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png 848w, https://substackcdn.com/image/fetch/$s_!zi9u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png 1272w, https://substackcdn.com/image/fetch/$s_!zi9u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zi9u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png" width="1456" height="944" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:944,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:275080,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/166332320?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zi9u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png 424w, https://substackcdn.com/image/fetch/$s_!zi9u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png 848w, https://substackcdn.com/image/fetch/$s_!zi9u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png 1272w, https://substackcdn.com/image/fetch/$s_!zi9u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86880f7a-5be6-48be-a27b-5c66e296e23d_1460x947.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We started with a dataset of over 120,000 earnings call transcripts from 3,000 U.S. stocks, spanning 20 years. To create our target labels ("STRONG BUY" to "STRONG SELL"), we calculated each stock's 1-month future return and then percentile-ranked it against its sector peers. This crucial step helped isolate company-specific performance by filtering out broader market or sector-wide trends.</p><p><strong>B. Establishing the Baselines</strong></p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CEqj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CEqj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png 424w, https://substackcdn.com/image/fetch/$s_!CEqj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png 848w, https://substackcdn.com/image/fetch/$s_!CEqj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png 1272w, https://substackcdn.com/image/fetch/$s_!CEqj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CEqj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png" width="518" height="166.74082840236687" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:272,&quot;width&quot;:845,&quot;resizeWidth&quot;:518,&quot;bytes&quot;:49535,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/166332320?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CEqj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png 424w, https://substackcdn.com/image/fetch/$s_!CEqj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png 848w, https://substackcdn.com/image/fetch/$s_!CEqj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png 1272w, https://substackcdn.com/image/fetch/$s_!CEqj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc88d3625-9e7e-4981-9bb9-b9efcbfc472b_845x272.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>We created three key baselines (higher F1 scores are better):</p><ul><li><p>The performance of a linear TF-IDF model to see how much predictive power was in the text itself. </p></li><li><p>The performance of human analyst consensus ratings.</p></li><li><p>The performance of the base LLM - the pre-trained Qwen3 1.7B</p></li></ul><p><strong>C. The Training Method</strong></p><p>Our approach wasn't just to throw the data at the model. We used a sequence of fine-tuning and reinforcement learning stages.</p><ul><li><p><strong>Supervised Fine-Tuning (SFT):</strong> To train our model to think like a financial analyst, we used a &#8220;frontier&#8221; model, Gemini Pro 2.5, to create synthetic &#8220;reasoning traces&#8221; &#8212; step-by-step explanations that show <em>why</em> a prediction makes sense, not just <em>what</em> it is. We generated these traces using two methods: rejection sampling (keeping only the traces that led to the correct answer) and a technique inspired by "Self-Taught Reasoner" (STaR), where we gave the model the correct answer and asked it to reason backward. We then fine-tuned our base Qwen3 model on these reasoning traces along with the answers.</p></li><li><p><strong>Reinforcement Learning with Verifiable Rewards (RLVR):</strong> This was the final and most critical stage. We used Group Relative Policy Optimization (GRPO), an algorithm designed for enhanced reasoning tasks. We rewarded the model with a simple binary score (1.0 for a correct label, 0 for incorrect) and a smaller reward for getting the output format right. </p></li></ul><p>For computational efficiency, we used Low-Rank Adaptation (LoRA) for all training stages and vLLM to speed up inference.</p><div><hr></div><h3>3. The Findings</h3><p>So, how did we do? The results were both encouraging and humbling.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0cPU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0cPU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png 424w, https://substackcdn.com/image/fetch/$s_!0cPU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png 848w, https://substackcdn.com/image/fetch/$s_!0cPU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png 1272w, https://substackcdn.com/image/fetch/$s_!0cPU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0cPU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png" width="434" height="269.25061425061426" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:505,&quot;width&quot;:814,&quot;resizeWidth&quot;:434,&quot;bytes&quot;:116925,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/166332320?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0cPU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png 424w, https://substackcdn.com/image/fetch/$s_!0cPU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png 848w, https://substackcdn.com/image/fetch/$s_!0cPU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png 1272w, https://substackcdn.com/image/fetch/$s_!0cPU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ffc1a42-2b53-4c8d-b0fb-1f917a9d6cd9_814x505.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The most sobering finding was that our <strong>simple linear TF-IDF baseline, trained on the full 90,000-sample training set, outperformed every other model</strong> with a macro F1 score of 0.2597. This result provided a valuable insight: even without sophisticated modeling, the information in earnings calls transcripts holds real predictive power.</p><p>However, the story of our LLM pipeline is one of consistent, measurable improvement:</p><ul><li><p>The <strong>base Qwen3 1.7B model</strong> performed poorly on its own (F1 score: 0.0897) and exhibited a strong "optimism bias," as indicated in the confusion matrix (a) of Figure 4 below. </p></li><li><p><strong>Human analyst consensus</strong> ratings, when mapped to our labels, scored an F1 of 0.1743&#8212;better than the raw LLM, but still below the TF-IDF model.</p></li><li><p>Our <strong>LLM pipeline, even when trained on a tiny fraction of the data, surpassed the human benchmark</strong>. The final model, which went through both SFT on traces and RLVR optimization, achieved an F1 score of 0.2084.</p><p></p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Brkb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Brkb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png 424w, https://substackcdn.com/image/fetch/$s_!Brkb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png 848w, https://substackcdn.com/image/fetch/$s_!Brkb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png 1272w, https://substackcdn.com/image/fetch/$s_!Brkb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Brkb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png" width="1456" height="904" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:904,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:238989,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/166332320?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Brkb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png 424w, https://substackcdn.com/image/fetch/$s_!Brkb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png 848w, https://substackcdn.com/image/fetch/$s_!Brkb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png 1272w, https://substackcdn.com/image/fetch/$s_!Brkb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f488a63-cce8-4c2a-aa5a-61f298c6cb20_1492x926.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>While we didn't beat the TF-IDF baseline, we successfully demonstrated that SFT on synthetic data and then refinement with RLVR improved the model's performance over the base LLM. We were held back by compute constraints; our entire pipeline took about 20 hours to run on a single NVIDIA H100, and we were only able to fine-tune on 414 synthetic examples and train the RLVR stage on 1,000 samples. Given that the full training set has 90,000 samples, there's a clear opportunity to scale.</p><h3>4. What More Could We Have Done?</h3><p>Every research project leaves unexplored avenues, and ours is no exception.</p><ul><li><p><strong>Integrating Tools and External Data Sources:</strong> A human analyst doesn't just read transcripts; they use calculators, refer to financial data from other sources. A major next step would be to give the model "tools" to allow it to perform quantitative calculations and access external data, making its analysis much more robust.</p></li><li><p><strong>Richer Reward Mechanisms:</strong> We used a simple binary reward for the final answer. Future work could employ an "LLM-as-a-judge" to provide more nuanced, process-based rewards that evaluate the quality of the reasoning itself, not just the outcome.</p></li><li><p><strong>More Rigorous Ablation Studies:</strong> We acknowledge the need for more rigorous ablation studies to precisely measure the impact of each component of our pipeline including, perhaps, tweaking the prompt to indicate to the model to be more skeptical when judging the transcripts. We also need to look for and eliminate data leakage, as our 20-year dataset may contain information the base LLM saw in its original pre-training.</p></li><li><p><strong>Scaling:</strong> Another obvious path forward is to apply this pipeline to larger foundation models and train it on our full dataset.</p></li></ul><h3>5. Lessons from the Project</h3><p>Beyond the technical findings, a project like this teaches one a lot about the practical realities of AI research.</p><p>First, <strong>planning is everything.</strong> A multi-stage pipeline with dependencies on data processing, baseline generation, supervised fine-tuning, and reinforcement learning requires a clear and structured plan from the outset. </p><p>Second, <strong>expect</strong> <strong>fast evolving frameworks/libraries and incomplete or missing documentation</strong>. Frontier reasoning models are still very new and getting the prompt formatting right and tuning the hyperparameters can be quite tricky. Documentation on the libraries and different hyperparameters them is still very sparse - one should be prepared to adjust one&#8217;s plans as more information is learned during implementation.</p><p>Finally, the project reinforced the sheer <strong>time-consuming nature of model training and fine-tuning.</strong> A single training and evaluation run took nearly a full day. One experiment we attempted had an estimated run time of 75 hours. This isn't a field for the impatient or the undisciplined; one needs to design experiments carefully because compute time is a precious resource.</p><p>Despite the challenges, we've established a promising foundation. Our findings show that LLMs, when equipped with structured reasoning from reinforcement learning, hold significant potential in financial forecasting. With more compute and refined techniques, this approach could become a powerful complement to traditional financial analysis.</p><div><hr></div><h3>6. Appendix</h3><h4>A. Example Prompt:</h4><pre><code>&lt;|im_start|&gt;user 
You are an expert institutional equity analyst. 

Given the following text, predict the stock&#8217;s relative performance to stocks in the same sector over the next month. 

You may rate a stock STRONG BUY, BUY, HOLD, SELL, or STRONG SELL. 

Think carefully. You must end your output with one of: STRONG BUY, BUY, SELL, STRONG SELL, or HOLD. 

Input: 
"""Operator: Ladies and gentlemen, thank you for standing by and welcome to the Third Quarter Oil-Dri Corporation of America Earnings Conference Call. At this time, all participants are in a listen-only mode. After the speakers presentation, there will be a question-and-answer session. [Operator Instructions] I would now like to hand the call over to Dan Jaffee, President and Chief Executive Officer. Please go ahead.  

Dan Jaffee: Thank you. Welcome everybody to the Oil-Dri third quarter investor
teleconference joining me both physically and virtually<strong>...</strong>

<strong>&lt;removed rest of the transcript of nearly six pages for brevity&gt;</strong>

<strong>...
</strong>Dan Jaffee: Great. And I hope you also spotted the share repurchase which again was a way of trying to deliver value back to our shareholders and we like we saw we are opportunistic so that we had a chance to buyback shares with higher dividend and our cash is earning 0.2% so and we have plenty of opportunity that we can deploy that cash on but we are going to still stay opportunistic on the share repurchase program so hopefully spotted that in the queue as well thank you guys its been a half an hour and we look forward to talking to you it will be our end of the year that will be our fourth quarter and fiscal year end I would tell you on one of our major metrics which is we look at pre tax pre bonus income and then we divvy up that income between the shareholders and then the teammates. We already, through nine months have made more than we have ever made and any fiscal year we have ever had so as I joke we are playing with a house of money at this point but we are having a record year it feels really good it feels really good that all the investments we have made in the last two and half years on people and infrastructure are why this is happening and its just it feels good it is predictable and we are up in front of our businesses and we just we have always taken a long term approach and the seeds we planted years ago are sprouting now and seeds we plant today will start in a few years so thank you for the long-time holders and we are happy to reward you with I think will results 17 year in a row of dividend increases which is fantastic. So thanks everybody. We will talk to you again in a quarter. Operator: Ladies and gentlemen, this concludes todays conference call. Thank you for participating. You may now disconnect. Everyone have a great day."""

Answer: 
&lt;|im_end|&gt; 
&lt;|im_start|&gt;assistant </code></pre><h4>B. Example Response with Thinking Trace:</h4><pre><code><code>&lt;think&gt;
Okay, let&#8217;s tackle this question. The user wants to predict the stock&#8217;s relative performance to stocks in the same sector over the next month, rating it as STRONG BUY, BUY, HOLD, SELL, or STRONG SELL. 

First, I need to understand the sector. The company mentioned is Oil-Dri Corporation of America, which is involved in cat litter, animal health products, and other related areas. The sector for this would be the consumer products or retail sector, specifically focusing on pet care and cat litter.

Looking at the earnings call, the company reported strong financial results, with a 11% increase in retail and wholesale product sales, 8% increase in net sales, and a 25% increase in animal health products. They also mentioned improvements in gross profit, lower freight and natural gas costs, and a strong income from operations. The CEO emphasized the team&#8217;s resilience during the pandemic and the company&#8217;s ability to meet demand. 

Now, the question is about relative performance compared to sector peers. The key factors here are the company&#8217;s financial performance, market trends, and sector-specific factors. The company is in the consumer goods sector, specifically pet care. 

Considering the sector&#8217;s general performance, the consumer products sector has been affected by the pandemic, with a shift towards e-commerce and online sales. However, the company&#8217;s performance in retail and wholesale products was strong, with a 11% increase. The animal health segment showed growth, particularly in the Americas, with a 25% increase in sales. 

The sector as a whole might be experiencing some challenges, like supply chain issues and market shifts, but the company&#8217;s strong financials and focus on value-driven products (like cat litter) could give them an edge. The CEO mentioned that the company is well-positioned to compete in the value segment, which is a key factor in the consumer products sector. 

However, there are risks. The company is in a sector that&#8217;s heavily influenced by consumer behavior and economic conditions. If the pandemic&#8217;s impact subsides, there might be competition from other brands, and the company&#8217;s ability to maintain pricing and quality could be tested. Also, the sector might face challenges with supply chain disruptions and the ongoing impact of the pandemic. 

Given the company&#8217;s strong financial results, improved gross margins, and focus on value, it&#8217;s possible that the stock could perform well relative to peers. However, the sector&#8217;s overall performance and market conditions could affect this. The user wants a relative performance prediction, so I need to consider if the stock is likely to outperform, underperform, or hold compared to others in the same sector. 

The answer should be based on the company&#8217;s strong performance, market position, and sector trends. Since the company is in a sector that&#8217;s adapting to digital trends and value strategies, and they have shown resilience, the relative performance might be positive. However, there&#8217;s a risk of sector-wide challenges, so the rating could be HOLD or BUY. But given the strong financial results and the company&#8217;s focus on value, a BUY might be appropriate. 
&lt;/think&gt;

BUY</code></code></pre><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Forget 'Vibe Coding': Why Karpathy's Talk Shows Fundamentals Still Matter]]></title><description><![CDATA[But the real story isn't just how we'll code, but what we're building towards. After decades of shifting to web and mobile, the next great business transformation is becoming agent-native.]]></description><link>https://www.modelanalysis.ai/p/forget-vibe-coding-why-karpathys</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/forget-vibe-coding-why-karpathys</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Fri, 20 Jun 2025 01:18:23 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3f601d2a-8649-49a3-b233-2aa9d3fc30f6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>What a <a href="https://www.youtube.com/watch?v=LCEmiRjPEtQ">great talk</a> by <strong>Andrej Karpathy</strong> - he's got to be one of the best communicators of this era of AI and LLMs! His talk is very high in information density - and there's something for everyone including the non-technical audience. Here are a few of my key takeaways and observations. Let me know what you think!&#8291;<br>&#8291;<br>&#120783;. &#119826;&#119848;&#119839;&#119853;&#119856;&#119834;&#119851;&#119838; &#120783;.&#120782;/&#120784;.&#120782;/&#120785;.&#120782;&#8291;<br>- Software 1.0 is the traditional programming paradigm that those of us in the industry are familiar with. Almost all the traditional software (e.g., operating systems, business applications, video games) that exists today is created in this paradigm. &#8291;<br>- Software 2.0 uses neural nets and lots of data to solve problems that are very hard or impossible to solve using Software 1.0 like autonomous driving, natural language processing and speech recognition. LLMs, of course, are the best known examples of Software 2.0.&#8291;<br>- Software 3.0 builds on top of LLMs (software 2.0) and writes code with natural language prompts from users.&#8291;<br>&#8291;<br>&#120784;. &#119827;&#119841;&#119838; &#119839;&#119854;&#119847;&#119837;&#119834;&#119846;&#119838;&#119847;&#119853;&#119834;&#119845; &#119853;&#119851;&#119834;&#119842;&#119853;&#119852; &#119848;&#119839; &#119834; &#119840;&#119851;&#119838;&#119834;&#119853; &#119837;&#119838;&#119855;&#119838;&#119845;&#119848;&#119849;&#119838;&#119851; &#119834;&#119851;&#119838;&#119847;'&#119853; &#119836;&#119841;&#119834;&#119847;&#119840;&#119842;&#119847;&#119840;.&#8291;<br>- Vibe coding gets all the memes but the maximum value will be generated by developers who are well versed in all the three paradigms (or at least in 1.0 and 3.0).&#8291;<br>- Developers that have a clear and accurate mental model of what is needed and effectively communicate that to the LLM will be more productive than those that do not.&#8291;<br>- Given the current limitations of LLMs (like context window length and unreliability), Developers need to know how to implement the changes in manageable chunks and how to debug and fix issues when things eventually go wrong.&#8291;<br>- Of course, the ability to design and architect software with all the x'ties (maintainability, supportability, scalability, etc.) remains as valuable as ever.&#8291;<br>&#8291;<br>&#120785;. &#119809;&#119854;&#119842;&#119845;&#119837;&#119842;&#119847;&#119840; &#119853;&#119848;&#119848;&#119845;&#119852; &#119839;&#119848;&#119851; &#119853;&#119841;&#119838; "&#119843;&#119834;&#119840;&#119840;&#119838;&#119837; &#119842;&#119847;&#119853;&#119838;&#119845;&#119845;&#119842;&#119840;&#119838;&#119847;&#119836;&#119838;" &#119848;&#119839; &#119819;&#119819;&#119820;&#119852;&#8291;<br>- AGI and ASI are getting all the attention with the promise and threat they pose to the world as we know it&#8291;<br>- But significant effort and innovation will be going into developing tools and applications with UI/UX capabilities that allow users to work with the "jagged intelligence" of LLMs with embedded verifiers and fine grained control on AI autonomy. &#8291;<br>&#8291;<br>&#120786;. &#119809;&#119838;&#119836;&#119848;&#119846;&#119842;&#119847;&#119840; &#119808;&#119840;&#119838;&#119847;&#119853;-&#119821;&#119834;&#119853;&#119842;&#119855;&#119838; &#119842;&#119852; &#119853;&#119841;&#119838; &#119847;&#119838;&#119857;&#119853; &#119840;&#119851;&#119838;&#119834;&#119853; &#119835;&#119854;&#119852;&#119842;&#119847;&#119838;&#119852;&#119852; &#119853;&#119851;&#119834;&#119847;&#119852;&#119839;&#119848;&#119851;&#119846;&#119834;&#119853;&#119842;&#119848;&#119847;&#8291;<br>- It took businesses over a decade to build a mature web presence, and another decade to go mobile-native.&#8291;<br>- I think the next major transformation for businesses is to become agent-native: both for exposing their services to external agents and for embedding agentic solutions in their internal processes.&#8291;<br>- IMO, this transformation is a marathon that will likely take the next decade to fully realize.&#8291;</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Evolution of Meta's Llama LLMs]]></title><description><![CDATA[Charting the Architectural Progression of its Open Weight Models]]></description><link>https://www.modelanalysis.ai/p/the-evolution-of-metas-llama-llms</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/the-evolution-of-metas-llama-llms</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Thu, 08 May 2025 02:31:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Jgeu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#171;All of this information presented in this article already out there, but seeing it all together might allow one to easily appreciate the evolution of the Llama family of LLMs.&#187;</em></p><p>Meta's Llama family of large language models has rapidly evolved from a research project into a major force in the AI landscape with its open weights license. While it&#8217;s no longer the open weight LLM leader (that spot is firmly in the hands of DeepSeek and Qwen&#8217;s models), there&#8217;s little doubt that Meta is working hard to reclaim its spot at the front.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Using key data points presented in three tables, this post provides a quick overview of how Llama&#8217;s architecture evolved over the past three years, focusing on its release timeline, the scaling and choices of its technical architecture, and how its choices stack up against its major competitors.</p><div><hr></div><h3>Table 1:  Release Timeline</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Jgeu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Jgeu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png 424w, https://substackcdn.com/image/fetch/$s_!Jgeu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png 848w, https://substackcdn.com/image/fetch/$s_!Jgeu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png 1272w, https://substackcdn.com/image/fetch/$s_!Jgeu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Jgeu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png" width="1456" height="877" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:877,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:230556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/161817410?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Jgeu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png 424w, https://substackcdn.com/image/fetch/$s_!Jgeu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png 848w, https://substackcdn.com/image/fetch/$s_!Jgeu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png 1272w, https://substackcdn.com/image/fetch/$s_!Jgeu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b3c6d62-fef6-4e97-9504-f4f3871ba3df_1650x994.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Highlights from Table 1</h4><p><strong>Rapid Iteration and Increasing Specialization</strong>: The Llama 3 generation saw accelerated point releases (3.1, 3.2, 3.3) within months in 2024, indicating a shift towards faster, more diverse model rollouts supporting coding, vision, edge deployments etc.</p><p><strong>Shift to Openness with Open Weights</strong>: Llama 1 started as a restricted research release, but Llama 2 marked a pivot towards broader commercial use, a strategy continued with subsequent versions, albeit with some restrictive licensing conditions.</p><p><strong>Architectural Leaps</strong>: Llama 4 introduced fundamental changes like Mixture of Experts (MoE) and native multimodality, distinguishing it significantly from the Llama 1-3 generations.</p><div><hr></div><h3>Table 2: Detailed Technical Specifications Comparison</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IFze!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IFze!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png 424w, https://substackcdn.com/image/fetch/$s_!IFze!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png 848w, https://substackcdn.com/image/fetch/$s_!IFze!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png 1272w, https://substackcdn.com/image/fetch/$s_!IFze!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IFze!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png" width="1456" height="965" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:965,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:279513,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/161817410?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IFze!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png 424w, https://substackcdn.com/image/fetch/$s_!IFze!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png 848w, https://substackcdn.com/image/fetch/$s_!IFze!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png 1272w, https://substackcdn.com/image/fetch/$s_!IFze!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77db4fed-1bd5-4144-9ae7-fe67d31b9440_1756x1164.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>*Note: NS = Not Specified in provided sources. </p><h4>Highlights from Table 2</h4><ul><li><p><strong>Exponential Growth in Context Length</strong>: Context window size saw dramatic increases, from 2K in Llama 1 to 128K in Llama 3.1, and finally increasing to 1M/10M in Llama 4. </p></li><li><p><strong>Massive Training Data Scaling</strong>: The volume of training data grew exponentially as well, from ~1.4T tokens for Llama 1 to 15T+ for Llama 3, and up to ~40T for Llama 4 Scout.</p></li><li><p><strong>Architectural Refinements</strong>: Key changes include the adoption of RMSNorm and SwiGLU from Llama 1 onwards; the introduction of Grouped Query Attention (GQA) in Llama 2 (70B) and its standardization across Llama 3 models; and the shift to a much larger 128K+ vocabulary tokenizer from Llama 3 onwards.</p></li><li><p><strong>Shift to Mixture of Experts (MoE)</strong>: Llama 4 introduced MoE, decoupling active parameters (used per token, e.g., 17B) from total parameters (overall knowledge, e.g., 109B), aiming for greater efficiency at scale.</p></li><li><p> <strong>Multimodality Integration</strong>: While Llama 3.2 added vision capabilities, Llama 4 implemented native multimodality from the ground up.</p></li></ul><div><hr></div><h3>Table 3: Comparative Technical Specifications (as of April 2025)</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!w_IN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!w_IN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png 424w, https://substackcdn.com/image/fetch/$s_!w_IN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png 848w, https://substackcdn.com/image/fetch/$s_!w_IN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png 1272w, https://substackcdn.com/image/fetch/$s_!w_IN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!w_IN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png" width="1456" height="810" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fb05a677-a26d-4311-9986-ec047113685e_1600x890.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:810,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:247244,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/161817410?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!w_IN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png 424w, https://substackcdn.com/image/fetch/$s_!w_IN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png 848w, https://substackcdn.com/image/fetch/$s_!w_IN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png 1272w, https://substackcdn.com/image/fetch/$s_!w_IN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb05a677-a26d-4311-9986-ec047113685e_1600x890.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Highlights from Table 3</h4><p><strong>Architectural Diversity</strong>: Llama 4 and DeepSeek V3 utilize MoE architectures, while Qwen 2.5 and Mistral Large 2/Nemo stick to dense designs, showcasing different strategies for scaling. MoE models (Llama 4, DeepSeek V3) have lower active parameters compared to dense models (Qwen2.5, Mistral Large 2) of similar capability, aiming for inference, but differ in their active/total parameter ratios.</p><p><strong>Context Length</strong>: Llama 4 Scout (10M) and Maverick (1M) offer significantly larger context windows than the 128K standard supported by Qwen 2.5, DeepSeek V3, and Mistral Large 2/Nemo.</p><p><strong>Tokenizer Variations</strong>: Vocabulary sizes and tokenizer types vary, with Llama 4, DeepSeek V3, and Mistral Nemo using larger vocabularies (~128K-131K) compared to Mistral Large 2 (32K). Qwen 2.5 has the largest vocab (~151K).</p><p><strong>Licensing Fragmentation</strong>: Licensing terms differ significantly, from the permissive Apache 2.0 (Mistral Nemo) and MIT/Commercial (DeepSeek V3) to custom licenses with MAU restrictions (Llama 4, Qwen 2.5) and research-focused licenses (Mistral Large 2 weights).</p><p><em>&#171;That&#8217;s all folks! Hope you got something out of this one :)&#187;</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The End of the App Store?]]></title><description><![CDATA[Why You Might Only Need One AI Super App]]></description><link>https://www.modelanalysis.ai/p/the-end-of-the-app-store</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/the-end-of-the-app-store</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Fri, 11 Apr 2025 22:50:12 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/688324a3-dc0e-4aed-b9bf-e8b75e735620_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>&#171;Series Note: This article is part of a series of articles I&#8217;m writing on MCP. See my blog modelanalysis.ai for the complete collection.&#187;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KVjE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KVjE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!KVjE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!KVjE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!KVjE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KVjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1712996,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/159798312?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KVjE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!KVjE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!KVjE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!KVjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff00c8695-633f-46b1-a73f-31a44b93dccb_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h1><strong>Introduction</strong></h1><p>Will our future be dominated by a single AI-driven &#8220;super app&#8221; that handles everything&#8212;from managing finances to entertainment to lifelong learning? It&#8217;s starting to look that way. In this article, I propose a scenario where our computing experience coalesces into one adaptive, AI-powered interface: a true all-in-one super app. I&#8217;ll explore how this shift might unfold and what such a transformative future could look like.</p><h1><strong>The Vision: From Many Apps to One</strong></h1><p>Today, we juggle multiple apps to achieve what we are looking to do&#8212;a few for shopping, another few for entertainment, yet another few for finances&#8212;all with unique logins, interfaces, and data silos. This fragmentation has become the norm, but it&#8217;s also a major source of friction.</p><p>But what if we could change the script? What if, instead of launching a dozen different apps, you could simply talk to one intelligent interface &#8211; an AI-powered "Super App" capable of handling almost any digital task?</p><p>For everyday users, this means unprecedented convenience and empowerment: the technology adapts to you, not the other way around. For service providers, it means rethinking how to offer services: instead of standalone apps, you might create modules that &#8220;plug into&#8221; a bigger AI ecosystem, with standard protocols like MCP ensuring compatibility.</p><p>The AI dynamically understands your natural language requests and invokes the necessary services behind the scenes. It's less about bundling existing apps and more about transcending the app paradigm altogether, moving towards a goal-oriented interaction model. The focus shifts from <em>which app to open</em> to <em>what you want to achieve</em>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1><strong>Peeking Inside the AI Super App</strong></h1><p>So, what might interaction with this AI Super App actually feel like? The key is blending conversational interactions with dynamic user interfaces that adapt to the task at hand.</p><p>Here&#8217;s how I envision the app interface:</p><ul><li><p><strong>The Conversation Stream:</strong> This is your primary form of input where you can talk or type. In the future, the Super App might even interpret visual cues like body language via video.</p></li><li><p><strong>The Adaptive Workspace:</strong> This central area morphs to fit your task. Drafting an email? You&#8217;ll see rich text editing tools. Planning a trip? Expect maps, calendars, and flight options. Working on finances? Interactive charts and spreadsheets. Early versions of this idea already exist in tools like Claude Desktop and ChatGPT&#8217;s Canvas&#8212;but we&#8217;re just getting started.</p></li><li><p><strong>Persistent, Accessible Memory:</strong> The AI learns your preferences (e.g., airlines, dietary needs, meeting times) and remembers relevant experiences (e.g., photos, past events, trips). This data&#8212;securely stored using a mix of local and cloud solutions&#8212;makes the AI feel like your personal assistant.</p></li><li><p><strong>Privacy and Safety By Design:</strong> With so much data centralized, privacy can&#8217;t be an afterthought. The Super App should share only the minimum necessary information with external services, acting as a secure gatekeeper rather than a leak point.</p></li></ul><h1><strong>MCP: A Key Enabler of the Super App Vision</strong></h1><p>So, how does an AI Super App seamlessly connect to and orchestrate potentially hundreds or even thousands of different tools and services? This is where the Model Context Protocol (MCP) comes in &#8211; paving the way for an ecosystem that is both extensible and secure:</p><ul><li><p><strong>Unified Access to Diverse Tools:</strong> MCP simplifies development by allowing a single uniform interface tools and services. Developers can more easily "wrap" their existing service or tool in an MCP server, making it instantly accessible to any AI application that speaks MCP.</p></li><li><p><strong>Agentic Workflows with Guardrails:</strong> MCP is designed with agentic AI in mind &#8211; AIs that can take actions autonomously while allowing for guardrails across the ecosystem. The user, typically through the host application, controls which tools and services the AI can access. The tools and services themselves can enforce permissions, checking if the user (or the AI acting on their behalf) is authorized to perform the requested action.</p></li><li><p><strong>Future-Proofing for AI Model Evolution:</strong> MCP decouples service integration from the specific AI model. You can upgrade the &#8220;brain&#8221; of your Super App without having to rewire every tool.</p></li></ul><h1><strong>Beyond MCP: A Road Map for What&#8217;s Next</strong></h1><p>While MCP represents a major step forward, for AI agents to become more sophisticated and autonomous in increasingly more complex domains, the demands on the underlying protocols will inevitably grow. I foresee several key areas where MCP (or its successors) will need to evolve:</p><ul><li><p><strong>Standardized Discovery &amp; Context:</strong> We need something akin to a universal directory listing and discovery protocol to enable AI Super Apps to reliably find out the tools and services.</p></li><li><p><strong>More powerful and dynamic UI:</strong> We will need to figure out how to create powerful yet simple to use interactive task specific UI components within one workspace provided by the Super App. Today&#8217;s tools such as Claude Desktop and ChatGPT are early steps in that direction.</p></li><li><p><strong>Multi-AI/Agent Collaboration:</strong> Complex tasks may require specialized AIs working together. We&#8217;ll need cross-agent protocols for context-sharing, task handoffs, and coordinated actions&#8212;perhaps through MCP Gateways. (Just as I write this, Google has released the open-source Agent Development Kit.)</p></li><li><p><strong>Scalability &amp; Multi-Tenancy:</strong> For MCP to thrive in SaaS environments, the protocol and supporting infrastructure will need robust multi-tenancy support.</p></li><li><p><strong>Advanced Security, Authentication &amp; Audit:</strong> We need more features such as standardized authentication methods between clients and servers, finer-grained permissions (especially for regulated industries like finance and healthcare), and robust, protocol-level audit trails.</p></li></ul><h1><strong>Reality Check: Navigating the Hurdles Ahead</strong></h1><p>As compelling as this vision of an all-powerful AI Super App feels like, there are significant hurdles that still stand in the way. Here are some of the key challenges I see:</p><ul><li><p><strong>LLM Reliability:</strong> For a Super App to manage critical tasks requires improvements in model fact-checking, grounding, and potentially curated knowledge bases, alongside robust error handling and human oversight mechanisms.</p></li><li><p><strong>The Fragmented Tool Ecosystem:</strong> Standardization is essential&#8212;but hard. Many companies maybe reluctant to open up or migrate legacy systems.</p></li><li><p><strong>Sophisticated Inter-App Interactions:</strong> Supporting rich, interactive capabilities within a standard API is a major architectural challenge - think about collaborative document editing, intricate image manipulation, or managing complex project workflows.</p></li><li><p><strong>User Trust: </strong>Asking users to trust one AI with everything is a big ask. Transparency and control are crucial to overcoming the &#8220;creepiness&#8221; factor.</p></li><li><p><strong>Navigating the Regulatory Maze:</strong> Global Super Apps must navigate data protection laws like GDPR and tackle hard questions about liability when things go wrong.</p></li></ul><h1><strong>So, Who Wins and Loses in This Scenario?</strong></h1><p>The emergence of AI-orchestrated Super Apps will rewire the economics of the entire software and services industry. The current app economy thrives on standalone applications vying for our attention, screen time, subscription dollars, and data. What happens when a single, intelligent interface becomes the primary gateway through which users access most, if not all, digital services?</p><p>I anticipate several major transformations in business models:</p><ul><li><p><strong>New Revenue Sharing and Monetization Models:</strong> We will see new revenue-sharing models emerge, analogous to current app store commissions with revenue distributed downstream based on the specific value add. Businesses might pay Super App platforms to have their services prioritized or featured.</p></li><li><p><strong>Market for Specialized AI Plugins &amp; Agents:</strong> Expert agents with domain knowledge in fields like medicine, law, or engineering could emerge&#8212;selling their expertise as plug-ins the Super App can call.</p></li><li><p><strong>Intensified Platform Competition and Market Concentration:</strong> Whoever owns the dominant Super App platform could gain enormous market power&#8212;far beyond what today&#8217;s app stores, search engines, or even OS vendors have. Expect a turf war between Big Tech (Google, Microsoft, Apple, etc.) and upstarts like OpenAI or Anthropic, along with inevitable antitrust debates.</p></li></ul><h1><strong>But what about the rest of us, the mainstream users?</strong></h1><p>For users, the biggest danger may be <strong>&#8220;learned helplessness.&#8221;</strong> If the AI handles everything, will we forget how to do things ourselves? A well-designed Super App, therefore, should aim to empower, not deskill. That means:</p><ul><li><p><strong>Transparency and Explainability:</strong> The AI should surface key decision points to the user rather than making all choices silently. When asked, it should be able to explain <em>why</em> it chose a particular tool, approach, or piece of information.</p></li><li><p><strong>User Control Points:</strong> Users must always have the ability to "grab the controls" and direct the process. This means being able to see which tools or APIs the AI is using, modify the plan, or switch to a fully manual mode if desired.</p></li><li><p><strong>Adaptive Interfaces:</strong> The system should cater to different user skill levels. Beginners get guidance; experts get control. The system should scale with user confidence and needs.</p></li></ul><h1><strong>Conclusion</strong></h1><p>The AI-powered Super App isn't just a futuristic fantasy. Increasingly powerful and agentic AI models are demonstrating capabilities that seemed like science fiction only a few years ago. We're witnessing a fundamental paradigm shift towards a human computer interaction model where we state our goals and an AI figure out how to achieve them.</p><p>That said, the AI Super App is but one facet of the impact AI will have on our society. This future sounds exhilarating but also scary given the profound consequences it will have on every aspect of how individuals, modern economies, and societies function.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Llama 4 Is a Context Monster—Just Don’t Ask It to Think (Yet)]]></title><description><![CDATA[Larger context, native support for multimodality, but no reasoning versions yet]]></description><link>https://www.modelanalysis.ai/p/llama-4-is-a-context-monsterjust</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/llama-4-is-a-context-monsterjust</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Sun, 06 Apr 2025 14:41:27 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6fddb90b-b640-4478-9b27-0ded564d5022_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TbaK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TbaK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!TbaK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!TbaK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!TbaK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TbaK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3555936,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/160707245?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TbaK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!TbaK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!TbaK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!TbaK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5b85194-24d6-4fe6-a77b-931699468fd7_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>tl;dr</strong> Meta's Llama 4 represents an evolutionary (not revolutionary) step forward in LLM development. Its move to a mixture of experts architecture and more native support for multimodal capabilities follow a well-trodden path. A context size ranging from 1M to 10M is probably the item that will grab the most attention, along with the fact that a reasoning model hasn't been released yet. The observed diminishing returns from scaling pre-training continues to highlight the need for new approaches to achieve substantial performance gains.</p><p>Here are some early observations for what it means for users:</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><strong>1. Expanded Context Windows</strong></p><p>Llama 4 models boast context windows ranging from 1 million to 10 million tokens, an 8x to 80x increase over previous models. This expansion, combined with its Needle in Haystack performance, allows entire codebases or large sets of documents to be processed in a single pass, reducing the need for complex retrieval-augmented generation (RAG) systems which have to deal with the chunking problem.</p><p><strong>2. Multimodal Capabilities</strong></p><p>The models support multimodal processing more natively with an early-fusion approach, enabling the integration of text, images, and audio. This positions Llama 4 competitively against other models, enabling applications that require simultaneous processing of diverse  formats, though actual performance still remains to be seen.</p><p><strong>3. Diminishing Returns in Base Model Scaling</strong></p><p>Llama 4 confirms the trend already seen with other SOTA models -  achieving substantial performance gains solely through scaling data and compute resources doesn&#8217;t seem possible anymore. Despite massive increase in parameters and training data, it demonstrated only modest improvements over its predecessors, confirming the need to find innovations beyond traditional scaling methods (RL-based test time training seems to be the one almost everyone&#8217;s betting on). Interestingly this generation&#8217;s largest model, Behemoth, which hasn&#8217;t yet been released for general use yet, continues to be trained and forms the basis for the smaller Maverick and Scout models. </p><p><strong>4. Absence of a Dedicated Reasoning Model</strong></p><p>While Llama 4 introduces newer/better models like Scout and Maverick, it does not include a dedicated reasoning model. But, I would expect Meta to release a reasoning model (just like everyone else!) later this year. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[MCP Demystified: Bridging a Gap in Official Documentation]]></title><description><![CDATA[A simple and clear insight into how Claude Desktop actually uses MCP &#8212; with diagrams and an explanation you won&#8217;t find in the official docs.]]></description><link>https://www.modelanalysis.ai/p/mcp-demystified-bridging-a-gap-in</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/mcp-demystified-bridging-a-gap-in</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Fri, 04 Apr 2025 18:39:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!H_Rl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#171;<strong>Terminology Note</strong>: Throughout this analysis, when I use the term "LLM" (Large Language Model), I&#8217;m referring to the class of foundation models that include not only text generation capabilities but also reasoning, planning, and multimodal capabilities (processing and generating images, audio, video, etc.). This expanded definition reflects the ongoing evolution of these models beyond pure language processing.</em></p><p><em><strong>Series Note</strong>: This article is part of a series of articles I&#8217;m writing on MCP. See my blog for the complete collection.&#187;</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Anthropic&#8217;s Model Context Protocol (MCP) has gained significant traction over the past few months, but the official documentation provided on its <strong>modelcontextprotocol.io</strong>  omits showing the role of LLM or conflates it with that of Claude Desktop (along with some other inconsistent terminology). This ambiguity can cause confusion, especially for new or casual readers trying to understand how MCP actually works &#8212; and can lead developers and architects to waste time muddling through with a hazy or incorrect mental model of MCP-based applications.</p><p>My goal in this article is to walk you through how things work under the hood, so that by the end, you&#8217;ll have a clearer and more accurate understanding of how MCP-based apps like Claude Desktop are designed. </p><p>(And if I&#8217;ve missed or misrepresented anything, I&#8217;d love to hear &#8212; I&#8217;ll keep this updated as things evolve.)</p><h1>A Quick Overview of Model Context Protocol</h1><p>Developed by Anthropic, MCP is an open protocol that is built from the ground up to help develop agentic applications powered by LLMs.  Using traditional, well known techniques and best practices of software architecture, MCP enables seamless integration between LLM applications and external data sources and tools.</p><p>There are four major components of MCP:</p><ol><li><p><strong>MCP servers</strong> provide access to data sources (e.g., portfolio data, market data), tools (e.g., Filesystem, Google Maps, weather), and prompts (e.g., reusable prompt templates and workflows). </p></li><li><p><strong>MCP clients</strong> reside within an MCP host process and connects to an MCP server.</p></li><li><p><strong>MCP hosts</strong> run one or more MCP clients that connect with the MCP servers.  MCP hosts act as both the technical orchestrators and the centralized dispatch system that enables LLMs to seamlessly interact with external data sources and tools while maintaining appropriate security boundaries.</p></li><li><p><strong>LLMs</strong> (Large Language Models), provide the core &#8220;intelligence&#8221; for the system.</p><p>They receive prompts from the MCP hosts and generate context-aware responses.</p></li></ol><h1>An Illustrative Example of Claude Desktop Application</h1><p>To illustrate how a typical MCP application can work, I&#8217;ll show below how Claude Desktop utilizes these MCP components to provide its functions. Claude Desktop is the application that sits on user&#8217;s computer and allows them to chat with one of Anthropic&#8217;s LLMs. It maintains their chat session context, chat history, and integrates with MCP based tools to enable the LLM&#8217;s agentic capabilities.</p><h2><strong>Initialization</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H_Rl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H_Rl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png 424w, https://substackcdn.com/image/fetch/$s_!H_Rl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png 848w, https://substackcdn.com/image/fetch/$s_!H_Rl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png 1272w, https://substackcdn.com/image/fetch/$s_!H_Rl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H_Rl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7a609016-327e-4378-979d-726217adfb11_2304x1256.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:237210,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/159688779?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H_Rl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png 424w, https://substackcdn.com/image/fetch/$s_!H_Rl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png 848w, https://substackcdn.com/image/fetch/$s_!H_Rl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png 1272w, https://substackcdn.com/image/fetch/$s_!H_Rl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a609016-327e-4378-979d-726217adfb11_2304x1256.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ol><li><p>User launches Claude Desktop application (the MCP Host Process). </p></li><li><p>Claude Desktop loads its configuration files either from local storage or from some remote config which include details about each MCP server (e.g. endpoints, authentication tokens, usage parameters, etc.</p></li><li><p>For every MCP Server in the config (Google Maps, Airbnb, Brave Search, Filesystem, or any other registered tool), Claude Desktop</p><ol><li><p>loads the server&#8217;s metadata (e.g. what capabilities it offers, how to call it, what parameters are needed, etc.).</p></li><li><p>instantiates a small &#8220;MCP Client&#8221; inside the Claude Desktop process for each server.</p></li></ol></li><li><p>Claude Desktop then sends the information about each tool to one of Anthropic&#8217;s LLM, say Claude 3.7 Sonnet. The LLM is essentially told, &#8220;Here are the available MCP servers (tools). Here&#8217;s how you can call them, which parameters they expect, and what they return.&#8221;</p></li></ol><h2>User Interaction with Claude Desktop</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0cYn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0cYn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png 424w, https://substackcdn.com/image/fetch/$s_!0cYn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png 848w, https://substackcdn.com/image/fetch/$s_!0cYn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png 1272w, https://substackcdn.com/image/fetch/$s_!0cYn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0cYn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png" width="1456" height="799" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:799,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:240978,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/159688779?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0cYn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png 424w, https://substackcdn.com/image/fetch/$s_!0cYn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png 848w, https://substackcdn.com/image/fetch/$s_!0cYn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png 1272w, https://substackcdn.com/image/fetch/$s_!0cYn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3164534-9277-476d-8447-35b0aeb78d04_2246x1232.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ol><li><p>The user enters a prompt or message. Claude Desktop passes the prompt to the LLM.</p></li><li><p>The LLM analyzes the user&#8217;s message in the context of the metadata about the available MCP servers. The LLM checks if it can answer the question on its own or if it needs to invoke one or more of these servers (for tools/resources/prompts). </p></li><li><p>If the LLM doesn&#8217;t need external information, it sends the answering response to Claude Desktop which then displays the final answer back on the user interfaces.</p></li><li><p>If the LLM does need information or to perform an action (e.g. searching, looking up a route, checking Airbnb listings, reading a local file, etc.):</p><ol><li><p>The LLMs send a response to Claude Desktop specifying which MCP server to invoke and with which parameters.</p></li><li><p>Claude Desktop (the &#8220;Host Process&#8221;) routes that request via the correct MCP client to the correct MCP Server.</p></li><li><p>The MCP Server (e.g. Google Maps) processes the request and returns the result to Claude Desktop.</p></li><li><p>Claude Desktop passes the result to the LLM.</p></li><li><p>The LLM uses this information to form an appropriate final answer to the user, and sends it back to Claude Desktop which displays it on the user interface.</p></li></ol></li></ol><h1>Couple More Key Takeaways</h1><p> A very important part of this architecture is the separation of concerns that enable implementation of guardrails and security controls. For example, Claude Desktop which is running with the user&#8217;s privileges can get explicit approvals from the user before invoking a server. Servers, on the other hand, can validate that users have appropriate privileges to make their requests. In addition, the host app/clients can validate the results returned by the LLMs and/or servers for additional business validation (not implemented by Claude Desktop at the moment). </p><p>At a high-level, I also find it useful to think of the LLMs as &#8216;business/user orchestrators&#8217; and the hosts as &#8216;technical/system orchestrators.&#8217; </p><h1>Conclusion</h1><p>For developers and architects building LLM-powered applications, understanding the true mechanics of MCP is crucial. Rather than viewing it as a black box, recognizing the distinct roles and interactions of each component &#8212; the host, the client, the server, and the LLM &#8212; enables clearer implementations, tighter security, and more powerful user experiences.</p><p>As MCP evolves and gains broader adoption, its architectural model may well become a foundational pattern for building the next generation of agentic applications.<br><br><br></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why the Future of Agentic AI Won’t Be Chained: The Case for LLM-Driven Agents]]></title><description><![CDATA[Predicting the shift from developer-defined chains to LLM-driven orchestration.]]></description><link>https://www.modelanalysis.ai/p/why-the-future-of-agentic-ai-wont</link><guid isPermaLink="false">https://www.modelanalysis.ai/p/why-the-future-of-agentic-ai-wont</guid><dc:creator><![CDATA[Ram  Komarraju]]></dc:creator><pubDate>Thu, 03 Apr 2025 08:09:46 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0fe685cb-9278-4717-85b6-60a5024e7a3a_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#171;<strong>Terminology Note</strong>: Throughout this analysis, when I use the term "LLM" (Large Language Model), I&#8217;m referring to the class of foundation models that include not only text generation capabilities but also reasoning, planning, and multimodal capabilities (processing and generating images, audio, video, etc.). This expanded definition reflects the ongoing evolution of these models beyond pure language processing.</em></p><p><em><strong>Series Note</strong>: This article is part of a series of articles I&#8217;m writing on MCP. See my blog for the complete collection.&#187;</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h1><strong>Introduction</strong></h1><p>Model Context Protocol (MCP) has been deservedly receiving significant attention lately. Anthropic calls MCP the "USB-C port for AI applications" in the sense that it provides a standardized way to connect AI models to different data sources and tools. However, another aspect of MCP remains underappreciated &#8211; and this is just as important as standardization. MCP positions the LLM at the center of agentic workflows, enabling a major paradigm shift. This paradigm shift is necessary to unlock the true potential of LLM-driven agentic workflows.</p><p>Meanwhile, many developers approach MCP after having worked with more traditional frameworks that put developers in complete control, viewing MCP as just another " USB-C port for AI applications." However, while both approaches aim to use LLM capabilities for orchestration, they differ significantly in philosophy, architecture, and implementation. Traditional frameworks favor a prescriptive approach, while MCP takes a descriptive approach, letting LLMs drive the workflows. Understanding how MCP compares to established frameworks like LangChain and its workflow extension LangGraph is key to realizing its full potential. (Note that MCP has many other key architectural features such as a clean separation of concerns along with an open standards-based protocol, but covering them isn't the purpose of this article.)</p><p>In this article, I will compare the two approaches and make predictions as to which approach is likely to dominate the future. These predictions may raise a few eyebrows, but I welcome your views on what I might have missed or misunderstood.</p><h1><strong>The Two Paradigms: Prescriptive vs. Descriptive</strong></h1><p>MCP lends itself naturally to a descriptive paradigm where capabilities are described without dictating a rigid path. The LLM's intelligence drives the workflow in real time given the current context and user input. In contrast, frameworks such as LangChain/LangGraph have been used primarily for prescriptive workflows where the developer defines the path. Under this prescriptive paradigm, the LLM has limited autonomy, following a fixed structure designed in code or graph.</p><h2><strong>1. "Prescriptive, Developer-Driven" Workflows</strong></h2><ol><li><p>Developers define the flow of tasks (chains or agents) in code. Each step (or node in a chain) is explicitly laid out&#8212;what input it takes, which tool it calls, what output it returns. This means the workflow is prescribed by the developers. They decide exactly which tools are available, in what order, and under which conditions they are invoked.</p></li><li><p>The framework manages control flow using conditionals, loops, and branching logic written by the developers. While these &#8220;agents&#8221; do allow some LLM-based autonomy, the agent's environment (available tools and instructions) remains largely predetermined in code.</p></li><li><p>Together, these design choices reflect a developer-driven, prescriptive workflow: the path for the LLM is structured in advance through code, and each stage or choice point is orchestrated primarily by the developer's logic. LLM&#8217;s role is typically limited to responding to an exacting prompt further constrained by a narrow context supplied by the orchestrating framework (e.g., Langchain).</p></li></ol><h2><strong>2. "Descriptive, Protocol-Centric, LLM-Driven" Workflows</strong></h2><ol><li><p>A protocol enables tools, data sources, and prompts to describe their capabilities. An LLM is told the capabilities at its disposal and uses these capabilities appropriately without needing a pre-coded workflow. This architecture is descriptive: each capability "announces" what it can do, and the LLM decides how (or whether) to use it.</p></li><li><p>The LLM is effectively "in charge." It reads descriptions of available endpoints, interprets user requests, and decides which capability to call next&#8212;no fixed sequence is imposed by the developer. Control flow emerges from the LLM's reasoning, rather than being prescribed in a Python script or graph.</p></li><li><p>New capability endpoints can appear at any time, and as long as they adhere to the MCP protocol, the LLM can discover and use them. This approach leverages LLM intelligence to figure out the best sequence or combination of capabilities for any given user query, making the workflow LLM-driven and adaptive.</p></li></ol><h2><strong>A comparison of Strengths and Limitations of the two paradigms:</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!K1x2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!K1x2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png 424w, https://substackcdn.com/image/fetch/$s_!K1x2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png 848w, https://substackcdn.com/image/fetch/$s_!K1x2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png 1272w, https://substackcdn.com/image/fetch/$s_!K1x2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!K1x2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png" width="1108" height="1058" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1058,&quot;width&quot;:1108,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:245431,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/160391597?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!K1x2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png 424w, https://substackcdn.com/image/fetch/$s_!K1x2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png 848w, https://substackcdn.com/image/fetch/$s_!K1x2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png 1272w, https://substackcdn.com/image/fetch/$s_!K1x2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4190545e-643d-4581-9edf-dd9050f00369_1108x1058.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MyfB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MyfB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png 424w, https://substackcdn.com/image/fetch/$s_!MyfB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png 848w, https://substackcdn.com/image/fetch/$s_!MyfB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png 1272w, https://substackcdn.com/image/fetch/$s_!MyfB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MyfB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png" width="1094" height="890" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:890,&quot;width&quot;:1094,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:219182,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.modelanalysis.ai/i/160391597?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MyfB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png 424w, https://substackcdn.com/image/fetch/$s_!MyfB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png 848w, https://substackcdn.com/image/fetch/$s_!MyfB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png 1272w, https://substackcdn.com/image/fetch/$s_!MyfB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a925b7b-b77a-4212-a5e4-c56dd9dfb1b0_1094x890.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h1><strong>Long-Term Strategic Prediction: Why the Descriptive Paradigm Will Likely Prevail</strong></h1><p>I predict that LLM-driven protocols like MCP that allow for descriptive approaches will ultimately emerge as the dominant paradigm for agentic orchestration in the medium to long term. Though significant effort will be needed to mature the implementation patterns for guardrails, workflow templating, and testing techniques.</p><h2><strong>The Case for the Descriptive Paradigm as the Long-Term Winner</strong></h2><p><strong>1. Alignment with AI's Evolution Toward Greater Agency: </strong>As LLMs continue to advance in reasoning capabilities, the fundamental limitation will increasingly be the rigidity of the systems around them rather than the models themselves. Descriptive paradigms that empower the LLM to make orchestration decisions align more naturally with the trajectory of AI development toward greater agency and reasoning.</p><p><strong>2. Scalability Across the Capability Explosion: </strong>If the past few months are any indicator, the coming years will see an explosion in the number and variety of specialized AI capabilities, tools, and services that support MCP natively. As the number of capabilities grows linearly, the number of possible workflows grows combinatorially. Explicitly programming all these workflows becomes untenable in a prescriptive approach.</p><p><strong>3. Developer Experience and Productivity: </strong>while frameworks like LangChain offer familiar paradigms for developers, the long-term productivity advantages favor MCP like approaches. Clear separation between capability implementation (e.g., tools, resources, and prompts) and orchestration logic creates better modularity. This allows developers to focus on building and improving individual capabilities rather than maintaining complex workflow graphs. New capabilities can be added or enhanced incrementally without redesigning entire workflows.</p><p><strong>4. Adaptability to User Needs: </strong>a long-term advantage of LLM-driven descriptive approaches is their ability to adapt to the infinite variability of user needs. Users can express their goals in natural language rather than conforming to predefined paths. The system can dynamically adapt based on the specific context of each interaction. And the LLM can incorporate user preferences expressed conversationally without requiring explicit personalization logic.</p><h2><strong>Counter-Arguments and Limitations</strong></h2><p>Despite these advantages, there are legitimate concerns about LLM-driven descriptive approaches that must be addressed:</p><p><strong>1. Control and Reliability Concerns: </strong>The primary argument against LLM-driven approaches is the risk of unpredictable behavior when the LLM makes orchestration decisions. For regulated domains like finance or healthcare, organizations need guaranteed execution of critical compliance steps.<strong> </strong>It's difficult to test all possible paths an LLM might take through a set of capabilities. Handling errors gracefully is more complex when the workflow isn't explicitly defined.</p><p><strong>2. Performance Considerations: </strong>Component-based approaches can potentially optimize performance in ways that dynamic orchestration cannot. Developers can create optimized sequences that minimize latency and resource usage.<strong> </strong>Explicit workflow definitions make it easier to identify opportunities for parallel execution.</p><p><strong>3. Evolution of Prescriptive Frameworks: </strong>As both ecosystems evolve, we're likely to see convergence in some areas. LangChain/LangGraphic can support increasing use of LLMs to dynamically select between predefined subgraphs or to modify graph structure at runtime. I&#8217;m sure there&#8217;ll be efforts to allow these frameworks to be used with MCP.</p><h2><strong>Long Term Architecture and Transition Timelines</strong></h2><p>The most likely long-term outcome is an architecture where:</p><ol><li><p>the core architecture remains protocol-centric, with clear separation of responsibilities, and standardized interfaces for capability discovery and invocation.</p></li><li><p>developers can define recommended workflow templates that guide but don't constrain the LLM.</p></li><li><p>instead of rigid workflows validated with traditional testing, the system provides sophisticated guardrails that ensure critical requirements are met while allowing flexibility within those constraints.</p></li><li><p>organizations can dial the level of LLM autonomy based on domain requirements, starting with more constrained workflows and gradually enabling more dynamic orchestration as trust builds.</p></li></ol><p>The transition will likely follow this pattern:</p><ol><li><p><strong>Short-term (1-2 years)</strong>: Prescriptive frameworks like LangChain/LangGraph will continue to dominate due to developer familiarity and concerns around the reliability of LLMs, especially for domains with strict, well-defined processes and high regulatory scrutiny. However, firms will likely try descriptive, MCP-like, solutions for use cases where the flexibility and power of a descriptive framework can be naturally paired with a human in the loop.</p></li><li><p><strong>Medium-term (2-5 years)</strong>: As the industry&#8217;s experience grows in adapting MCP-like architectures (e.g., how to build workflow templates and guardrails effectively), and LLMs become more reliable and capable, the descriptive paradigm will dominate, with prescriptive approaches used selectively for special cases.</p></li><li><p><strong>Long-term (5+ years)</strong>: Will AGI eventually emerge and obviate the need for any of these discussions?</p></li></ol><h1><strong>Conclusion</strong></h1><p>Both MCP and LangChain/LangGraph represent significant advancements in LLM-powered applications, but with fundamentally different philosophies. LangChain/LangGraph excels in scenarios where workflows need to be precisely defined and controlled, while MCP enables more adaptive, emergent workflows driven by the LLM's understanding of user needs and available capabilities.</p><p>The prescriptive paradigm "works around AI" while the descriptive paradigm "works with AI." The history of technology shows that flexible, adaptable architectures that take advantage of the growing sophistication of underlying technologies (LLMs in this case) tend to outlast more rigid, prescriptive approaches in the long run.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.modelanalysis.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ModelAnalysis.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>