<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Sergiu's Substack - Deep Learning Beyond the Plateau]]></title><description><![CDATA[An interpretive blog on deep learning research, focusing on why models are designed the way they are and how structure shapes learning behavior.]]></description><link>https://sergiunistor.substack.com</link><image><url>https://substackcdn.com/image/fetch/$s_!C6JB!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a1f5ad-2f7c-4338-8c1c-4f424ab56af8_1271x1271.png</url><title>Sergiu&apos;s Substack - Deep Learning Beyond the Plateau</title><link>https://sergiunistor.substack.com</link></image><generator>Substack</generator><lastBuildDate>Tue, 14 Apr 2026 12:16:16 GMT</lastBuildDate><atom:link href="https://sergiunistor.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Sergiu Nistor]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[sergiunistor@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[sergiunistor@substack.com]]></itunes:email><itunes:name><![CDATA[Sergiu Nistor]]></itunes:name></itunes:owner><itunes:author><![CDATA[Sergiu Nistor]]></itunes:author><googleplay:owner><![CDATA[sergiunistor@substack.com]]></googleplay:owner><googleplay:email><![CDATA[sergiunistor@substack.com]]></googleplay:email><googleplay:author><![CDATA[Sergiu Nistor]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Recursive Reasoning with Tiny Networks]]></title><description><![CDATA[How a single 7M-parameter network, a reinterpreted latent space, and a clean ablation study dismantled everything HRM thought it needed.]]></description><link>https://sergiunistor.substack.com/p/recursive-reasoning-with-tiny-networks</link><guid isPermaLink="false">https://sergiunistor.substack.com/p/recursive-reasoning-with-tiny-networks</guid><dc:creator><![CDATA[Sergiu Nistor]]></dc:creator><pubDate>Tue, 17 Mar 2026 17:37:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b826f553-7b60-4b45-8f9f-981335aa7f1c_1200x2372.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>The simplest recursive reasoner</h1><p>There is a pattern in research where a genuinely powerful idea gets buried under the weight of its own justification. Biological analogies, theoretical guarantees, multi-component architectures. All of it accumulated to explain something that turned out to be much simpler once someone actually tested each piece. </p><p><a href="https://arxiv.org/abs/2510.04871">Tiny Recursive Model (TRM)</a>, by Alexia Jolicoeur-Martineau, is an exercise in cutting all of that away and finding out what was load-bearing to begin with.</p><p>It starts as a critique of <a href="https://arxiv.org/abs/2506.21734">Hierarchical Reasoning Model (HRM)</a>. It ends as a replacement.</p><h1>What HRM got right, and where it got complicated</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TFxJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TFxJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 424w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 848w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 1272w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png" width="1456" height="657" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:657,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Hierarchical Reasoning Model: A Brain-Inspired AI Breakthrough&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Hierarchical Reasoning Model: A Brain-Inspired AI Breakthrough" title="Hierarchical Reasoning Model: A Brain-Inspired AI Breakthrough" srcset="https://substackcdn.com/image/fetch/$s_!TFxJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 424w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 848w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 1272w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The original brain-inspired, 2-stage <a href="https://arxiv.org/abs/2506.21734">HRM</a> architecture.</figcaption></figure></div><p>HRM earned its attention. With 27 million parameters and around 1000 training samples, it solved reasoning tasks that stump frontier language models: extremely hard Sudoku, optimal pathfinding through large mazes, and the abstract pattern recognition challenge of <a href="https://arcprize.org/arc-agi">ARC-AGI</a>.</p><p>The core mechanism was a loop. Two networks, alternating through a shared latent space while a training scheme called &#8220;deep supervision&#8221; pushed intermediate representations toward the correct answer at each step.</p><p>To justify the two-network design, the authors drew on neuroscience: a fast low-level module for local computation mirroring rapid cortical processing, a slow high-level module for global planning mirroring deliberate cognition.</p><p>Neither assumption was strictly necessary. That is what TRM demonstrates.</p><h1>One network is enough</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kwWj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kwWj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png 424w, https://substackcdn.com/image/fetch/$s_!kwWj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png 848w, https://substackcdn.com/image/fetch/$s_!kwWj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png 1272w, https://substackcdn.com/image/fetch/$s_!kwWj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kwWj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png" width="390" height="574.86" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:737,&quot;width&quot;:500,&quot;resizeWidth&quot;:390,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Less is More: Recursive Reasoning with Tiny Networks (Paper Review) |  Towards AI&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Less is More: Recursive Reasoning with Tiny Networks (Paper Review) |  Towards AI" title="Less is More: Recursive Reasoning with Tiny Networks (Paper Review) |  Towards AI" srcset="https://substackcdn.com/image/fetch/$s_!kwWj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png 424w, https://substackcdn.com/image/fetch/$s_!kwWj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png 848w, https://substackcdn.com/image/fetch/$s_!kwWj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png 1272w, https://substackcdn.com/image/fetch/$s_!kwWj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd00e3cc2-2116-4bbc-bb90-1f3587ad8cfb_500x737.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The TRM architecture, as per the <a href="https://arxiv.org/pdf/2510.04871">original paper</a>.</figcaption></figure></div><p>HRM used two separate networks, L and H, justified through biological analogy about different processing hierarchies in the brain.</p><p>Replacing both with a single shared network cuts the parameter count in half and still improves accuracy over the two-network version. Shared weights push the network toward learning a more general transformation rather than two partially overlapping specializations.</p><h1>Fewer layers, more recursions</h1><p>Adding layers made things worse. With only around 1000 training examples, additional parameters overfit before they can do anything useful.</p><p>The fix is to go in the opposite direction: drop to 2 layers and scale the number of recursion steps to compensate. Depth from iterating a small network generalizes better than depth from stacking a tall one. This relates to overfitting dynamics on small data, and the empirical gap is large and consistent. </p><h1>Match the architecture to the problem</h1><p>For tasks with a small, fixed context length like <a href="https://ieee-dataport.org/documents/sudoku-extreme">Sudoku Extreme</a> on a 9x9 grid, self-attention is doing more work than the problem requires. A plain MLP applied over the sequence dimension fits the structure of the problem naturally and generalizes better in that setting.</p><p>For larger inputs like the 30x30 mazes in <a href="https://huggingface.co/datasets/sapientinc/maze-30x30-hard-1k">Maze-Hard</a> and the grids in <a href="https://arcprize.org/arc-agi">ARC-AGI</a>, self-attention recovers its advantage. The MLP&#8217;s capacity becomes excessive relative to the problem geometry at that scale, and the inductive biases of attention prove useful.</p><h1>Using EMA for training stability</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8U6w!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8U6w!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png 424w, https://substackcdn.com/image/fetch/$s_!8U6w!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png 848w, https://substackcdn.com/image/fetch/$s_!8U6w!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png 1272w, https://substackcdn.com/image/fetch/$s_!8U6w!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8U6w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png" width="451" height="403" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:403,&quot;width&quot;:451,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:101184,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://sergiunistor.substack.com/i/191251729?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8U6w!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png 424w, https://substackcdn.com/image/fetch/$s_!8U6w!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png 848w, https://substackcdn.com/image/fetch/$s_!8U6w!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png 1272w, https://substackcdn.com/image/fetch/$s_!8U6w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7b43dcb-065c-47bf-a2df-b81dcce2e731_451x403.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Pseudocode for the TRM implementation, as per the <a href="https://arxiv.org/pdf/2510.04871">original paper</a>.</figcaption></figure></div><p>Recursive models trained on tiny datasets are unstable. Sharp overfitting followed by divergence is the typical failure mode.</p><p>Exponential Moving Average of the weights, a stabilization technique established in <a href="https://arxiv.org/abs/1806.04498">GAN training</a>, prevents the worst collapses and raises the generalization ceiling consistently. The gap between training with and without EMA is large enough that it is not an optional detail.</p><h1>Results</h1><p>TRM outperforms HRM across all four benchmarks, <a href="https://ieee-dataport.org/documents/sudoku-extreme">Sudoku Extreme</a>, <a href="https://huggingface.co/datasets/sapientinc/maze-30x30-hard-1k">Maze-Hard</a>, <a href="https://arcprize.org/arc-agi/1/">ARC-AGI-1</a>, and <a href="https://arcprize.org/arc-agi/2/">ARC-AGI-2</a>, while using a fraction of the parameters.</p><p>The LLM comparisons deserve some qualification. The language models it is compared against are evaluated zero-shot, while TRM is trained specifically on these tasks.</p><p>The tasks also test structurally different capabilities: recursive latent refinement over a bounded grid is not the same problem as autoregressive generation over an open vocabulary.</p><p>The puzzle accuracy is real. The cross-paradigm comparison needs to be read carefully.</p><h1>What this leaves open</h1><p>TRM works as a rigorous ablation study that happened to produce a better architecture. Each simplification was tested individually, and each one helped. The result is a system that achieves more with fewer parameters, fewer layers, fewer theoretical assumptions, and a cleaner conceptual framing than what it replaced.</p><p>It is interesting to note how in this particular case recursion over a small network generalizes better than equivalent depth. There is something about iterating a small network on its own outputs, keeping an explicit answer state and a separate reasoning state, and using deep supervision to make each iteration accountable to the final answer, that outperforms simply making the network deeper.</p><p>From an inference systems perspective, there is also an observation the paper does not develop but which follows naturally from the architecture. A 2-layer network run many times has a very different execution profile from a deep network run once, even at a similar compute budget. The recursive structure makes early exit, adaptive computation, and interpretability through the exposed answer state at each step all straightforward extensions. None of that is explored here, but it is a natural next direction.</p><p>The code is open-source at <a href="https://github.com/SamsungSAILMontreal/TinyRecursiveModels">github.com/SamsungSAILMontreal/TinyRecursiveModels</a>.</p>]]></content:encoded></item><item><title><![CDATA[Hierarchical Reasoning Model]]></title><description><![CDATA[How a 27M-parameter brain-inspired architecture beat the world on ARC-AGI.]]></description><link>https://sergiunistor.substack.com/p/hierarchical-reasoning-model</link><guid isPermaLink="false">https://sergiunistor.substack.com/p/hierarchical-reasoning-model</guid><dc:creator><![CDATA[Sergiu Nistor]]></dc:creator><pubDate>Sun, 15 Mar 2026 19:20:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TFxJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>Introduction</h1><p>There is a quiet but radical assumption baked into modern large language models: that thinking happens in language. Ask a frontier model to solve a hard problem, and it will reason by writing, producing chains of tokens that trace the path from question to answer. This works remarkably well. It is also, arguably, completely backwards.</p><p>A growing body of neuroscience suggests that human abstract thought does not primarily occur in language. The brain solves problems first, and translates solutions into words afterward. Language, in this view, is a communication interface, not a cognitive engine. This distinction has a concrete architectural consequence: if reasoning doesn&#8217;t require tokens, maybe we&#8217;ve been building our reasoners that are simply inefficient.</p><p>Developed by <a href="https://sapient.inc/">Sapient Intelligence</a>, the <a href="https://arxiv.org/pdf/2506.21734">Hierarchical Reasoning Model (HRM)</a> is a bet on exactly this idea. Released in August 2025, it achieved State-of-the-Art results on both ARC-AGI-1 and ARC-AGI-2 (widely regarded as one of the hardest tests of abstract generalization in AI) with just <strong>27 million parameters</strong>, trained on only <strong>1,000 samples</strong>, and <strong>no pretraining</strong> on large corpora. The frontier LLMs it outperformed on these benchmarks are orders of magnitude larger.</p><p>This post breaks down what HRM is, where it comes from intellectually, and why its design choices might represent something more than a clever trick.</p><h1>Reasoning without tokens</h1><p>Chain-of-Thought prompting works by making the model externalize its reasoning as text. The intuition is solid: intermediate scaffolding helps. But it&#8217;s slow, expensive, and the token space (which is discrete, surface-level, linguistic) may just be the wrong medium for certain kinds of abstract computation.</p><p>HRM&#8217;s core idea is that reasoning can happen entirely in the latent state space of a recurrent network, never surfacing as language. The model doesn&#8217;t narrate its thinking. It performs its thinking internally, across multiple recurrent steps, and only produces output once that process has converged. Critically, none of those intermediate steps are supervised. the model only receives feedback on the final answer, not on how it got there.</p><p>The intellectual backing for this comes from the idea that <a href="https://www.nature.com/articles/s41586-024-07522-w">language is primarily a tool for communication rather than thought</a>. Non-human primates solve complex problems without it. Patients with severe aphasia retain intact reasoning. Mathematical and musical cognition run on entirely separate neural systems. If thinking doesn&#8217;t require language in the brain, a model forced to write out its reasoning has an architectural bottleneck the brain simply doesn&#8217;t have.</p><h1>The architecture</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TFxJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TFxJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 424w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 848w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 1272w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png" width="1456" height="657" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:657,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Hierarchical Reasoning Model: A Brain-Inspired AI Breakthrough&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Hierarchical Reasoning Model: A Brain-Inspired AI Breakthrough" title="Hierarchical Reasoning Model: A Brain-Inspired AI Breakthrough" srcset="https://substackcdn.com/image/fetch/$s_!TFxJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 424w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 848w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 1272w, https://substackcdn.com/image/fetch/$s_!TFxJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e73879-52a0-4af0-ac6e-f3f28bf7743a_1600x722.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">HRM&#8217;s brain-inspired, hierarchical architecture.</figcaption></figure></div><p>HRM is a recurrent network with two modules, inspired directly by <a href="https://dn790002.ca.archive.org/0/items/DanielKahnemanThinkingFastAndSlow/Daniel%20Kahneman-Thinking%2C%20Fast%20and%20Slow%20%20.pdf">Kahneman&#8217;s System 1 / System 2 framework</a> and by three observations from neuroscience:</p><ul><li><p>Different cortical regions of the brain operate at different speeds, and this layered organization is what enables deep, multi-step reasoning (<a href="https://www.nature.com/articles/s41467-023-37613-7">Zeraati et al.</a>, <a href="https://www.nature.com/articles/nn.3862">Murray et al.</a>, <a href="https://pubmed.ncbi.nlm.nih.gov/29203085/">Huntenburg et al.</a>)</p></li><li><p>Recurrent feedback loops continuously refine the network&#8217;s internal representations, with slow higher-level areas guiding the fast lower-level ones, without losing global coherence (<a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC3777738/">Bastos et al.</a>, <a href="https://www.sciencedirect.com/science/article/abs/pii/S016622360001657X">Lamme et al.</a>, <a href="https://openreview.net/forum?id=xavWvnJTST">Kaleb et al.</a>)</p></li><li><p>The brain achieves this computational depth without the credit-assignment problem that makes training deep recurrent networks so expensive (<a href="https://www.sciencedirect.com/science/article/pii/S0959438818302009">Lillicrap et al.</a>)</p></li></ul><p>This maps directly onto two coupled recurrent modules:</p><ul><li><p><strong>High-level (H) module</strong> - slow, abstract, responsible for planning and goal maintenance</p></li><li><p><strong>Low-level (L) module</strong> - fast, detailed, executing rapid computations guided by H</p></li></ul><p>The L module runs multiple steps until it reaches a local equilibrium. Only then does H advance, incorporating L&#8217;s converged state to update its higher-level representation. L resets and begins a new phase. The slow areas wait for the fast ones, which is also, broadly, what the brain does.</p><h3>Solving the RNN collapse problem</h3><p>Anyone who&#8217;s worked with recurrent networks knows the failure mode: they collapse early in training.</p><p>HRM introduces the concept of <strong>hierarchical convergence</strong> to fix this. H is only allowed to update <em>after</em> L has completed its full cycle and settled. This forces the network to actually use its recurrent structure and not short-circuit.</p><p>It&#8217;s both the paper&#8217;s most important technical contribution and its most elegant one.</p><h1>The results</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-DtP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-DtP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png 424w, https://substackcdn.com/image/fetch/$s_!-DtP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png 848w, https://substackcdn.com/image/fetch/$s_!-DtP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png 1272w, https://substackcdn.com/image/fetch/$s_!-DtP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-DtP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png" width="1400" height="567" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:567,&quot;width&quot;:1400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Killing LMMs with a Tiny 27M model: Hierarchical Reasoning Model (HRM)  Briefly Explained | by Allen Liang | Aug, 2025 | Medium | Medium&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Killing LMMs with a Tiny 27M model: Hierarchical Reasoning Model (HRM)  Briefly Explained | by Allen Liang | Aug, 2025 | Medium | Medium" title="Killing LMMs with a Tiny 27M model: Hierarchical Reasoning Model (HRM)  Briefly Explained | by Allen Liang | Aug, 2025 | Medium | Medium" srcset="https://substackcdn.com/image/fetch/$s_!-DtP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png 424w, https://substackcdn.com/image/fetch/$s_!-DtP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png 848w, https://substackcdn.com/image/fetch/$s_!-DtP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png 1272w, https://substackcdn.com/image/fetch/$s_!-DtP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff954000d-0f1b-4b2c-9234-7823eb77e8e8_1400x567.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">HRM performance on <a href="https://arcprize.org/arc-agi/1/">ARC-AGI-1</a>, <a href="https://arcprize.org/arc-agi/2/">ARC-AGI-2</a>, <a href="https://ieee-dataport.org/documents/sudoku-extreme">Sudoku-Extreme</a> and <a href="https://huggingface.co/datasets/sapientinc/maze-30x30-hard-1k">Maze-Hard</a>.</figcaption></figure></div><p><a href="https://arcprize.org/arc-agi">ARC-AGI</a> was designed by Fran&#231;ois Chollet specifically to resist memorization. The tasks require genuine abstract generalization across novel visual grids. ARC-AGI-2 is harder still, designed to resist the pattern-matching tricks that helped models game the first benchmark.</p><p>HRM achieved State-of-the-Art on both benchmarks (at that time).</p><p>The model couldn&#8217;t have memorized its way there. The architecture is doing the work, and that&#8217;s a meaningful data point for a few ongoing debates: whether scale can always substitute for structure and invariance, and whether token-level reasoning is a fundamental requirement or a workaround.</p><p>It doesn&#8217;t resolve any of those debates. But it sharpens them.</p><h1>Caveats and conclusions</h1><p>HRM isn&#8217;t a general-purpose language model. It doesn&#8217;t generate text or handle open-ended tasks. The benchmark comparison to much larger models needs to be made carefully. Scaling behavior is unknown. And the biological inspiration, while compelling, is a motivation rather than a guarantee.</p><p>The code is open-source at <a href="https://github.com/sapientinc/HRM">github.com/sapientinc/HRM</a>.</p><p>The next post will cover <strong>TRM (Tiny Recursive Model)</strong>, an optimized successor described in <a href="https://arxiv.org/pdf/2510.04871">Alexia Jolicoeur-Martineau's paper</a>, which pushes the hierarchical recurrent framework further and starts to answer some of the scaling questions.</p>]]></content:encoded></item><item><title><![CDATA[When Scaling Isn't Enough]]></title><description><![CDATA[A look at where deep learning research stands today, why scaling alone won't take us further, and what the field actually needs.]]></description><link>https://sergiunistor.substack.com/p/when-scaling-isnt-enough</link><guid isPermaLink="false">https://sergiunistor.substack.com/p/when-scaling-isnt-enough</guid><dc:creator><![CDATA[Sergiu Nistor]]></dc:creator><pubDate>Fri, 13 Mar 2026 16:47:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/f9d3fc99-5bc5-447a-8a0b-8ad9d222985f_2048x1152.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Deep learning has come a long way, but the field is showing signs of stagnation. Most progress today comes from scaling existing architectures rather than rethinking them, and the gap between what large labs can build and what independent researchers can meaningfully explore keeps widening. </p><p>This blog is a response to that: a space for examining the reasoning behind model design, questioning assumptions that have gone unquestioned for too long, and making the case that the next wave of breakthroughs will come from new ideas, not just more compute.</p><h1>When Research Plateaus</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6NWF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6NWF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png 424w, https://substackcdn.com/image/fetch/$s_!6NWF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png 848w, https://substackcdn.com/image/fetch/$s_!6NWF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png 1272w, https://substackcdn.com/image/fetch/$s_!6NWF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6NWF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png" width="728" height="550.9546279491833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:834,&quot;width&quot;:1102,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:855884,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://sergiunistor.substack.com/i/182040647?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6NWF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png 424w, https://substackcdn.com/image/fetch/$s_!6NWF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png 848w, https://substackcdn.com/image/fetch/$s_!6NWF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png 1272w, https://substackcdn.com/image/fetch/$s_!6NWF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd529e36-3aa9-4181-ad32-c9be6cd95d5f_1102x834.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Depiction of some of the most cited Computer Science research papers, according to <a href="https://paperscape.org/">Paperscape</a>.</figcaption></figure></div><p>Deep learning has achieved remarkable successes, yet the field is increasingly converging around a few dominant architectures, particularly large transformers and LLMs. Most current innovation focuses on <a href="https://arxiv.org/abs/2001.08361">scaling existing models</a> rather than exploring fundamentally new ideas, creating a plateau where architectural decisions are repeated and the reasoning behind them is rarely questioned.</p><p>Exceptions exist, but they are rare, and most still do not deviate much from the current deep learning standards. Some notable recent examples are hierarchical models (<a href="https://arxiv.org/abs/2506.21734">HRMs</a>, <a href="https://arxiv.org/abs/2510.04871">TRMs</a>), state-space models (<a href="https://arxiv.org/abs/2312.00752">Mamba</a>), attention-free transformer alternative models (<a href="https://arxiv.org/pdf/2305.13048">RWKV</a>), and hybrid SSM-transformer models (<a href="https://research.nvidia.com/labs/nemotron/files/NVIDIA-Nemotron-3-White-Paper.pdf">Nemotron 3</a>).</p><p>The deeper issue is that a single idea can only take us so far. The field needs new approaches and perspectives, because scaling alone has limits. Some researchers consider that <a href="https://www.businessinsider.com/meta-yann-lecun-scaling-ai-wont-make-it-smarter-2025-4">scaling laws may eventually stop holding</a>. Beyond performance, progress should not be measured only by model size: <a href="https://arxiv.org/abs/2104.10350">Environmental impact</a> and resource requirements are real constraints that cannot be ignored.</p><p>At the same time, accessibility in research is increasingly constrained, as the validation of most new ideas now requires large-scale infrastructure and computational power. This creates a barrier for smaller labs and independent researchers, limiting their ability to explore innovative approaches, test unconventional hypotheses, and meaningfully contribute to the advancement of the field. The current situation underscores the reality of an inevitable <a href="https://en.wikipedia.org/wiki/AI_winter">AI winter</a>.</p><h1>Driving the Next Wave of AI</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Uv-1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Uv-1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png 424w, https://substackcdn.com/image/fetch/$s_!Uv-1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png 848w, https://substackcdn.com/image/fetch/$s_!Uv-1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png 1272w, https://substackcdn.com/image/fetch/$s_!Uv-1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Uv-1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png" width="1456" height="890" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:890,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:389444,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://sergiunistor.substack.com/i/182040647?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Uv-1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png 424w, https://substackcdn.com/image/fetch/$s_!Uv-1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png 848w, https://substackcdn.com/image/fetch/$s_!Uv-1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png 1272w, https://substackcdn.com/image/fetch/$s_!Uv-1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb53d04d-53c1-480e-9777-b138e0a482b4_2400x1467.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Frontier model training compute requirement evolution over time, as per <a href="https://epoch.ai/blog/training-compute-of-frontier-ai-models-grows-by-4-5x-per-year">Epoch AI</a>.</figcaption></figure></div><p>True advancement in the field depends on researchers exploring and inventing novel architectures, techniques, and methods, rather than merely scaling up existing models. By focusing on creativity and innovation, new researchers can uncover more efficient, interpretable, and versatile solutions, pushing the boundaries of what is possible without being limited by the brute-force approach of ever-larger models.</p><p>The field needs people who combine creativity, intuition, and rigorous thinking to tackle underexplored ideas, propose alternative architectures, and test unconventional learning mechanisms. Diversifying who contributes and what approaches are investigated is essential to break out of incremental cycles and drive the next wave of breakthroughs in deep learning.</p><p>However, starting research today can feel like navigating a landscape full of proven successes but few clear paths.</p><p>There is a pressing need for a centralized resource that traces the development of brilliant deep learning and machine learning ideas, showing how concrete objectives, such as implementing memory or attention, led to the design of specific mathematical mechanisms that guide the model&#8217;s learning.</p><p>These insights are typically buried in papers, which naturally emphasize scientific rigor over clarity for a wider audience. As a result, the ideas are often too complex or lengthy for most readers to fully absorb.</p><h1>Understanding Model Design</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5QGS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5QGS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png 424w, https://substackcdn.com/image/fetch/$s_!5QGS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png 848w, https://substackcdn.com/image/fetch/$s_!5QGS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png 1272w, https://substackcdn.com/image/fetch/$s_!5QGS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5QGS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png" width="234" height="339.195067264574" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1293,&quot;width&quot;:892,&quot;resizeWidth&quot;:234,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Attention is All You Need. Breaking Down the Architecture Behind&#8230; | by  Hadis | Towards AI&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Attention is All You Need. Breaking Down the Architecture Behind&#8230; | by  Hadis | Towards AI" title="Attention is All You Need. Breaking Down the Architecture Behind&#8230; | by  Hadis | Towards AI" srcset="https://substackcdn.com/image/fetch/$s_!5QGS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png 424w, https://substackcdn.com/image/fetch/$s_!5QGS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png 848w, https://substackcdn.com/image/fetch/$s_!5QGS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png 1272w, https://substackcdn.com/image/fetch/$s_!5QGS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e803a7e-80ce-4516-b192-3ebd88623152_892x1293.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The original transformer architecture, <a href="https://arxiv.org/pdf/1706.03762">Vaswani et al.</a></figcaption></figure></div><p>Deep learning remains largely a black box. We know <em>what</em> works, but not always <em>why</em>. While models achieve impressive results, the internal mechanisms that drive their success often remain opaque. This uncertainty has spurred the field of mechanistic interpretability, with groups like Anthropic working to uncover the principles behind neural network behavior (<a href="https://www.anthropic.com/research/interpretability-dreams">Interpretability Dreams</a>, <a href="https://www.anthropic.com/research/decomposing-language-models-into-understandable-components">Decomposing Language Models Into Understandable Components</a>, <a href="https://www.anthropic.com/research/mapping-mind-language-model">Mapping the Mind of a Large Language Model</a>).</p><p>Many design choices in deep learning only make sense in hindsight. Architectural tweaks, normalization methods, and optimization strategies are frequently justified empirically rather than theoretically. Mechanisms that seem arbitrary at first reveal their value only after extensive experimentation.</p><p>Much of our intuition comes from the brain, our only partially understood example of intelligence. Concepts such as <a href="https://arxiv.org/abs/1706.03762">attention</a>, <a href="https://ieeexplore.ieee.org/abstract/document/6795963">memory</a>, and hierarchical processing are translated into mechanisms that neural models can implement. These abstractions often succeed in practice, but the reasons for their effectiveness or optimality are rarely clear.</p><p>Meaningful research requires a rare blend of skills: the ability to translate abstract, human-intuitive concepts into precise mathematical constructs that a neural network can internalize. It&#8217;s this bridge between practical intuition and implementation that drives innovation beyond mere trial-and-error.</p><h1>What Readers Can Expect</h1><p>This blog explores the reasoning behind model design, how structure shapes learning, and the principles that guide research. By examining the foundations of deep learning and machine learning, we aim to spark innovation, encourage critical thinking, and support a more diverse and capable research community.</p><p>You&#8217;ll find practical explorations of deep learning and machine learning research: analyses of novel architectures, experiments that question conventional assumptions, and reflections on why certain design choices succeed or fail. Each post links core intuitions with the mathematical principles that govern model behavior.</p><p>This space is meant for active engagement. I welcome discussions, critiques, and collaborations, whether you are a researcher testing new concepts or a learner exploring the fundamentals.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://sergiunistor.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading my blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>