<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Upcurious]]></title><description><![CDATA[We track where capital is going and what AI tools can actually do. Research for investors. Guides for builders]]></description><link>https://upcurious.substack.com</link><image><url>https://substackcdn.com/image/fetch/$s_!Pkhj!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ddf3a53-a048-4f0e-bb10-96f06dbfe75f_160x160.jpeg</url><title>Upcurious</title><link>https://upcurious.substack.com</link></image><generator>Substack</generator><lastBuildDate>Wed, 06 May 2026 09:38:55 GMT</lastBuildDate><atom:link href="https://upcurious.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Upcurious]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[upcurious@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[upcurious@substack.com]]></itunes:email><itunes:name><![CDATA[Upcurious]]></itunes:name></itunes:owner><itunes:author><![CDATA[Upcurious]]></itunes:author><googleplay:owner><![CDATA[upcurious@substack.com]]></googleplay:owner><googleplay:email><![CDATA[upcurious@substack.com]]></googleplay:email><googleplay:author><![CDATA[Upcurious]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[How Agentic Work Broke Token Pricing]]></title><description><![CDATA[Seven frontier AI models in 18 days. The price page no longer tells you what a task costs.]]></description><link>https://upcurious.substack.com/p/how-agentic-work-broke-token-pricing</link><guid isPermaLink="false">https://upcurious.substack.com/p/how-agentic-work-broke-token-pricing</guid><dc:creator><![CDATA[Upcurious]]></dc:creator><pubDate>Mon, 27 Apr 2026 09:50:37 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/afc9a547-4dca-4fbe-aa76-b62020181b04_1200x634.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The instinct is to rank AI models on price-per-token and SWE-Bench score. That ranking no longer tracks what users actually pay.</p><p>Three signals broke at once:</p><ul><li><p><strong>Tokenizers became a hidden price lever.</strong> Opus 4.7 raised bills up to 35% without touching the sticker.</p></li><li><p><strong>Benchmarks became an optimization target.</strong> Three Chinese open-weights labs cleared SWE-Bench Pro inside the 18-day window.</p></li><li><p><strong>The harness layer now drives more cost-per-task variance than the model name does.</strong> SemiAnalysis measured Claude Code consuming roughly 25% more input context per output token than Codex on equivalent coding work, on the same underlying model.</p></li></ul><p>The unit enterprises actually care is <strong>cost per successful task</strong>: token price &#215; tokenizer expansion &#215; harness token behaviour &#215; retries, divided by task success rate. </p><div><hr></div><h2>What Landed in 18 Days</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rkG1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rkG1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png 424w, https://substackcdn.com/image/fetch/$s_!rkG1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png 848w, https://substackcdn.com/image/fetch/$s_!rkG1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png 1272w, https://substackcdn.com/image/fetch/$s_!rkG1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rkG1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png" width="1456" height="876" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:876,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4751290,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://upcurious.substack.com/i/195602700?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rkG1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png 424w, https://substackcdn.com/image/fetch/$s_!rkG1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png 848w, https://substackcdn.com/image/fetch/$s_!rkG1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png 1272w, https://substackcdn.com/image/fetch/$s_!rkG1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce0f6f6b-e5ee-48d2-ac52-c868021ffb91_2450x1474.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Three patterns visible across the cluster.</p><p><strong>Closed labs are raising prices, sometimes invisibly.</strong> Opus 4.7 took the hidden route: the sticker did not move, the tokenizer did. GPT-5.5 took the visible route, doubling per-token pricing against GPT-5.4 to $5 / $30, with a Pro tier at $30 / $180. Anthropic&#8217;s other play, Mythos (unreleased frontier model), restricted to a small set of launch partners and not slated for general release, has reported SWE-Bench Pro at 77.8%, well above anything publicly comparable. The price moves and the Mythos restriction are two angles on the same play: protect margin where capability is genuinely above the open-weights wave; raise prices where it is not.</p><p><strong>Open-weights labs are pricing for adoption, not yield.</strong> MiMo V2.5 Pro at $1 / $3 per million tokens is the load-bearing data point on this side, roughly a fifth of GPT-5.5 at frontier-adjacent capability on the public benchmarks. That is not a sustainable yield-curve price. It is a market-entry price: get production workloads onto the platform, accumulate harness-tuning data, price up later. Z.ai and Moonshot are running similar economics. Real per-task gaps are smaller than the 5x sticker once harness behaviour is accounted for, still large enough to move procurement.</p><p><strong>Meta added a closed consumer model at the open-weights peak.</strong> Muse Spark landed two days after GLM-5.1 became the first open-weights model to top SWE-Bench Pro. That does not mean Meta has abandoned open source. Zuckerberg has said Meta still plans new open-source models. It does mean Meta is separating its consumer-product strategy from the raw model-access strategy: frontier capability can be routed first into Meta AI, with planned rollout across WhatsApp, Instagram, Facebook, Messenger, and Meta smart glasses, without being released as weights on day one. It is a different vertical-integration move from Anthropic&#8217;s, on a different end of the stack.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zKgS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zKgS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png 424w, https://substackcdn.com/image/fetch/$s_!zKgS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png 848w, https://substackcdn.com/image/fetch/$s_!zKgS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png 1272w, https://substackcdn.com/image/fetch/$s_!zKgS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zKgS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png" width="1456" height="772" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:772,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:444838,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://upcurious.substack.com/i/195602700?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zKgS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png 424w, https://substackcdn.com/image/fetch/$s_!zKgS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png 848w, https://substackcdn.com/image/fetch/$s_!zKgS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png 1272w, https://substackcdn.com/image/fetch/$s_!zKgS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7151c3a6-79db-4d0f-993a-6d6919c99073_3440x1824.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This is the cost-per-task argument in action. Running the full Artificial Analysis Intelligence Index evaluation costs $4,811 on Opus 4.7 max-config and $3,357 on GPT-5.5 xhigh, against $1,071 on DeepSeek V4 Pro Max and $462 on MiMo V2.5 Pro. The closed-source max configurations sit at roughly 10x the open-weights frontier on the same benchmark suite, and reasoning cost dominates that bill: 60&#8211;75% of the total on closed-source max configurations, close to zero on the open-weights side. Configuration is itself a cost lever within a single model. GPT-5.5 ranges from $1,199 (medium) to $3,357 (xhigh), a 2.8x spread on the same weights. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZnnR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZnnR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png 424w, https://substackcdn.com/image/fetch/$s_!ZnnR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png 848w, https://substackcdn.com/image/fetch/$s_!ZnnR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png 1272w, https://substackcdn.com/image/fetch/$s_!ZnnR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZnnR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png" width="1456" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1542032,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://upcurious.substack.com/i/195602700?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZnnR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png 424w, https://substackcdn.com/image/fetch/$s_!ZnnR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png 848w, https://substackcdn.com/image/fetch/$s_!ZnnR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png 1272w, https://substackcdn.com/image/fetch/$s_!ZnnR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ce943c-d557-4d2c-874b-776a290a803a_3126x1288.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>The Public Signal Is Now Broken</h2><p>Five things have gone wrong with the signals at the same time.</p><p><strong>Benchmark optimization has saturated the open-weights cluster.</strong> SWE-Bench Pro was meant to be the contamination-resistant successor to SWE-Bench, with private repositories used to keep training data clean. Within months of the benchmark becoming the industry reference, three Chinese open-weights labs have published models that occupy the top of the leaderboard, all within a 18-day window. That is not impossible to achieve honestly, and the underlying capability is real. It also fits the pattern of a benchmark that has become the optimization target rather than the measurement instrument. The contamination asymmetry runs in one direction: open-weights labs face fewer institutional checks on training data composition than closed labs do, and the cost of getting caught is lower. The more defensible read is not that the scores are fake; it is that leaderboard position now needs production evidence before it should be treated as durable capability.</p><p><strong>Closed labs curate benchmark framing, too.</strong> OpenAI publicly campaigned for SWE-Bench Pro as the right industry benchmark in February 2026, and the GPT-5.5 announcement did publish a SWE-Bench Pro score: 58.6%, with a memorization caveat. That is better than omission, but it does not remove the disclosure problem. Closed labs still choose which benchmarks get headline treatment, which internal evals stay private, and how caveats are framed. SemiAnalysis&#8217;s broader read is that Mythos and Opus 4.7 remain stronger in some important zones even where GPT-5.5 looks better on selected public numbers. The benchmark game is asymmetric on both sides. Open-weights labs optimize for the score; closed-source labs optimize the disclosure frame. Both compromise the public signal.</p><p><strong>Opus 4.6 and 4.7 had visible production trust issues.</strong> Anthropic&#8217;s Claude Code product carried three production bugs that affected many users before the fix posted. Engineers using Claude Code through the 4.6 and 4.7 cycle reported behaviour changes that did not match the clean release-note story: more brittle instruction-following, unexpected regressions, and tool-use issues that varied by harness and effort level. The important point is not intent. It is that the closed-source trust premium (the reason an enterprise pays $5 / $25 for Opus rather than $1 / $3 for MiMo) depends on stability, disclosure, and predictable migration behaviour. Anthropic spent some of that premium in this cycle.</p><p><strong>Tokenizer drift, in detail.</strong> Opus 4.7&#8217;s change is the cleanest case. A workload migrated from 4.6 to 4.7 today can cost up to 35% more for the same inputs and outputs, with all of the change driven by how the model encodes text rather than what the model does with it. </p><p><strong>Harness as the dominant cost variable.</strong> A coding agent does not run a single forward pass. It plans, reads files, drafts, runs tests, reads errors, redrafts. Each step is a token decision driven by the harness: what context to load, when to compact, what tools to call, when to stop. Codex and Claude Code on the same underlying model produce materially different per-task costs. Cursor and OpenCode on the same model produce materially different costs again. The harness is now part of the product, and it is where most of the cost-per-task delta lives. Two of the three inputs to real cost &#8212; tokenizer behaviour and harness behaviour &#8212; sit outside the standard pricing comparison.</p><div><hr></div><h2>Updating the Demand Engine Thesis</h2><p>Two weeks ago, Anthropic&#8217;s Managed Agents launch read as a vertical-integration play: Anthropic capturing the runtime moat by selling a managed agent platform on top of its model rather than letting customers wire up their own. That framing is intact. The April cluster bounds the upside of that thesis tighter than the prior piece suggested.</p><p>Even discounting the benchmark-gaming concern from the prior section, the floor under managed pricing is bounded by two things that do not depend on leaderboard truth: open-weights pricing and architectural memory efficiency. MiMo at $1 / $3 versus GPT-5.5 at $5 / $30 is a procurement decision regardless of who tops SWE-Bench Pro. DeepSeek V4 cutting KV cache by an order of magnitude is structural inference economics regardless of any capability claim. The floor is no longer one open-weights model. It is a wave. GLM-5.1 cleared the closed-source coding leaderboard at zero per-token cost beyond inference compute. Kimi K2.6 followed within two weeks. An enterprise that wanted to run a long-running agent workflow on closed infrastructure two weeks ago was paying to keep HBM pinned for the length of the session. That same enterprise today has at least four credible self-hosting options, three with open weights, one with first-party pricing that approaches commodity.</p><p>This does not collapse the managed-runtime business. It bounds the pricing power. Anthropic and OpenAI can charge a premium where reliability, security review, accountability, and integrated tooling matter more than unit cost, which is most regulated enterprises today. They cannot charge an unbounded premium, because the cost of the alternative just dropped sharply on multiple axes simultaneously. The premium is bounded above by what an enterprise can replicate with open weights and competent infrastructure, and that ceiling moved down materially.</p><p>Deepseek V4&#8217;s reception is the cleanest case of the broken signal. Every public signal interprets this release as underwhelming. The architecture says the opposite.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!r-8-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!r-8-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png 424w, https://substackcdn.com/image/fetch/$s_!r-8-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png 848w, https://substackcdn.com/image/fetch/$s_!r-8-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png 1272w, https://substackcdn.com/image/fetch/$s_!r-8-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!r-8-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png" width="1412" height="434" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:434,&quot;width&quot;:1412,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:102824,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://upcurious.substack.com/i/195602700?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!r-8-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png 424w, https://substackcdn.com/image/fetch/$s_!r-8-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png 848w, https://substackcdn.com/image/fetch/$s_!r-8-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png 1272w, https://substackcdn.com/image/fetch/$s_!r-8-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb135764e-faa3-4607-9fb7-4481f230d9f7_1412x434.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The disappointment narrative is the public signal failing to capture the structural one.</p><div><hr></div><h2>Efficiency, Jevons, and the Demand Engine</h2><p>The compression wave is real. The demand engine is also real. The empirical question is whether efficiency outpaces usage growth, or usage growth absorbs the efficiency gain.</p><p>The OpenRouter scoreboard is the cleanest public signal on this. Weekly token volume on the platform is up roughly four times year-over-year, from approximately 5 trillion in April 2025 to over 20 trillion in April 2026. The fastest-growing behaviour is agentic inference: single requests pushed from thousands of tokens to 100K-1M tokens, run in long sequences with tool calls. That growth happened against continuous efficiency improvement: better quantization, sparse attention, prompt caching, KV compaction. None of it slowed the demand curve.</p><p>The composition shift on OpenRouter is the second signal worth holding. Chinese-origin models now account for over 45% of platform traffic, up from under 2% in late 2024. That number reflects two things at once: real-world quality adoption (engineers running production workloads on cheaper open-weights models that perform well enough) and price-driven traffic that does not reflect quality choices. Both interpretations advance the cost-per-task argument: enterprises are increasingly willing to migrate workloads onto cheaper models, either because the quality holds or because the cost differential is large enough to absorb the quality risk.</p><p>DeepSeek V4 will follow the same pattern. This is the Jevons shape: cutting long-context memory by an order of magnitude does not deflate inference demand, it expands the set of workloads where million-token context is economically justifiable, which means more million-token sessions, which means more sustained HBM demand in aggregate even as each session uses less HBM per unit of work. Total memory load goes up unless capability ceilings hold demand back, and the demand ceiling has been receding for three years.</p><p>The counter evidence worth holding: capability dominance. If Mythos, Opus 4.7, and GPT-5.5 widen the outcome-quality gap meaningfully against open weights (and if benchmark-optimization on the open-weights side reflects gaming rather than real capability), cost-per-task comparisons matter less because only the frontier closed model finishes the job reliably. SemiAnalysis&#8217;s read is that the capability gap still applies in narrow zones: Mythos on cybersecurity, Opus on long open-ended tasks, GPT-5.5 on dense reasoning. Outside those zones, the cost frame increasingly drives selection. Leaderboard dominance from the open-weights cluster does not by itself prove the cost frame has won. It does confirm the cost frame is now being seriously tested.</p><div><hr></div><h2>What This Means for Where Value Sits</h2><p>The thesis from this series holds, with the following updates.</p><p><strong>Memory, packaging, and power names: thesis intact and reinforced.</strong> Seven frontier-class products land on the same constrained infrastructure. SK Hynix sold out through 2026, Micron sold out through 2026, TSMC CoWoS reserved by NVIDIA. The DeepSeek and MiMo efficiency improvements expand the demand pool rather than relieve the supply constraint. The bottleneck names remain the cleanest public expression of this trade.</p><p><strong>Harness layer: analytically real, structurally private.</strong> Cursor, Cognition, OpenCode, and the first-party harnesses inside frontier labs are where cost-per-task is determined. There is no clean public-equity expression of this layer today. Watch for acquisitions, IPO filings, and any disclosure of session-level economics. Anthropic&#8217;s eventual S-1, rumoured for October 2026, will be the first published view of harness economics from the inside.</p><p><strong>Managed-runtime thesis: bounded above by an open-weights wave, not one model.</strong> The prior piece&#8217;s read on Managed Agents was directionally right. The ceiling on that thesis is now visible and lower than two weeks ago. The product is real, the moat is real, the pricing power is bounded by what an enterprise can replicate with open-weights infrastructure at frontier-adjacent capability, which is now four models, including one (MiMo) with first-party pricing at roughly a fifth of the closed-source frontier.</p><p><strong>At risk: API-resold middleware.</strong> Companies that built features on top of frontier APIs without owning either the model or the harness layer face compression from both directions. Frontier labs are descending into the harness. Open weights are rising into capability ranges that previously required closed APIs. The middle of the stack is where the squeeze lands first.</p><p><strong>Trust premium: under explicit pressure.</strong> Open-weights pricing is now low enough and capability close enough that enterprises will require predictable migration behaviour, stable tokenizers, and a low-babysitting harness to justify the closed-source premium. If closed labs spend that capital faster than Mythos-tier capability replenishes it, the cost-per-task argument starts winning workloads where it currently does not.</p><p><strong>Subscription pricing is breaking too, for the same reason.</strong> Cursor shifted from &#8220;unlimited&#8221; to capped and usage-metered tiers in 2025 after power users running coding agents burned through subscription compute orders of magnitude faster than the plans were designed to absorb. Claude Pro and Max have weekly limits and throttling on heavy users. ChatGPT Plus has had message caps from the start. GitHub Copilot, Replit, and Cognition&#8217;s Devin have all moved toward usage-metered or &#8220;premium request&#8221; tiers. Per-token pricing breaks the procurement comparison; subscription pricing breaks the consumer plan. Both were structures designed for an earlier consumption pattern, and both are compressing under the same agentic load.</p><div><hr></div><h2>What We Are Watching</h2><p><strong>Big Tech earnings, in flight now.</strong> Meta, Google, Microsoft, and Amazon report through late April and early May. The number that matters is whether 2026 capex guidance moves up against the agent-driven demand signal. Hyperscaler capex revisions are the next confirmation point.</p><p><strong>Real-world adoption of the open-weights cluster, and migration patterns inside the closed-source pair.</strong> OpenRouter share is the cleanest leading indicator on the open-weights side. If GLM-5.1, Kimi K2.6, and MiMo V2.5 Pro accumulate share on production workloads, not just price-driven test traffic, the benchmark-optimization skepticism gets resolved one way. The companion signal sits inside the closed-source pair: if teams using Claude Code start switching to Codex on GPT-5.5 for execution work, the cost-per-task argument is being priced in real time. If migration stalls because the harness gap closes more slowly than the model gap, the harness layer just got more valuable.</p><p><strong>Mythos release timeline.</strong> A general-availability date for Mythos would be a meaningful re-rating event for both inference demand and capability competition. It would also be the cleanest counter evidence for the cost-per-task thesis, if Mythos&#8217;s capability lead is large enough to make the cost frame irrelevant in the workloads that matter most.</p><p><strong>Anthropic S-1 filing.</strong> Rumoured for October 2026. The single most informative document this thesis is waiting on. It will resolve the gross-versus-net revenue question, expose session economics, and put the managed-runtime moat on a balance sheet the market can price.</p><div><hr></div><p><em>This is part of an ongoing series on AI infrastructure economics. Earlier installments covered the bottleneck sequence, the memory crunch and the compression wave moving against it, and the model race as the demand engine behind the infrastructure thesis. This piece updates the demand engine read for the seven-release April cluster, and bounds the managed-runtime thesis against the open-weights wave.</em></p><p><em>For more on the AI stack and where value flows, visit <a href="https://www.theupcurious.com/">theupcurious.com</a>.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://upcurious.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Model Race Is Heating Up]]></title><description><![CDATA[What Anthropic vs. OpenAI means for enterprise software, compute demand, and how you work]]></description><link>https://upcurious.substack.com/p/the-model-race-is-heating-up</link><guid isPermaLink="false">https://upcurious.substack.com/p/the-model-race-is-heating-up</guid><dc:creator><![CDATA[Upcurious]]></dc:creator><pubDate>Thu, 16 Apr 2026 06:02:19 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/70ce00ad-9e9f-4df4-bb47-957e87f268ec_1200x634.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Anthropic reported $30 billion in annualised revenue, launched a fully managed agent infrastructure platform, and quietly distributed its most capable model to a restricted group of partners without releasing it to the public. OpenAI&#8217;s chief revenue officer responded with a four-page internal memo arguing Anthropic is gaining momentum in enterprise &#8212; and separately, Sam Altman confirmed that pre-training for OpenAI&#8217;s next frontier model finished on March 24, with a release expected within weeks.</p><p>More enterprise customers means more sustained inference load on the same constrained stack. The model race is the demand engine.</p><div><hr></div><h2>What the Revenue Numbers Actually Say</h2><p>Anthropic said on April 6 that its run-rate revenue had surpassed $30 billion in 2026 &#8212; the first time any rival has publicly claimed to match or surpass OpenAI&#8217;s run rate of $25 billion. There is a caveat.</p><p>OpenAI&#8217;s CRO pushed back in an internal memo, first reported by CNBC and later published in full by The Verge, arguing Anthropic&#8217;s number is inflated by roughly $8 billion. The accounting difference: Anthropic counts gross revenue from cloud partners, booking the full amount billed through AWS and Google Cloud before those platforms take their cut. OpenAI reports net of Microsoft&#8217;s share. Adjusted for that, the real comparison is closer to $22 billion versus $25 billion &#8212; OpenAI still ahead.</p><p>The accounting dispute matters less than the enterprise momentum in Ramp&#8217;s data. In Ramp&#8217;s April 15 AI Index update &#8212; built from corporate card and bill-pay flows across 50,000+ US businesses &#8212; Anthropic reached 30.6% of AI-paying businesses versus OpenAI&#8217;s 35.2%, cutting the gap to 4.6 percentage points. That is not a direct measure of token demand, but it is one of the clearest public indicators of enterprise customer momentum. OpenAI&#8217;s own CRO considered that momentum serious enough to write a four-page memo.</p><p>The revenue composition gap reinforces it. Anthropic&#8217;s mix skews heavily toward enterprise &#8212; CEO Dario Amodei has cited roughly 80% of revenue from business customers &#8212; while OpenAI carries a larger consumer share. Enterprise workloads run longer, burn more tokens per session, and sustain higher memory loads than chatbot interactions. As Anthropic gains enterprise share, the per-customer infrastructure demand rises with it.</p><div><hr></div><h2>What Anthropic Is Not Releasing &#8212; and Why</h2><p>On April 7, Anthropic announced it had built what it described as the most capable model it has ever created. Then it did not release it.</p><p>Anthropic&#8217;s Project Glasswing gives Mythos Preview to a restricted set of launch partners and additional critical-infrastructure organizations for defensive cybersecurity work, underscoring both the model&#8217;s capability and Anthropic&#8217;s decision not to release it broadly. Twelve named launch partners &#8212; including AWS, Apple, Google, Microsoft, and NVIDIA &#8212; received access, alongside more than 40 additional organisations that build or maintain critical software infrastructure.</p><p>The benchmark numbers explain the caution. Mythos leads 17 of 18 benchmarks Anthropic measured. On SWE-Bench Verified &#8212; real-world software engineering &#8212; it scores 93.9%, up from 80.8% for the prior model. A 13-point gain in a domain where 2-3 points was previously considered meaningful progress. On CyberGym, the cybersecurity capability benchmark, it scores 83.1%. In one test, it developed 181 working browser exploits from vulnerabilities it autonomously identified, versus 2 for the prior model.</p><p>That last number is the reason for restricted release, stated plainly. The same capability that finds vulnerabilities at scale can be redirected to weaponise them.</p><p>The strategic implication is less obvious. Anthropic is holding a capability tier above what is publicly available &#8212; deliberately, with named enterprise partners, as a demonstration of safety discipline. This is not just caution. It is positioning. The enterprises deploying AI in regulated, high-stakes environments have a procurement story to tell their boards. Anthropic is building that story one announcement at a time.</p><p>Whether Mythos eventually reaches general availability is not confirmed. What the restricted access does establish is a demand signal: when a model at this capability level deploys at broader scale, inference load goes up again. That timeline is unknown, but the direction is not.</p><div><hr></div><h2>The Inference Supercycle</h2><p>In 2023, training dominated AI compute. Building the models consumed most of the resources. Running them was secondary.</p><p>That has reversed. Inference now accounts for roughly two-thirds of all AI compute in 2026, up from one-third in 2023 and half in 2025, according to Deloitte&#8217;s 2026 TMT Predictions report. The shift is structural. Enterprises are not experimenting anymore &#8212; they are running AI in production, continuously, at scale.</p><p>Claude Managed Agents, launched in public beta on April 8, moves Anthropic further from selling model access alone toward operating agent infrastructure on behalf of customers. Define the agent in natural language or YAML (a configuration file), set the guardrails, and Anthropic handles hosting, scaling, and monitoring. Sessions are persistent. Context is maintained across the full task lifecycle. Runtime is billed at $0.08 per session hour, on top of standard token costs.</p><p>What Managed Agents actually represents is a shift in the unit of compute. A chatbot interaction is a round trip &#8212; a few seconds, a few thousand tokens, done. A managed agent session working through a codebase or a complex document runs for minutes or hours, consuming 100K to 500K tokens before it finishes. Previously, deploying that kind of session in production required your team to build and operate the infrastructure. Anthropic is now absorbing that entirely. The compute demand per customer goes up. The barrier to creating it just dropped by an order of magnitude.</p><p>Jevons paradox is already visible &#8212; the idea that making something cheaper and easier to use tends to grow total demand, not shrink it. Managed Agents reduces the fixed engineering cost and operational friction of running an agent workflow. The runtime fee is only part of the bill; the bigger change is that Anthropic is absorbing the sandboxing, orchestration, session management, and monitoring work that previously took teams months to build. When agent infrastructure gets easier to deploy, more companies deploy agents. When more companies deploy agents, total token volume rises. That is the pattern that has defined every prior technology supercycle &#8212; efficiency enabling more usage, not less.</p><div><hr></div><h2>What This Means for How You Work</h2><p>The practitioner shift is already visible in the tooling. Ninety days ago, AI coding tools mostly behaved like assistants: you wrote the prompt, they returned a draft, and you reviewed it. Now the workflow is increasingly supervisory. Tools like Claude Code can break work into subtasks, operate over longer sessions, and come back with code, tests, and follow-up actions. The job is moving from writing better prompts to defining the task, setting constraints, and reviewing the result.</p><p>That changes the economics as well as the workflow. A short assistant interaction might use a few thousand tokens; an agentic session working through a real codebase can run far longer and consume materially more compute before it is done. Managed runtimes extend that pattern beyond code into research, operations, and back-office workflows. The human is still in the loop, but later in the process: less step-by-step execution, more specification and review. From the infrastructure side, that means more compute consumed per completed task, not less.</p><div><hr></div><h2>Where Value Accretes in the Stack</h2><p>The model race has one clear beneficiary at the infrastructure layer: NVIDIA. More inference sessions mean more GPU utilisation. Longer, persistent sessions mean more HBM (high-bandwidth memory) demand per GPU. Neither Anthropic nor OpenAI is building infrastructure that reduces demand for Blackwell or Vera Rubin. They are filling it.</p><p>The underlying supply picture hasn&#8217;t changed &#8212; HBM is sold out through end of 2026, CoWoS (TSMC&#8217;s advanced chip packaging) capacity is dominated by NVIDIA reservations, and neither constraint eases because the demand driver changed form. Managed Agents is a new demand signal layered on top of an already constrained stack. Watch for capex guidance in Big Tech earnings starting April 22 &#8212; any upward revision in inference infrastructure spend is the first public confirmation of the agent-driven shift.</p><p>The more interesting question is what happens in the middle. Enterprise software companies that built their value proposition on &#8220;we&#8217;ll handle the AI complexity for you&#8221; face a different competitive environment when Anthropic is running the agents directly. Notion is an early Managed Agents adopter &#8212; which means Notion is now a customer of Anthropic&#8217;s infrastructure layer, not just its models. The orchestration, session management, and monitoring that middleware vendors and integration platforms previously sold as differentiation is now a platform feature at $0.08 per hour.</p><p>The vertical integration runs in both directions. As the labs move down toward infrastructure, the value of the application layer becomes harder to defend on anything except proprietary data, workflow depth, or distribution. Building on top of a lab that is also building your product&#8217;s core infrastructure is a different strategic position than it was eighteen months ago. Companies that assumed the frontier labs would stay in their lane are finding that assumption was wrong.</p><div><hr></div><h2>What Could Break This</h2><p>The inference demand thesis weakens if efficiency compounds faster than usage.</p><p>Managed Agents supports built-in prompt caching, compaction, and other context optimizations. If those systems deliver high cache hit rates or reduce repeated long-context reads, the effective compute footprint per workflow could come in below the raw token math. Anthropic has not published workload composition or realized cache-hit data. Until it does, this is an unknown variable with real weight.</p><p>The enterprise crossover story deflates if OpenAI&#8217;s push accelerates faster than Anthropic can convert. OpenAI is already at 40% enterprise revenue, up from 30% a year ago, and is not standing still. The Ramp data shows Anthropic momentum, not a completed crossover.</p><p>Open-weights models are the structural floor under both. Meta&#8217;s Llama 4 supports a ten-million token context window and runs on commodity infrastructure. If open-source frontier models reach enterprise-grade reliability, enterprises running their own inference infrastructure do not add demand to the constrained GPU clusters that Anthropic and OpenAI are filling. They move the load elsewhere &#8212; and the managed inference moat both companies are building becomes harder to price above cost.</p><div><hr></div><h2>What We Are Watching</h2><p><strong>Enterprise customer share crossover.</strong> The Ramp AI Index has Anthropic closing fast &#8212; from 24.4% to 30.6% in the latest monthly update. The next Ramp release will either confirm the trajectory or show it stalling.</p><p><strong>Managed Agents adoption velocity.</strong> The list of enterprise names announcing integrations over the next 90 days will indicate whether the managed-runtime model is clearing the market.</p><p><strong>Big Tech earnings, starting April 22.</strong> TSMC&#8217;s Q1 results confirmed 35% revenue growth and CoWoS capacity dominated by NVIDIA reservations. The question for Meta, Google, Microsoft, and Amazon is whether capex guidance reflects agent-driven inference load. Any upward revision in inference infrastructure spend is direct confirmation of the demand signal we are tracking.</p><p><strong>Next frontier models: Mythos and Spud.</strong> Mythos Preview is now accessible through Amazon Bedrock as a gated research preview &#8212; allowlisted organisations, cybersecurity use cases only, US East region. That is a narrower channel than general availability, but it is no longer purely internal. Watch how quickly the allowlist expands and whether use cases beyond cybersecurity get cleared. On the OpenAI side, Spud remains a rumour with no public benchmarks or confirmed release date, despite pre-training completing on March 24. Altman described it as a model that could &#8220;really accelerate the economy.&#8221; If Spud launches with a step-change in capability, it raises the inference demand ceiling and resets the competitive framing &#8212; in both directions.</p><p><strong>IPO filings.</strong> Anthropic is rumoured for October 2026. The S-1 (IPO prospectus) will resolve the gross versus net revenue question definitively, and will be the first public view of actual token consumption, session economics, and infrastructure cost structure. That filing will contain more useful data about the inference supercycle than any estimate produced before it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://upcurious.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><p><em>This is part of an ongoing series on AI infrastructure economics. Part 1 covered the full bottleneck sequence &#8212; power, packaging, fabs, and EUV. Part 2 covered the memory crunch and the compression wave building against it.</em></p><p><em>For more on the AI stack and where value flows, visit <a href="https://www.theupcurious.com/">theupcurious.com</a>.</em></p><div><hr></div><h2>Sources</h2><p><strong>Revenue and enterprise share</strong></p><ul><li><p>Anthropic $30B ARR (Google/Broadcom compute partnership announcement, April 2026): <a href="https://www.anthropic.com/news/google-broadcom-partnership-compute">https://www.anthropic.com/news/google-broadcom-partnership-compute</a></p></li><li><p>OpenAI CRO internal memo: The Verge &#8212; <a href="https://www.theverge.com/ai-artificial-intelligence/911118/openai-memo-cro-ai-competition-anthropic">https://www.theverge.com/ai-artificial-intelligence/911118/openai-memo-cro-ai-competition-anthropic</a></p></li><li><p>Ramp AI Index April 2026 &#8212; &#8220;As AI adoption crosses 50%, the tokenmaxxing economy splits off and up&#8221; (Anthropic 30.6% / OpenAI 35.2%): <a href="https://ramp.com/leading-indicators/the-tokenmaxxing-economy-splits-off-and-up">https://ramp.com/leading-indicators/the-tokenmaxxing-economy-splits-off-and-up</a></p></li></ul><p><strong>Mythos and Project Glasswing</strong></p><ul><li><p>Project Glasswing launch (April 7, 2026): <a href="https://www.anthropic.com/glasswing">https://www.anthropic.com/glasswing</a></p></li><li><p>Mythos Preview benchmarks and technical write-up: <a href="https://red.anthropic.com/2026/mythos-preview/">https://red.anthropic.com/2026/mythos-preview/</a></p></li><li><p>Mythos Preview on Amazon Bedrock (gated research preview): <a href="https://aws.amazon.com/about-aws/whats-new/2026/04/amazon-bedrock-claude-mythos/">https://aws.amazon.com/about-aws/whats-new/2026/04/amazon-bedrock-claude-mythos/</a></p></li></ul><p><strong>OpenAI Spud</strong></p><ul><li><p>Leaked memo and Altman pre-training confirmation: <a href="https://the-decoder.com/openais-leaked-memo-says-new-spud-model-will-make-all-its-products-significantly-better/">https://the-decoder.com/openais-leaked-memo-says-new-spud-model-will-make-all-its-products-significantly-better/</a></p></li></ul><p><strong>Inference supercycle</strong></p><ul><li><p>Inference = 2/3 of AI compute: Deloitte 2026 TMT Predictions &#8212; <a href="https://www.deloitte.com/us/en/insights/industry/technology/technology-media-and-telecom-predictions/2026/compute-power-ai.html">https://www.deloitte.com/us/en/insights/industry/technology/technology-media-and-telecom-predictions/2026/compute-power-ai.html</a></p></li><li><p>Computerworld CES 2026 supporting coverage: <a href="https://www.computerworld.com/article/4114579/ces-2026-ai-compute-sees-a-shift-from-training-to-inference.html">https://www.computerworld.com/article/4114579/ces-2026-ai-compute-sees-a-shift-from-training-to-inference.html</a></p></li><li><p>Claude Managed Agents overview: <a href="https://platform.claude.com/docs/en/managed-agents/overview">https://platform.claude.com/docs/en/managed-agents/overview</a></p></li><li><p>Claude Managed Agents pricing: <a href="https://platform.claude.com/docs/en/about-claude/pricing">https://platform.claude.com/docs/en/about-claude/pricing</a></p></li></ul><p><strong>Infrastructure</strong></p><ul><li><p>SK Hynix HBM sold out through 2026: <a href="https://www.techspot.com/news/110058-sk-hynix-completely-sells-out-semiconductor-supply-ai.html">https://www.techspot.com/news/110058-sk-hynix-completely-sells-out-semiconductor-supply-ai.html</a></p></li><li><p>Micron HBM sell-out confirmation: <a href="https://www.trendforce.com/news/2025/08/13/news-hbm-battle-heats-up-micron-reportedly-hints-2026-sell-out-sk-hynix-yet-to-confirm/">https://www.trendforce.com/news/2025/08/13/news-hbm-battle-heats-up-micron-reportedly-hints-2026-sell-out-sk-hynix-yet-to-confirm/</a></p></li><li><p>TSMC CoWoS / NVIDIA capacity dominance: <a href="https://www.digitimes.com/news/a20260410VL204/packaging-capacity-tsmc-nvidia-demand.html">https://www.digitimes.com/news/a20260410VL204/packaging-capacity-tsmc-nvidia-demand.html</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[The Quiet Force Reshaping AI Infrastructure]]></title><description><![CDATA[Why the most important cost driver in AI isn't compute &#8212; and what happens when it gets compressed]]></description><link>https://upcurious.substack.com/p/the-quiet-force-reshaping-ai-infrastructure</link><guid isPermaLink="false">https://upcurious.substack.com/p/the-quiet-force-reshaping-ai-infrastructure</guid><dc:creator><![CDATA[Upcurious]]></dc:creator><pubDate>Thu, 26 Mar 2026 04:45:16 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4a6bfa55-d829-4832-a6f7-3cdfd3b76bd1_2162x1158.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Most AI infrastructure conversations revolve around compute: how many H100s a hyperscaler has ordered, what NVIDIA&#8217;s next chip roadmap looks like, how many exaflops a data centre can deliver. This framing is understandable but incomplete. As AI models get deployed at scale for longer and longer conversations, a different bottleneck is quietly becoming the dominant cost driver in production inference: memory.</p><p>Specifically, a component called the KV cache &#8212; and the expensive, constrained chip technology required to hold it.</p><p>The memory crunch is no longer a niche concern. Micron just reported the strongest quarter in its history. HBM is sold out through 2026 across all three producers. Lead times are years, and the supply side cannot catch up. Everyone knows this part of the story. The harder question is why demand broke so far past what anyone modeled &#8212; and whether the efficiency wave building against it will move fast enough to change the investment math.</p><div><hr></div><h2>The Demand Math Broke</h2><p>The memory demand was based on chatbots. A user sends a message. The model responds. A few thousand tokens. The GPU moves on. Under that model, memory was a commodity input &#8212; important but predictable.</p><p>Two things broke that assumption simultaneously.</p><p><strong>The KV cache explosion.</strong> Every transformer stores intermediate results &#8212; key and value matrices &#8212; for every token in the conversation. Think of it as the model keeping a detailed notepad of everything it has already read, so it only needs to process what is genuinely new. This KV cache lives in HBM, the high-bandwidth memory stacked on the GPU, and it must stay there for the entire session.</p><p>The size scales linearly with context length. And context windows have gone from 8K to 200K to 1M in three years. Here is what that does to a single session on a large model:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gd63!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gd63!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png 424w, https://substackcdn.com/image/fetch/$s_!gd63!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png 848w, https://substackcdn.com/image/fetch/$s_!gd63!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png 1272w, https://substackcdn.com/image/fetch/$s_!gd63!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gd63!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png" width="1456" height="640" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:640,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1993034,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://upcurious.substack.com/i/192170383?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gd63!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png 424w, https://substackcdn.com/image/fetch/$s_!gd63!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png 848w, https://substackcdn.com/image/fetch/$s_!gd63!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png 1272w, https://substackcdn.com/image/fetch/$s_!gd63!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67cb9ca7-ca0b-448a-8a2d-02a19e6f529c_2198x966.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The industry is building toward the bottom of that table. Every row down doubles the memory bill.</p><p><strong>Agents changed the unit economics.</strong> A chatbot interaction costs under 2,000 tokens. An agentic task &#8212; reading a codebase, writing code, running tests, debugging, iterating &#8212; costs 100,000 to 500,000 tokens compressed into a single growing context, with the model reasoning internally at every step. The model sees the full history on every forward pass. Run two agents in parallel and you hold two full KV caches in HBM simultaneously. The multiplier versus chatbot economics is not 5x. It is 50 to 100x per unit of user-facing work.</p><p>This is not hypothetical. OpenClaw &#8212; an autonomous agent that taps multiple models including ChatGPT, Gemini, and Claude, running them continuously and in parallel &#8212; has become the hottest productivity trend around the world. OpenRouter data shows weekly token consumption jumping from 1.62 trillion in March 2025 to 20.4 trillion in March 2026. A 12x increase in one year. A human programmer might use several million tokens a day. OpenClaw can burn through billions, with costs reaching thousands of dollars per session.</p><p>Jensen Huang made the implication explicit at GTC 2026 on March 16: NVIDIA is reviewing a plan to provide engineers with token allocations equal to half their base salary &#8212; roughly $187,500 per engineer per year in token resources. When the CEO of the company that makes the GPUs starts treating tokens as compensation, the demand curve has moved past &#8220;early adoption.&#8221;</p><p>The infrastructure models everyone built before 2024 assumed chatbot economics. The agent era broke that assumption, and nothing about the trajectory suggests it is slowing down.</p><div><hr></div><h2>Micron Just Showed You What a Supply Ceiling Looks Like</h2><p>We covered Micron&#8217;s Q2 FY2026 results in detail last time. The short version: flat volumes, exploding prices. DRAM prices up mid-60s percent sequentially while bit shipments grew only mid-single digits. That is what a physical ceiling looks like &#8212; producers cannot ship more, so every dollar of incremental demand shows up in price.</p><p>The supply picture has not changed. HBM is sold out through 2026 across all three producers. Micron guided Q3 revenue to $33.5 billion at 81% gross margins, and raised FY2026 capex above $25 billion &#8212; money that produces wafers in 2027 and 2028, not now. The hyperscalers have pre-bought the available supply: Microsoft&#8217;s 2026 capex is projected above $120 billion, and the Stargate joint venture &#8212; backed by SoftBank, OpenAI, Oracle, and others &#8212; is a $500 billion, four-year commitment.</p><p>The supply crunch is real. But the cause is not primarily a chip story. It is a software story &#8212; agents, reasoning, long context &#8212; that shows up in the hardware. Micron&#8217;s income statement is the proof.</p><div><hr></div><h2>Now the Software Is Attacking the Problem It Created</h2><p>Here is where it gets interesting.</p><p>On March 24, Google Research published TurboQuant. If you only read one AI paper this quarter, make it this one.</p><p>The KV cache &#8212; the notepad that consumes all that expensive HBM &#8212; is traditionally stored at 16-bit floating point precision. TurboQuant compresses it to 3 bits per value, with no retraining or fine-tuning required. That is a roughly 6x reduction in KV cache size. The 70B model that needed 305 GiB of KV cache at 1 million tokens? Now it needs about 50 GiB. A workload that required six H100s can potentially be served on two.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ClDv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ClDv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png 424w, https://substackcdn.com/image/fetch/$s_!ClDv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png 848w, https://substackcdn.com/image/fetch/$s_!ClDv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png 1272w, https://substackcdn.com/image/fetch/$s_!ClDv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ClDv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png" width="1456" height="650" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:650,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2116115,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://upcurious.substack.com/i/192170383?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ClDv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png 424w, https://substackcdn.com/image/fetch/$s_!ClDv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png 848w, https://substackcdn.com/image/fetch/$s_!ClDv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png 1272w, https://substackcdn.com/image/fetch/$s_!ClDv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4ebbbac-5551-4d51-afc2-1cb85cf43ba5_2234x998.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The mechanism is elegant. TurboQuant applies a random rotation to the data vectors (a technique called PolarQuant), which simplifies the geometry enough to apply aggressive quantization. Then it uses a 1-bit error correction algorithm (QJL) to eliminate the bias that aggressive compression normally introduces. The result: 3-bit KV cache that scores identically to full precision across every standard benchmark Google tested &#8212; LongBench, Needle in a Haystack, ZeroSCROLLS, RULER. No quality loss. No retraining. Drop-in deployment.</p><p>Google tested it on Gemma and Mistral. At 4 bits, TurboQuant achieves up to 8x speedup in computing attention on H100 GPUs versus unquantized keys. The paper will be presented at ICLR 2026.</p><p>The &#8220;training-free&#8221; property is what makes this different from prior KV compression work. Most techniques &#8212; KIVI, KVQuant &#8212; require model fine-tuning or custom infrastructure. TurboQuant applies at inference time. Any model that already exists can benefit. That means adoption can move fast.</p><div><hr></div><h2>TurboQuant Is Part of a Larger Wave</h2><p>TurboQuant matters. But zoom out. It is one of three vectors attacking the same problem, and they compound rather than add.</p><p><strong>Quantization compresses the bytes.</strong> TurboQuant, KIVI, KVQuant, NVIDIA&#8217;s NVFP4 &#8212; all reduce how many bits each cached value occupies in HBM. This directly shrinks capacity demand.</p><p><strong>Sparse attention reduces the reads.</strong> Methods like DeepSeek&#8217;s NSA architecture and SparQ Attention do not shrink the cache &#8212; they reduce how much of it the model reads on each generation step. In long-context inference, the bottleneck is often bandwidth, not capacity. Sparse attention can cut data read per step by up to 8x.</p><p><strong>CXL tiering moves cold data off-chip.</strong> CXL lets accelerators address large pools of external DRAM at low latency. KV cache that is not actively needed can spill from expensive HBM to cheap rack-level memory. Only the recent tokens stay on-device.</p><p>Combine all three: quantize the cache to 3 bits, read only the relevant fraction per step, and offload the cold tail to CXL memory. The maths gets aggressive fast. And none of these techniques existed in production two years ago.</p><div><hr></div><h2>The Market Is Already Repricing This</h2><p>The tension between demand and compression is not theoretical. The market started pricing it in the week TurboQuant dropped.</p><p>Micron hit an all-time high of $471 on March 18, the day it reported the strongest quarter in its history. By March 25 &#8212; the day after Google published TurboQuant &#8212; the stock had fallen to roughly $382. A 19% drawdown from peak, despite record revenue, record margins, and sold-out HBM through 2026. Western Digital dropped 4.7%. SanDisk fell 5.7%. Samsung and SK Hynix weakened at the same time. Meanwhile, CPU stocks &#8212; Intel, AMD &#8212; surged as investors recalculated whether &#8220;insatiable&#8221; HBM demand might be tempered by algorithmic efficiency.</p><p>The sell-the-news reaction after earnings was predictable. The acceleration after TurboQuant was not. The market is telling you it heard the compression thesis &#8212; and it moved before most analysts had read the paper.</p><p>The standard memory cycle ends when demand cools or supply catches up. What makes this cycle different is that both sides are moving simultaneously &#8212; and in tension.</p><p>On the demand side, the forces are structural, not cyclical. KV cache requirements grow with every context window expansion, and context windows keep expanding. Agents consume more tokens as the technology matures, not fewer &#8212; longer loops, larger codebases, more parallel sessions. These are new floors that compound.</p><p>On the compression side, TurboQuant-class innovations are also structural. Training-free, drop-in deployment means adoption can move faster than traditional semiconductor cycles. And the compression vectors compound with each other.</p><p>So the question becomes: which force wins, and when does it flip?</p><p><strong>Through 2026&#8211;2027</strong>, the demand side wins. Supply is already sold out. Compression techniques take time to deploy at scale across production inference stacks. Micron trades at roughly 10x forward earnings &#8212; a multiple that prices a cyclical peak, not a supercycle. The memory names still have the better near-term visibility.</p><p><strong>From 2028 onwards</strong>, the compression side gets heavier. If TurboQuant-class techniques deploy broadly and CXL tiering becomes production-standard, the per-token HBM demand equation shifts. Total HBM demand probably still grows &#8212; AI workloads are still compounding &#8212; but the growth rate decelerates. SK Hynix&#8217;s pricing power compresses. Samsung, which has lagged on HBM yield and quality, gets hit at the worst time.</p><div><hr></div><h2>So Is This a Buy?</h2><p>The memory crunch thesis is not broken. But it is no longer open-ended.</p><p>Start with what has not changed. HBM is sold out through 2026. Micron just guided $33.5 billion in Q3 revenue with 81% gross margins. The hyperscalers have pre-committed hundreds of billions in infrastructure capex against agentic workloads that are still ramping. TurboQuant is a research paper presented at a conference &#8212; it is not yet running in production inference at any major provider. The gap between &#8220;published&#8221; and &#8220;deployed at scale across Google, Microsoft, Amazon, and Meta&#8217;s inference fleets&#8221; is measured in quarters, probably years.</p><p>Micron at ~$382, down 19% from its all-time high with the strongest earnings in its history behind it and sold-out supply in front of it, looks like a buy on a 12-month view. The market is pricing TurboQuant&#8217;s impact as if it will compress HBM demand tomorrow. It will not. The production deployment timeline gives the memory names at least four to six more quarters of pricing power before compression can realistically dent contract economics.</p><p>he compression wave introduces a time dimension that did not exist six months ago. The near-term supply picture is clear; the 2028 picture is not. Trade accordingly.</p><p>Here is how I think about Micron at this junction:</p><p><strong>Micron (MU)</strong> &#8212; a tactical trade, not a thesis reset. The fundamentals have not deteriorated: 10x forward earnings, Q3 guidance implying acceleration, sold-out supply through 2026. The 19% drawdown looks like the market front-running a risk that is real but not imminent. The compression wave will matter &#8212; just not in the next four to six quarters. That window is the trade. The risk is that the re-rating sticks regardless of earnings, and the multiple stays compressed even as the business keeps growing. Size accordingly.</p><p><strong>The names that win regardless.</strong> The more interesting positioning might be the companies that benefit whether demand or compression dominates:</p><p><strong>Teradyne (TER)</strong> &#8212; semiconductor test equipment. AI chips require ~10x more test time per die than prior generations. HBM4 is a substantially more complex package. Whether HBM demand keeps surging or the industry transitions to more complex, lower-volume HBM at higher ASPs, test intensity per wafer goes up. Teradyne&#8217;s demand is driven by complexity, not volume alone.</p><p><strong>NVIDIA</strong> &#8212; the quiet winner of this entire dynamic. Memory efficiency increases GPU utilisation. A more efficient model runs at higher batch sizes and larger contexts on the same hardware. Compression opens markets that were previously too expensive to serve. NVIDIA sells the compute. Cheaper memory makes that compute more useful, not less.</p><p><strong>Broadcom and Marvell</strong> &#8212; the CXL plays. If CXL tiered memory becomes the standard way to handle KV cache overflow, both companies benefit from switching fabric, interconnect, and custom silicon. CXL grows faster if HBM supply stays constrained, because CXL is part of the solution. This is a hedge that does not require the memory crunch to resolve in either direction.</p><div><hr></div><h2>What Could Break This</h2><p>The demand thesis weakens if:</p><ul><li><p><strong>KV cache compression deploys faster than expected.</strong> TurboQuant is training-free and drop-in. If the major inference providers adopt it within quarters rather than years, the per-session HBM demand curve flattens sooner. Watch for API pricing changes at the major labs &#8212; falling per-token costs at long context lengths would be the leading signal.</p></li><li><p><strong>Agent token efficiency improves.</strong> The 12x token growth from OpenRouter is real &#8212; but a meaningful share of it is architectural slop. OpenClaw sends full conversation history on every request, triggers four to five background API calls per visible response, and pulls unranked context that forces the model to process thousands of tokens it does not need. Engineers are already cutting bills by 90% with basic configuration changes. Prompt caching, model routing, and smarter context management could eliminate 60&#8211;80% of current agent token spend without sacrificing quality. The counterargument is Jevons paradox &#8212; cheaper tokens historically mean more usage, not less. But the efficiency headroom is large enough to temper any straight-line extrapolation of current demand.</p></li><li><p><strong>Alternative architectures gain traction.</strong> Groq&#8217;s SRAM-based LPU design sidesteps HBM entirely for inference. If architectures like this reach commercial scale, they structurally reduce HBM addressable market.</p></li><li><p><strong>DRAM pricing plateaus in Q2&#8211;Q3 2026.</strong> If memory prices stop climbing despite supply remaining tight, it signals demand absorption is slowing. This is the simplest and most direct disconfirming indicator.</p></li></ul><p>The compression thesis weakens if:</p><ul><li><p><strong>Hyperscalers reinvest savings into longer contexts rather than banking them.</strong> This is what has historically happened &#8212; efficiency gains get consumed by expanded capability, not passed through as cost reduction. Google is already testing 10 million token contexts. Llama 4 Scout supports 10 million tokens. If the ceiling keeps rising, compression just enables more demand rather than reducing it.</p></li><li><p><strong>Quality degradation at extreme compression proves worse in production than in benchmarks.</strong> TurboQuant reports zero quality loss at 3 bits in controlled tests. Production workloads at scale, with adversarial edge cases and long tails, may be less forgiving.</p></li></ul><div><hr></div><h2>What We Are Watching</h2><p><strong>Agentic token consumption.</strong> The number to watch is what share of total tokens are consumed by agentic workloads versus simple chat. No lab has published that breakdown yet. When agents cross 50% of token volume at a major provider, the per-session memory profile of the entire fleet shifts dramatically. The first lab to disclose workload composition data will be a significant data point.</p><p><strong>TurboQuant adoption velocity.</strong> The paper is at ICLR 2026. The question is how fast it moves from research to production inference at Google, and whether NVIDIA integrates similar techniques into TensorRT. If NVIDIA ships native KV cache quantization in its inference stack, adoption accelerates across the entire GPU ecosystem.</p><p><strong>DRAM spot pricing &#8212; already flattening.</strong> This one is worth watching closely. InSpectrum spot data through late March 2026 shows DDR4 16Gb spot prices went near-vertical from October to December 2025 &#8212; roughly $5 to $70 &#8212; then flatlined from January through March. DDR5 shows the same pattern: stepped up through late 2025, sideways since January at around $32. Three months of zero price momentum after a parabolic move. NAND is lagging by about a quarter &#8212; still stepping up, but the slope is flattening. Meanwhile, Micron&#8217;s contract prices are still surging (mid-60s% sequential DRAM increases in Q2). That divergence between flat spot and rising contracts is the kind of signal that resolves in one of two directions: either spot is a leading indicator and contract pricing follows down with a lag, or spot is simply too thin to matter because the hyperscalers have pre-bought everything on contract. If spot stays flat through Q2 while contract re-rates lower in Q3, the demand-compression balance may be shifting sooner than the bull case assumes. <em>(Source: inSpectrum Tech Inc DRAM/NAND spot pricing, March 2026)</em></p><p><strong>CXL deployment timelines.</strong> Production CXL memory tiering at hyperscaler scale would be a structural shift in how KV cache is served. Broadcom and Marvell benefit directly. HBM-per-GPU requirements decline.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://upcurious.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>This is part of an ongoing series on AI infrastructure economics. The previous piece covered the full bottleneck sequence: power, packaging, fabs, and EUV</em></p><p><em>For more on the AI stack and where value flows, visit <a href="https://www.theupcurious.com/">theupcurious.com</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Which AI Bottleneck Matters Most Right Now?]]></title><description><![CDATA[The most important question in AI investing right now is not who has the best model. It is what the system can physically support]]></description><link>https://upcurious.substack.com/p/which-ai-bottleneck-matters-most</link><guid isPermaLink="false">https://upcurious.substack.com/p/which-ai-bottleneck-matters-most</guid><dc:creator><![CDATA[Upcurious]]></dc:creator><pubDate>Wed, 18 Mar 2026 07:30:55 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/55e59c49-d8c6-4e93-8f1b-f22048204053_1424x752.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The most important question in AI investing right now is not who has the best model. It is what the system can physically support &#8212; and which constraint is most underpriced.</p><p>Dylan Patel, founder of SemiAnalysis, lays out this framework on the Dwarkesh Podcast. His argument: AI does not have one bottleneck. It has a sequence of them &#8212; and the important question is always which one matters most right now, which one is coming next, and which one sets the outer limit on the whole system.</p><p>This is still an infrastructure story. But in Patel&#8217;s framing, it is becoming more of a semiconductor manufacturing story than a construction story.</p><div><hr></div><h2>The Scale of the Buildout</h2><p>Hyperscaler capex is running at $660-750B in 2026 &#8212; up 67% year-over-year. Cloud revenue growth is 25-30%. The capex-to-revenue ratio has hit 45-57%, which is historically unsustainable. Microsoft reports $80B in Azure backlog. The market keeps asking whether capex will slow down. So far, the answer is no.</p><p>The demand side is not in question. OpenAI hit $25B in annualized revenue (ARR) by February 2026, up from $6B in 2024. Anthropic surged to $19B ARR, up from $9B at the end of 2025. These are the fastest-scaling software businesses in history, and they are the primary drivers of compute demand going forward.</p><p>That is what makes Patel&#8217;s framework useful. The question is not whether AI demand is strong. It is which part of the system is failing to keep up with that demand &#8212; and who owns the bottleneck.</p><div><hr></div><h2>The Bottleneck Sequence</h2><h3>Now: The Deployment Constraints</h3><p>These are the bottlenecks the market is focused on: CoWoS packaging, data center construction, power availability. They are serious &#8212; TSMC&#8217;s CoWoS capacity remains sold out through 2026, Vertiv&#8217;s organic orders surged 252% in Q4 2025. But Patel&#8217;s argument is that the biggest bottleneck is already shifting past them.</p><p>His point is that these sit higher in the stack, have simpler supply chains, and can be expanded with enough capital and execution. That does not mean they are easy, and it does not mean they stop being bottlenecks. It means he views them as more tractable than semiconductor manufacturing itself.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KZ-i!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KZ-i!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png 424w, https://substackcdn.com/image/fetch/$s_!KZ-i!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png 848w, https://substackcdn.com/image/fetch/$s_!KZ-i!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png 1272w, https://substackcdn.com/image/fetch/$s_!KZ-i!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KZ-i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png" width="1456" height="671" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:671,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:860603,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://upcuriousinvest.substack.com/i/191330543?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!KZ-i!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png 424w, https://substackcdn.com/image/fetch/$s_!KZ-i!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png 848w, https://substackcdn.com/image/fetch/$s_!KZ-i!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png 1272w, https://substackcdn.com/image/fetch/$s_!KZ-i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f75736-ca51-413f-bdba-87af9fcdbfce_1527x704.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Data centers can be built in less than a year. Power &#8212; where Patel parts company with the &#8220;AI runs out of watts&#8221; crowd &#8212; can be sourced flexibly through industrial gas turbines, solar, nuclear, grid expansion, even ship engines. GE Vernova booked $2B+ in data center electrification orders in 2025 (3x the prior year) and has only addressed roughly 10% of its addressable market. Constellation Energy has nuclear PPAs contracted directly with hyperscalers. GPUs are sitting idle due to power constraints, not demand constraints. These are real bottlenecks with real pricing power &#8212; but in Patel&#8217;s framing, they are deployment constraints, not structural ceilings. Difficult but addressable with enough capital, risk-taking, and engineering effort.</p><p><strong>The trade:</strong> Two clusters, different risk profiles.</p><p>The packaging and inspection names &#8212; TSMC (TSM), KLA (KLAC), Camtek (CAMT) &#8212; own the CoWoS chokepoint directly. TSMC controls roughly 60% of advanced packaging capacity and is embedded in every leading GPU shipment. KLA and Camtek sit on the yield management side: as CoWoS complexity increases, inspection becomes more critical and more recurring. These are not cyclical equipment names &#8212; they are process-critical infrastructure with multi-year demand visibility.</p><p>The power and electrical names &#8212; GE Vernova (GEV), Eaton (ETN), Vertiv (VRT), Constellation Energy (CEG) &#8212; own the constraint between &#8220;GPU purchased&#8221; and &#8220;GPU producing revenue.&#8221; Vertiv&#8217;s $15B backlog is up 109% year-over-year. Eaton&#8217;s backlog equals 11 years of 2025-level construction. The market still prices some of these like industrials. The backlog data says otherwise.</p><p>The honest risk: Patel views these deployment constraints as solvable. As power and packaging bottlenecks ease, the market will rotate toward the deeper constraints. These names have durable cash flows for now, but the long-term compounding story sits further down the stack.</p><h3>Next: Fabs, Clean Rooms, and the Memory Crunch</h3><p>The system has shifted from deployment chokepoints into something harder: physical semiconductor capacity. Fabs are the most complex buildings humans make, with lead times of two to three years. And the industry underbuilt when memory economics were weak &#8212; vendors halted new fab construction in 2023 when they were losing money. Micron&#8217;s own supply roadmap shows how long the lag really is: first wafer output from Idaho is expected in mid-calendar 2027, the Tongluo site supports meaningful shipments in fiscal 2028, Singapore HBM packaging is expected to contribute meaningfully in calendar 2027 (per prior guidance), and a new Singapore NAND fab is not expected to produce wafers until the second half of 2028.</p><p>This is what is actually driving the memory crunch. It is not just that HBM is tight. There is literally nowhere to place the manufacturing tools needed to scale production.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vlf3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vlf3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png 424w, https://substackcdn.com/image/fetch/$s_!vlf3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png 848w, https://substackcdn.com/image/fetch/$s_!vlf3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png 1272w, https://substackcdn.com/image/fetch/$s_!vlf3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vlf3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png" width="1456" height="685" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:685,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2685647,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://upcuriousinvest.substack.com/i/191330543?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!vlf3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png 424w, https://substackcdn.com/image/fetch/$s_!vlf3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png 848w, https://substackcdn.com/image/fetch/$s_!vlf3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png 1272w, https://substackcdn.com/image/fetch/$s_!vlf3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e9e21af-9a63-4858-b0b3-f5d3b2147a2a_2042x960.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>One gigawatt of Rubin-scale AI compute requires approximately 170,000 DRAM wafers (Memory), plus roughly 55,000 wafers of 3nm logic (GPU). Memory is not riding alongside the GPU story. It is the GPU story &#8212; just measured in a different unit. And the supply side is an oligopoly:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mjUb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mjUb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png 424w, https://substackcdn.com/image/fetch/$s_!mjUb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png 848w, https://substackcdn.com/image/fetch/$s_!mjUb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png 1272w, https://substackcdn.com/image/fetch/$s_!mjUb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mjUb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png" width="1240" height="254" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:254,&quot;width&quot;:1240,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:50706,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://upcurious.substack.com/i/191342106?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mjUb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png 424w, https://substackcdn.com/image/fetch/$s_!mjUb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png 848w, https://substackcdn.com/image/fetch/$s_!mjUb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png 1272w, https://substackcdn.com/image/fetch/$s_!mjUb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338e6219-4c40-4504-8b8b-04f1c943f02b_1240x254.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>HBM consumes roughly 3x the wafer capacity per gigabyte versus commodity DRAM &#8212; widening toward 4x with HBM4. AI is expected to consume 20% of total global DRAM wafer capacity in 2026. DRAM contract prices are up 40-70%.</p><p>Micron&#8217;s latest quarter makes Patel&#8217;s point more concrete. The company now says both DRAM and NAND supply-demand conditions should remain tight beyond calendar 2026. More importantly, management explicitly tied the DRAM constraint to cleanroom shortages, long construction lead times, a higher HBM trade ratio, and declining bits per wafer from node migrations. That is almost a public-company version of Patel&#8217;s framework.</p><p>We are looking at a memory supercycle, not a normal cyclical upswing. The demand curve rises hard and fast if compute roadmaps hold. Supply is concentrated in three players maintaining discipline. And the missing fabs mean this crunch cannot be solved quickly &#8212; the capacity decisions that would have eased it needed to happen two years ago.</p><p><strong>The trade:</strong> Three names, three different ways to own the same thesis.</p><p>Micron (MU) is the cleanest US public exposure, and the latest quarter materially strengthened the case. Fiscal Q2 revenue reached $23.86B with non-GAAP gross margin at 74.9%, and Q3 guidance points to roughly $33.5B of revenue with gross margin around 81%. DRAM inventory days remain especially tight at below 120 days, and Micron now expects fiscal 2026 capex to come in above $25B as it pushes harder on cleanrooms and long-duration capacity. MU now trades around ~10x forward earnings, and closer to ~5-6x if you annualize the current Q3 run rate.</p><p>The most important detail is what drove the quarter. This was not primarily a units story. DRAM bit shipments were up only mid-single digits sequentially, while DRAM prices rose in the mid-60s percentage range. NAND bit shipments were up low-single digits, while NAND prices increased in the mid-to-high 70s percentage range. That is what a bottleneck looks like in financial form: constrained supply showing up first in price, margin, and backlog urgency rather than just shipment volume.</p><p>SK Hynix is the dominant player &#8212; ~57% of global HBM supply, entire 2026 allocation sold out. The forward P/E sits around ~5x, which prices in Korea market risk and some skepticism about cycle duration. If you believe the supercycle extends past 2026, that discount is the opportunity. Accessible partly via EWY ETF for US investors, or directly on the Korean exchange.</p><p>Samsung is the wildcard. They are planning a ~50% capacity increase in 2026. Historically, Samsung has been willing to flood supply to gain share in NAND and commodity DRAM. But so far in this cycle, all three players are raising prices &#8212; Samsung and SK Hynix both hiked HBM3E prices ~20% for 2026, and server DRAM is up 60-70%. Samsung&#8217;s own exec said demand has already outpaced even the expanded supply. The risk is that discipline breaks as new capacity comes online. It hasn&#8217;t yet. Accessible partly via EWY alongside SK Hynix.</p><h3>Later: The Hard Ceiling &#8212; EUV and ASML</h3><p>By the end of the decade, Patel&#8217;s framework says the bottleneck drops to the deepest layer of the supply chain: the production of EUV lithography tools. These are the most complicated machines humans make, built by a single company in Veldhoven, relying on an artisanal supply chain of thousands of specialized suppliers.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ifrt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ifrt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png 424w, https://substackcdn.com/image/fetch/$s_!ifrt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png 848w, https://substackcdn.com/image/fetch/$s_!ifrt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png 1272w, https://substackcdn.com/image/fetch/$s_!ifrt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ifrt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png" width="1376" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1376,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1401833,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://upcuriousinvest.substack.com/i/191330543?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!ifrt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png 424w, https://substackcdn.com/image/fetch/$s_!ifrt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png 848w, https://substackcdn.com/image/fetch/$s_!ifrt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png 1272w, https://substackcdn.com/image/fetch/$s_!ifrt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c82916-c9d8-4b74-aa7a-0362c1511551_1376x768.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>ASML is expected to produce 60-70 EUV tools in 2026 &#8212; and will only reach a little over 100 per year by 2030. Each gigawatt of Rubin-scale AI compute requires around 3.5 EUV tools. That math implies an estimated upper bound of roughly 200 gigawatts of AI chip capacity by the end of the decade if enough of that tool base is allocated to AI.</p><p>Other constraints can ease. You can build more fabs. You can add power generation. You can expand packaging lines. But you can&#8217;t meaningfully accelerate EUV tool production &#8212; the ramp is measured in years, not quarters.</p><p>High-NA EUV &#8212; the next-generation TWINSCAN EXE:5200B at $350M per unit &#8212; has only shipped 8 tools in total, with high-volume manufacturing not expected until 2027-2028.</p><p><strong>The trade:</strong> The 10-year compounder</p><p>ASML trades at ~40x forward earnings with a EUR 38.8B backlog. The market understands the monopoly. What it may not fully price is that ASML does not just supply the system &#8212; it defines the outer limit of how large the system can get. It doesn&#8217;t matter which chip architecture wins, which hyperscaler spends the most, or which AI model is the best. Every leading-edge chip on earth goes through an ASML tool. The question isn&#8217;t whether you own it &#8212; it&#8217;s how much.</p><p>One caveat. ASML&#8217;s upside is constrained by physics. They can only produce ~60-70 EUV tools per year, ramping to ~100 by 2030. Raising prices helps at the margin &#8212; High-NA tools are $370-400M versus $200M for standard EUV &#8212; but it does not change the unit constraint that governs the system. Gross margins sit at 52-53% today, guided to 56-60% by 2030. That is gradual expansion, not a step function. At ~40x forward earnings, you are paying for a quality compounder where volume is physically capped and margin expansion is slow. The backlog is durable, but the growth rate has a ceiling built into the manufacturing process itself. Size accordingly.</p><div><hr></div><h2>The H100 Argument</h2><p>One of Patel&#8217;s more provocative claims: an H100 may be worth more today than it was several years ago.</p><p>Sounds backwards. But think about what an H100 actually is right now &#8212; not a depreciating asset on a balance sheet, but productive capital inside a world where models are getting better, inference demand is increasing, and useful AI work per installed accelerator keeps rising. If the software improves faster than supply expands, older hardware appreciates in economic usefulness even as newer hardware arrives.</p><p>Jensen Huang&#8217;s GTC 2026 keynote reinforces this. The market has shifted from training-heavy to inference-heavy. Every installed GPU is doing more valuable work than it was a year ago &#8212; not less.</p><p>Traditional compute depreciates. AI compute, in a supply-constrained world with improving models, may not. The installed base matters as much as the new shipment number.</p><div><hr></div><h2>Custom Silicon Is Part of the Map</h2><p>Google&#8217;s TPUs, Amazon&#8217;s Trainium, and the custom ASICs that Broadcom and Marvell design for hyperscalers all run through the same physical constraints &#8212; CoWoS packaging, DRAM wafers, EUV tools. They change the competitive dynamics within each layer, but they do not escape the constraints.</p><p>Broadcom&#8217;s AI semiconductor revenue is up 74% year-over-year, with Google and Amazon representing roughly 60% of that AI revenue. Marvell is growing AI silicon revenue 22%. These are not NVIDIA competitors in the traditional sense &#8212; they are parallel demand streams through the same constrained supply chain.</p><p>The bottleneck framework is not just a bull case for NVIDIA. It is a bull case for the physical supply chain regardless of whether the marginal AI dollar goes to GPUs or custom silicon. TSMC, ASML, Micron, the equipment names &#8212; they win either way. The GPU vs. ASIC debate is a distraction from the deeper question, which is whether the fabs and tools exist to build any of it fast enough.</p><div><hr></div><h2>China Is a Question of Timing</h2><p>Patel does not make a flat claim that the West wins and China loses. His view is more conditional, and it connects directly to the bottleneck framework.</p><p>Fast AI progress widens the Western advantage &#8212; the leading labs and infrastructure stack pull further ahead. Slower progress gives China time to catch up through manufacturing iteration and system-level adaptation.</p><p>The policy landscape adds texture. H200 exports were shifted to case-by-case review in January 2026 &#8212; partially reopened, but with conditions. Blackwell and more advanced architectures remain blocked. The AI Overwatch Act would tighten this further, treating AI chips as weapons exports.</p><p>The competitive map is shifting. China is making real progress in memory (CXMT), mature-node logic (SMIC), and system-level AI deployment (Huawei Ascend). Where it is not closing the gap is leading-edge logic &#8212; and that is exactly where ASML&#8217;s EUV monopoly and export controls create the deepest moat.</p><p>The faster the frontier moves, the more valuable the bottleneck owners become. Speed is what determines whether the Western infrastructure advantage compounds or narrows.</p><div><hr></div><h2>What Could Break the Thesis</h2><p>Every thesis has failure modes. Here is where this one breaks:</p><p><strong>Capex slows before application revenue catches up.</strong> Hyperscaler capex-to-revenue ratios at 45-57% are historically unsustainable. If even one major hyperscaler signals a capex pause, the infrastructure layers reprice quickly.</p><p><strong>The inference inflection disappoints.</strong> Jensen Huang&#8217;s $1T projection covers cumulative chip sales through 2027 across Blackwell and Vera Rubin, underpinned by his view that agentic AI requires significantly more compute and that inference demand is at an inflection point. If agentic workloads do not scale that aggressively, the demand extension that supports the HBM supercycle and NVIDIA&#8217;s $1T outlook loses its foundation.</p><p><strong>Deployment constraints ease faster than expected.</strong> If power, packaging, and construction bottlenecks resolve quickly &#8212; as Patel expects they will &#8212; the market may reprice those names before the deeper semiconductor constraints become fully visible. The transition from deployment bottleneck to manufacturing bottleneck is where positioning risk lives.</p><p><strong>Fab capacity arrives faster than expected.</strong> If the fabs halted in 2023 come online ahead of schedule, or if Samsung and SK Hynix accelerate clean room buildouts, the memory crunch eases sooner than the 2027-2028 timeline suggests. Samsung&#8217;s planned 50% HBM capacity surge in 2026 is the first test. If DRAM pricing momentum fades materially in mid-2026, the supercycle narrative needs revisiting.</p><p><strong>China closes the gap faster than expected.</strong> CXMT scaling DRAM production, SMIC advancing to N+2, or Huawei&#8217;s Ascend chips gaining meaningful inference share would compress the Western infrastructure premium. The lower-probability but higher-impact version: a genuine breakthrough in domestic EUV or alternative lithography &#8212; China has been investing heavily in DUV multi-patterning and homegrown lithography development &#8212; that reduces dependence on ASML tools at the leading edge.</p><div><hr></div><h2>What We Are Watching Next</h2><p>Five items tied to the next reporting cycle:</p><ol><li><p><strong>Micron Q3 FY26 execution against guidance</strong> &#8212; Management guided to $33.5B of revenue and roughly 81% gross margin. The next question is whether pricing, inventory tightness, and capex commentary keep confirming the supercycle thesis.</p></li><li><p><strong>ASML Q1 2026 earnings</strong> &#8212; EUV shipment count and High-NA EUV delivery timeline. Any acceleration in tool shipments changes the logic ceiling math.</p></li><li><p><strong>Hyperscaler capex commentary (Q1 2026 earnings cycle)</strong> &#8212; Watch for any signal of capex moderation from Microsoft, Google, Amazon, or Meta. The $660-750B 2026 estimate needs reconfirmation.</p></li><li><p><strong>DRAM contract pricing in Q2-Q3 2026</strong> &#8212; If HBM and DRAM prices plateau, the memory scarcity thesis weakens. If they re-accelerate, the supercycle extends.</p></li><li><p><strong>NVIDIA Vera Rubin shipment timeline</strong> &#8212; Any delay or acceleration reshapes the entire demand curve downstream.</p></li></ol><div><hr></div><p><em>Based on Dylan Patel&#8217;s interview on the Dwarkesh Podcast. Source: <a href="https://www.dwarkesh.com/p/dylan-patel">dwarkesh.com/p/dylan-patel</a></em></p><p><em>For more on the AI and how we track the stack, visit <a href="https://www.theupcurious.com/">theupcurious.com</a>.</em></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://upcurious.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Upcurious! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item></channel></rss>