<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[zach's tech blog]]></title><description><![CDATA[building chips and writing words]]></description><link>https://www.zach.be</link><image><url>https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png</url><title>zach&apos;s tech blog</title><link>https://www.zach.be</link></image><generator>Substack</generator><lastBuildDate>Tue, 28 Apr 2026 13:54:02 GMT</lastBuildDate><atom:link href="https://www.zach.be/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[zach]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[zachbe@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[zachbe@substack.com]]></itunes:email><itunes:name><![CDATA[zach]]></itunes:name></itunes:owner><itunes:author><![CDATA[zach]]></itunes:author><googleplay:owner><![CDATA[zachbe@substack.com]]></googleplay:owner><googleplay:email><![CDATA[zachbe@substack.com]]></googleplay:email><googleplay:author><![CDATA[zach]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Crossbar’s new security chip isn’t actually secure.]]></title><description><![CDATA[Designing tamper-proof chips is hard.]]></description><link>https://www.zach.be/p/crossbars-new-security-chip-isnt</link><guid isPermaLink="false">https://www.zach.be/p/crossbars-new-security-chip-isnt</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 16 Mar 2026 17:38:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://www.crossbar-inc.com/">Crossbar, Inc</a> is a company that was founded to commercialize <a href="https://en.wikipedia.org/wiki/Resistive_random-access_memory">resistive random access memory (RRAM)</a>, an emerging nonvolatile memory technology that&#8217;s one of the candidates to replace <a href="https://en.wikipedia.org/wiki/Flash_memory">eFlash</a> in modern process nodes. In the past few years, they&#8217;ve been starting to focus on ReRAM&#8217;s unique security properties. Compared to flash memory, it&#8217;s more robust to <a href="https://en.wikipedia.org/wiki/Laser_voltage_prober">laser voltage probing attacks</a>, and it can be used as a <a href="https://en.wikipedia.org/wiki/Physical_unclonable_function">physically unclonable function</a> to generate stable, secure random keys inside the chip.</p><p>Licensing secure memory and PUF IP isn&#8217;t super lucrative, though, so Crossbar decided to vertically integrate and start designing their own security chips and <a href="https://www.crossbar-inc.com/phsm.html">hardware products that integrate those chips</a>. But Crossbar didn&#8217;t want to design their chip in the way that most security chips are designed: closed-source and highly secretive. Instead, they worked with the <a href="https://www.bunniestudios.com/">legendary hardware hacker Bunnie Huang</a> to partially open-source their chip design.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> This is really cool, and it&#8217;s exciting to see more companies embracing open-source silicon; <a href="https://github.com/baochip/baochip-1x">you can go look at all the RTL right here</a>.</p><p>But there&#8217;s one issue with making your silicon open-source: everybody can see your mistakes. That includes people like me, who have been designing highly secure hardware for years. And Crossbar&#8217;s implementation of masked AES has a significant flaw that makes it fundamentally insecure against side-channel attacks. Today, we&#8217;re going to be looking at what went wrong, and how they could have done better.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>What makes a chip side-channel secure?</h1><p>Whenever the value on a wire in a chip switches from a 0 to a 1 or vice-versa, it consumes a small amount of power.  If you&#8217;re trying to design a secure chip that&#8217;s processing sensitive cryptographic keys, this can be a problem. For example, let&#8217;s say you want to send a key from one part of a chip to another along a <a href="https://en.wikipedia.org/wiki/Bus_(computing)">bus</a>. The amount of power consumed by that transaction is going to be related to the value of that cryptographic key. If an attacker measures the power consumption of this chip over a lot of transactions, they can reconstruct the value of the secret key. This is called a power analysis side-channel attack.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JTDs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JTDs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png 424w, https://substackcdn.com/image/fetch/$s_!JTDs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png 848w, https://substackcdn.com/image/fetch/$s_!JTDs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png 1272w, https://substackcdn.com/image/fetch/$s_!JTDs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JTDs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png" width="848" height="267" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/257911ff-3f18-497c-9f3a-bb373a082058_848x267.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:267,&quot;width&quot;:848,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JTDs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png 424w, https://substackcdn.com/image/fetch/$s_!JTDs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png 848w, https://substackcdn.com/image/fetch/$s_!JTDs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png 1272w, https://substackcdn.com/image/fetch/$s_!JTDs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F257911ff-3f18-497c-9f3a-bb373a082058_848x267.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Luckily, there are ways to prevent these side-channels. By XOR-ing the key data with random data, the power consumption of sending data along wires becomes randomized.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!s-SO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!s-SO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png 424w, https://substackcdn.com/image/fetch/$s_!s-SO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png 848w, https://substackcdn.com/image/fetch/$s_!s-SO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png 1272w, https://substackcdn.com/image/fetch/$s_!s-SO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!s-SO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png" width="848" height="267" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:267,&quot;width&quot;:848,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!s-SO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png 424w, https://substackcdn.com/image/fetch/$s_!s-SO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png 848w, https://substackcdn.com/image/fetch/$s_!s-SO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png 1272w, https://substackcdn.com/image/fetch/$s_!s-SO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f923d5e-9e69-4d23-9383-298fb5b6010f_848x267.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Unfortunately, it&#8217;s harder to mask actual computations this way. If we want to implement a masked AND gate, we can&#8217;t just XOR each input with a mask and then XOR the output by the same mask to recover the desired output value. We have to be more clever.</p><p>In 2003, Elena Trichina developed <a href="https://eprint.iacr.org/2003/236.pdf">an implementation of an AND gate that was supposed to be secure against side channel attacks</a>. Specifically, the Trichina AND gate was &#8220;1-probe secure&#8221;: the value of the secret key never appeared directly on a wire in the chip, so a single probe connected to any wire on the chip could not recover it. That (roughly) meant that the power consumption of the chip wouldn&#8217;t be proportional to the secret key.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> This notion is important: <strong>for a chip to be secure against power analysis, the real value of the secret key should never appear directly on a wire.</strong></p><p>Crossbar&#8217;s implementation is similar to Elena Trichina&#8217;s design, with some minor improvements. Unfortunately, this kind of design was shown to be insecure as early as 2005.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/crossbars-new-security-chip-isnt?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/crossbars-new-security-chip-isnt?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/crossbars-new-security-chip-isnt?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>Why isn&#8217;t Crossbar&#8217;s implementation secure?</h1><p>The physical logic gates and wires that make up a digital circuit all have different amounts of propagation delay through them; data doesn&#8217;t flow through these circuits instantly. This can result in behavior called &#8220;<a href="https://en.wikipedia.org/wiki/Glitch_removal">glitches</a>&#8221;, where intermediate values in a digital circuit transition from 0 to 1 and back multiple times during a single cycle as values propagate through the logic. In practice, glitches account for a significant portion of a chip&#8217;s power consumption&#8230; so they should also contribute to side-channel leakage.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kakc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kakc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png 424w, https://substackcdn.com/image/fetch/$s_!kakc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png 848w, https://substackcdn.com/image/fetch/$s_!kakc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png 1272w, https://substackcdn.com/image/fetch/$s_!kakc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kakc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png" width="342" height="395.352" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:289,&quot;width&quot;:250,&quot;resizeWidth&quot;:342,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kakc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png 424w, https://substackcdn.com/image/fetch/$s_!kakc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png 848w, https://substackcdn.com/image/fetch/$s_!kakc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png 1272w, https://substackcdn.com/image/fetch/$s_!kakc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F292ca72c-52a6-4187-944b-5d9cfd82b57c_250x289.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">An example of a glitch in a logic circuit. <em>(Wikipedia)</em></figcaption></figure></div><p>In early 2005, <a href="https://link.springer.com/chapter/10.1007/978-3-540-30574-3_24">a paper</a> confirmed that these sorts of glitches actually do cause side-channel leakage. By late 2005, glitching power consumption had been used to <a href="http://link.springer.com/content/pdf/10.1007/11545262_12.pdf">compromise masked AES implementations in practice</a>.</p><p>These glitching attacks work because, even when the value of the secret key is never exposed as the final state of a wire, it can become exposed as the intermediate state of a wire during a glitching event. This has the effect of compromising 1-probe security, which means that the power consumption of the chip starts to be correlated to the value of the secret key. This means that Crossbar&#8217;s chip isn&#8217;t actually secure against power analysis attacks.</p><p>There are <a href="https://www.zach.be/p/the-basics-of-side-channel-analysis?utm_source=publication-search">various modern techniques</a> that hardware designers use to implement side-channel secure algorithms even in the presence of glitches. The most common implementation I&#8217;ve seen in industry is <a href="https://eprint.iacr.org/2016/486.pdf">domain-oriented masking (DOM)</a>, which inserts register stages into a design to force signals to arrive in a certain order and prevent glitches that would leak data. Applying DOM to a circuit increases its latency, but overall, it&#8217;s a fairly resource-efficient technique for implementing a masked algorithm. Unfortunately, Crossbar didn&#8217;t decide to leverage DOM or a similar, glitch-secure masking scheme in their design.</p><p>I have to give kudos to Crossbar for open-sourcing their RTL for this chip. So much of the silicon we rely on every day is closed source -- especially security chips. But open-source security silicon poses some challenges. This isn&#8217;t like software, where bugs can be fixed with a simple update. Crossbar is putting these chips, which have clear security flaws, <a href="https://www.crossbar-inc.com/phsm.html">into real products</a>, and expecting people to use them to store their passkeys and custody their cryptocurrency. I respect Crossbar for trying to make open-source silicon, but they&#8217;re lacking the collaborative spirit that makes open-source development work so effectively.</p><h1>On open-source silicon.</h1><p>The right way to build open-source silicon isn&#8217;t just to take your design and publish the RTL online. It&#8217;s important to collaborate with the broader open-source community, which is aware of these sorts of attacks and how to prevent them. OpenTitan has <a href="https://opentitan.org/book/hw/ip/aes/doc/theory_of_operation.html#side-channel-analysis">a side-channel secure AES implementation</a> that leverages <a href="https://opentitan.org/book/hw/ip/aes/doc/theory_of_operation.html#side-channel-analysis">domain-oriented masking</a>. On the topic of a fully combinational masked S-box, like the one Crossbar is using, they say <a href="https://opentitan.org/book/hw/ip/aes/doc/theory_of_operation.html#side-channel-analysis">this</a> (emphasis mine):</p><blockquote><p>Alternatively, the two original versions of the masked Canright S-Box can be chosen as proposed by<a href="https://eprint.iacr.org/2009/011.pdf"> Canright and Batina: &#8220;A very compact &#8220;perfectly masked&#8221; S-Box for AES (corrected)&#8220;.</a> These are fully combinational (one S-Box evaluation every cycle) and have lower area footprint, but they are significantly less resistant to SCA. They are mainly included for reference but <strong>their usage is discouraged due to potential vulnerabilities</strong> to the correlation-enhanced collision attack as described by<a href="https://eprint.iacr.org/2010/297.pdf"> Moradi et al.: &#8220;Correlation-Enhanced Power Analysis Collision Attack&#8221;.</a></p></blockquote><p>If Crossbar had leveraged OpenTitan&#8217;s DOM AES implementation, not only would they have a more secure implementation, but it would serve as additional silicon proof of an agreed-upon implementation that&#8217;s gathering support among many different open-source silicon developers. Instead, Crossbar developed their own implementation, with flaws, and released it. Again, I respect them for making their silicon open source, but I hope open-source silicon developers in the future work more closely with the open-source community to ensure their designs are secure.</p><p>And one last note before I go: I would discourage any readers from buying any Crossbar PHSM devices due to this security flaw.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p> <a href="https://www.bunniestudios.com/blog/2026/baochip-1x-a-mostly-open-22nm-soc-for-high-assurance-applications/#comments">Bunnie didn&#8217;t work on the cryptographic hardware accelerators</a>, and this post is not a critique of him or any of his work on the Baochip-1x project. It is, however, a critique of Crossbar, because a company selling security chips should not be making this mistake.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>The d-probing model and how it relates to power side-channels is way beyond the scope of this article. Another time!</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Taalas is what Etched should have been.]]></title><description><![CDATA[I&#8217;m generating 17,000 tokens per second, and they&#8217;re all wrong!]]></description><link>https://www.zach.be/p/taalas-is-what-etched-should-have</link><guid isPermaLink="false">https://www.zach.be/p/taalas-is-what-etched-should-have</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 09 Mar 2026 12:55:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I&#8217;ve been an outspoken critic of Etched, the chip company that claims to be making AI chips that beat Nvidia by aggressively focusing on accelerating transformers and reducing programmability. This is a bad idea, because Nvidia is already focused on accelerating transformers, and they have more money and talent than Etched.</p><p>Recently, though, a new chip company called <a href="https://taalas.com/">Taalas</a> announced <a href="https://taalas.com/products/">their first silicon</a>. They&#8217;re also focusing on accelerating transformers by reducing programmability, but in a much, much more aggressive way. Etched wanted to build a chip that could only accelerate transformers; Taalas built a chip that could only <a href="https://taalas.com/the-path-to-ubiquitous-ai/">accelerate a specific transformer model, namely Llama 3.1 8</a>B. That specific model is hardwired into read-only memory (ROM) inside the chip, and can&#8217;t be changed.</p><p>This is a huge cost to pay. If you want to run a new model, you have to throw away the old chip and get a new one. But the benefits are also enormous. Taalas&#8217; HC1 proof-of-concept chip generates <a href="https://taalas.com/the-path-to-ubiquitous-ai/">17,000 tokens per second</a>, over two orders of magnitude more than their competitors. It&#8217;s truly absurdly fast, so naturally, people started going nuts about the tech demo. Today, let&#8217;s look at Taalas, separate the hype from the truth, and determine whether baking your model into ROM is actually a good idea.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Things people are getting wrong about Taalas</h1><p>In the excitement about the Taalas announcement, a lot of people have gotten a lot of things wrong. Before we jump into where I think Taalas could carve out a potential market niche, I want to say what they definitely aren&#8217;t going to be able to do. First off: Taalas is never going to be a chip for local inference on consumer devices.</p><p>While the Taalas chip is smaller and lower-power than an all-SRAM chip like the ones Groq, d-Matrix, or Cerebras make, it&#8217;s still not well-suited for edge devices. These devices care about memory density more than compute. For example, the HC1 stores Llama 3.1 8B, an 8 billion parameter model, with 3 bits per parameter, on a 800mm^2, reticle size logic die. That&#8217;s about 3 gigabytes of storage, and requires a large PCIe card&#8217;s worth of packaging, power delivery, and cooling. My lower-end iPhone, on the other hand, has 256GB of storage, so it can easily store and run Llama 3.1 8B alongside all of the pictures I&#8217;ve taken of my cats. It&#8217;ll just be much slower, because weights will be swapped around between the SSD, DRAM, and on-chip SRAM, but it&#8217;s much more practical than strapping a PCIe card to the back of your smartphone or laptop.</p><p>I also don&#8217;t think this is as big of a game-changer for coding agents that some people think it is. At least in my experience, developers significantly prefer using the newest, smartest models for coding tasks, even if they are more expensive. Giving an old model a ton of reasoning time is more likely to ruin your codebase than suddenly fix all your bugs. And given that most users of coding agents are enterprise customers with large budgets, I&#8217;d expect most of them will care about quality and running the most up-to-date models for their coding tasks. That makes the Taalas solution a poor fit.</p><p>On the other hand, a lot of the more &#8220;banal&#8221; consumer use-cases for AI aren&#8217;t nearly as sensitive to the underlying model quality. If you&#8217;re using an LLM to write high-performance Rust code, help diagnose rare diseases, or analyze the structural integrity of a bridge, you want the smartest model available. If you want an AI boyfriend to vent to or an alt girl AI gf to get off to, you&#8217;ll be fine with last year&#8217;s version of DeepSeek. I think those consumer use-cases, where price and speed matter more than quality, are going to be Taalas&#8217;s bread and butter.</p><p>There&#8217;s another, more technical reason why I think Taalas&#8217; chips are going to be better suited to ultra-low-cost consumer use-cases, rather than enterprise use-cases: how large context lengths affect the side of a model&#8217;s KV cache.</p><h1>KV cache and context length</h1><p>If you want to use an LLM for coding, you usually need to feed it a ton of input &#8211; usually, a significant portion of your codebase. It also usually generates fairly long <a href="https://openai.com/index/learning-to-reason-with-llms/">reasoning traces</a>, where the model talks itself through a complex problem before presenting a solution. These both pose challenges for Taalas&#8217;s architecture.</p><p>Transformer models consist of two main compute-heavy operations, each containing a ton of matrix multiplications: feed-forward layers, and multi-head attention. In the feed-forward layers, input data is multiplied by weights; because weights are fixed after training, Taalas can store them in ROM and access them extremely quickly and efficiently. In the multi-head attention layers, however, the input sequence is transformed to produce key (K), query (Q), and value (V) matrices that get longer as the input sequence grows. Because these K, Q and V matrices change with the input sequence, they have to be stored in on-chip SRAM, rather than being pre-stored in ROM like the model weights.</p><p>For extremely long context lengths, Taalas&#8217; chips will start to be bottlenecked by the KV cache operations in SRAM, even if their feed-forward layers run extremely quickly. These long context lengths show up all the time in enterprise use-cases (legal briefs, medical reports, and especially code), but not in consumer use-cases. This is just further evidence that Taalas&#8217;s niche may be delivering models for consumers to talk to, rather than for enterprises to develop software with.</p><p>However, it&#8217;s unclear to me whether consumer AI is going to be nearly as large of a market as enterprise AI. And Taalas is paying a massive cost to deliver chips that are so fast.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/taalas-is-what-etched-should-have?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/taalas-is-what-etched-should-have?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/taalas-is-what-etched-should-have?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>The logical conclusion of &#8220;speed at any cost&#8221;</h1><p>Taalas chips may be much more power efficient than their competitors, but the total cost of ownership (TCO) of a Taalas system is truly staggering. The Taalas HC1 may just be one chip, but if you want to implement a larger model using Taalas&#8217;s technology, all of its weights won&#8217;t fit on a single chip. Instead, you need to manufacture multiple unique chips for different sections of the model. Taalas&#8217;s own predictions assume that a model like Deepseek R1 would require <a href="https://www.eetimes.com/taalas-specializes-to-extremes-for-extraordinary-token-speed/">30 unique chip tape-outs.</a></p><p>Now, if you stay up-to-date with the industry or <a href="https://semianalysis.com/2022/07/24/the-dark-side-of-the-semiconductor/?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web">read SemiAnalysis</a> you&#8217;re probably gasping at the astronomical cost of 30 tapeouts. A cutting edge tapeout normally costs tens of millions of dollars &#8211; with those numbers, a Taalas system would cost hundreds of millions in <a href="https://en.wikipedia.org/wiki/Photomask">mask sets</a> alone, before you even manufacture a single chip. But they&#8217;re doing something clever. They can change the weights, matrix dimensions, and other various control parameters with only two mask layers. The other mask layers in the chip, of which there are often over 10, are unchanged. That means that most of the cost of doing many different tapeouts is actually shared, and a Taalas system might cost &#8220;only&#8221; $100M or less in mask sets to deploy.</p><p>Taalas says that their chips are so efficient that it makes sense to pay this massive upfront cost. If a datacenter leverages Taalas chips, they&#8217;re spending less power, and therefore less money, on each token, and can also use fewer chips to achieve the same throughput. So the longer they use the Taalas system, the bigger its cost advantages are. But because the Taalas systems are hardwired to only use one model, it&#8217;s also the case that the longer they have the Taalas system, the more out-of-date it becomes. Taalas says the economics work out <a href="https://www.eetimes.com/taalas-specializes-to-extremes-for-extraordinary-token-speed/">if you assume a 1-year lifetime of each Taalas datacenter.</a></p><p>Ultimately, from a technology perspective, Taalas has a new, cool idea. Unlike Etched, there&#8217;s no hand-wavey magic here &#8211; it makes complete sense how wiring weights into ROM can deliver massive performance increases. But if nobody is willing to use out-of-date LLMs, all those performance issues are for naught. Given that the biggest LLM application right now is coding, which should have a strong preference for the most up-to-date models, I&#8217;m not sure if Taalas will be a massive success. But if consumer AI use-cases continue to grow, Taalas hardware could be powering that wave of consumer AI apps that leverage last year&#8217;s models running lightning fast.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why is OpenAI partnering with Cerebras?]]></title><description><![CDATA[TSMC's manufacturing capacity rules everything around me.]]></description><link>https://www.zach.be/p/why-is-openai-partnering-with-cerebras</link><guid isPermaLink="false">https://www.zach.be/p/why-is-openai-partnering-with-cerebras</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 02 Feb 2026 16:09:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A few weeks ago, OpenAI announced a massive new partnership with Cerebras, the wafer-scale AI chip company. This came as a big surprise to those of us who have been watching Cerebras for years -- the AI chip startup had, until recently, <a href="https://www.zach.be/p/why-is-cerebras-trying-to-go-public">no major customers other than the UAE-based G42, who was also an investor in the company</a>. That lack of customers primarily stemmed from the total cost of ownership (TCO) of Cerebras systems.</p><p>Like other AI accelerators that lack high-bandwidth memory (HBM), Cerebras needed to network together <a href="https://www.zach.be/p/most-ai-chips-arent-cost-effective">many chips to host a single large model</a>. The relatively small Llama 3.1 70B LLM takes 4 racks of Cerebras wafer-scale systems, <a href="http://www.zach.be/p/why-is-cerebras-trying-to-go-public">each of which costs between two and three million dollars</a>. An Nvidia DGX H100 can run the same workload at a tenth of the cost. And the TCO difference only grows for larger models, which require more and more Cerebras chips to run without p<a href="http://www.zach.be/p/most-ai-chips-arent-cost-effective">erformance plummeting due to model weights being moved off-chip</a>.</p><p>Cerebras would claim that this cost overhead is worth it for the ultra-low latency that they provide. They&#8217;re not wrong that Cerebras chips, by storing all of their data in SRAM, offer significantly lower inference latency than equivalent Nvidia solutions. The same is true for other SRAM-based inference chips, like <a href="http://www.zach.be/p/most-ai-chips-arent-cost-effective">d-Matrix</a> and the <a href="http://www.zach.be/p/why-did-nvidia-acqui-hire-groq">recently-acquired Groq</a>. But players like <a href="http://www.zach.be/p/why-is-sambanova-giving-up-on-ai">SambaNova</a> and <a href="https://www.zach.be/p/ai-chip-startups-dont-need-to-be">Furiosa</a> also offer relatively high-speed inference without the same cost tradeoff, due to their inclusion of HBM. So if Cerebras&#8217; chips are so expensive, and their low-latency inference isn&#8217;t that unique, why is OpenAI partnering with them?</p><p>In my opinion, what actually makes Cerebras unique is that, after the Groq acquisition, they are the only major AI chip startup that has already deployed chips at scale that isn&#8217;t limited by TSMC&#8217;s chip-on-wafer-on-substrate (CoWoS) manufacturing capacity.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>CoWoS, HBM, and SRAM</h1><p>As I discussed in <a href="https://www.zach.be/p/why-did-nvidia-acqui-hire-groq">my article on Nvidia&#8217;s Groq acquisition</a>, Nvidia&#8217;s chip manufacturing volume is limited by TSMC&#8217;s CoWoS manufacturing capacity. CoWoS packaging is expensive, difficult, and has limited production volume. Nvidia isn&#8217;t the only company fighting for that limited volume; <a href="https://www.tomshardware.com/news/amd-and-nvidia-gpus-consume-lions-share-of-tsmc-cowos-capacity">AMD uses CoWoS packaging for their AI chips as well</a>. Plus, startups are trying to leverage CoWoS, with <a href="https://arxiv.org/abs/2405.07518v1">SambaNova also using CoWoS packaging</a>, and other startups like Furiosa, Etched, and MatX probably will as well. CoWoS capacity shortages are the norm, and <a href="https://www.tomshardware.com/news/amd-and-nvidia-gpus-consume-lions-share-of-tsmc-cowos-capacity">are predicted to persist for years</a>.</p><p>CoWoS packaging is key to the HBM-based architectures that Nvidia, AMD, and startups like SambaNova and Furiosa are selling. With massive, low-latency banks of HBM, these chips can store LLM weights, biases, and KV cache values off-chip while still accessing them relatively quickly. This high memory density enables a datacenter to serve small models like Llama 3.1 70B with less than one rack of servers, offering a cost-efficient way to deliver inference.</p><p>On the other hand, SRAM-based architectures, like the ones developed by <a href="https://www.zach.be/p/why-is-cerebras-trying-to-go-public?utm_source=publication-search">Cerebras</a>, <a href="https://www.zach.be/p/why-did-nvidia-acqui-hire-groq?utm_source=publication-search">Groq</a>, and <a href="https://www.zach.be/p/most-ai-chips-arent-cost-effective?utm_source=publication-search">d-Matrix</a>, have to store all that data on-chip. On-chip SRAM memory is far, far less dense and far, far more expensive than off-chip SRAM. That&#8217;s why it takes over ten million dollars to build out a Cerebras system for even a small model. On the other hand, SRAM is much faster than HBM, which lets these chips run inference workloads faster than their HBM counterparts.</p><p>For years, analysis and the market have been looking at this tradeoff and deciding it wasn&#8217;t worth it. Nobody wanted to pay orders of magnitude more for a system that was slightly faster. That&#8217;s why <a href="https://www.zach.be/p/why-groq-might-struggle-as-an-api?utm_source=publication-search">Groq had to pivot from selling chips, which nobody would buy, to running their own inference clusters</a>. It&#8217;s also why <a href="https://www.zach.be/p/why-is-cerebras-trying-to-go-public?utm_source=publication-search">Cerebras&#8217; only major customer was G42</a>, the Emirati company that was also one of their major investors.</p><p>But the limited supply of CoWoS packaging changes the equation. If every bit of TSMC capacity is used for Nvidia and AMD chips that are already flying off the shelves, what do you do when you want even more compute? You might need to start buying chips with a worse cost-to-performance tradeoff.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-is-openai-partnering-with-cerebras?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-is-openai-partnering-with-cerebras?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-is-openai-partnering-with-cerebras?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>The AI bubble doesn&#8217;t care about cost</h1><p><a href="https://www.nytimes.com/2026/01/29/technology/openai-in-talks-to-raise-as-much-as-100-billion.html">OpenAI is in talks to raise as much as $100 billion dollars</a>. <a href="https://techcrunch.com/2026/01/27/anthropic-reportedly-upped-its-latest-raise-to-20b/">Anthropic is raising $20 billion</a>. And existing tech giants like Meta and Google are pouring countless billions into their own AI efforts. Put simply, the AI giants have more money than they know what to do with. And many of them see the AI race as a winner-take-all effort. Any small advantage now could be worth trillions of dollars in the future if it gives a company a substantial competitive edge in developing and deploying the next models sooner.</p><p>What Cerebras is offering is just that. It&#8217;s not actually a better chip than what Nvidia has -- in fact, it offers a worse cost-to-performance tradeoff. But the demand for AI chips is so high that &#8220;almost as good as Nvidia but still worse&#8221; has become a multi-billion dollar value proposition. Every single one of Nvidia&#8217;s chips is already being deployed in AI datacenters, and they can&#8217;t make more unless TSMC adds additional CoWoS manufacturing capacity. Cerebras is able to fill the excess demand for compute that exists past that limited TSMC manufacturing capacity.</p><p>As a brief aside &#8212; Cerebras is not the only player offering all-SRAM inference chips. <a href="https://www.zach.be/p/why-did-nvidia-acqui-hire-groq">But Groq was just acquired by Nvidia</a>, and the only other player in this space, d-Matrix doesn&#8217;t have chips deployed at scale in the way Cerebras does. If OpenAI wants to source additional compute, even at a relatively high cost, it makes sense to pick the player with the lowest risk. In this case, that&#8217;s Cerebras.</p><p>Ultimately, as the demand for compute grows, the cost-efficiency and power-efficiency of different chips starts to matter less and less, and the availability of those chips starts to matter more. This explains <a href="http://www.zach.be/p/ai-chip-startups-dont-need-to-be">Nvidia&#8217;s acquisition of Groq</a>, and it explains OpenAI&#8217;s Cerebras deal. Sure, these chips might be slightly worse than Nvidia&#8217;s H100 and B200s at serving large models at scale, but Nvidia&#8217;s chips are a limited resource. When the demand for AI chips is so high that Nvidia is running out of stock, &#8220;almost as good as Nvidia&#8221; becomes a viable strategy.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why did AMD’s random number generator start returning 0x0000?]]></title><description><![CDATA[And why cryptographic randomness matters.]]></description><link>https://www.zach.be/p/why-did-amds-random-number-generator</link><guid isPermaLink="false">https://www.zach.be/p/why-did-amds-random-number-generator</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 05 Jan 2026 18:20:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!eu3h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>If there&#8217;s one thing we love here at Zach&#8217;s Tech Blog, it&#8217;s <a href="https://www.zach.be/p/how-did-amds-secure-enclave-get-hacked">weird</a> security <a href="https://www.zach.be/p/the-most-secure-chip-in-the-world">vulnerabilities</a> in <a href="https://www.zach.be/p/apples-m-series-cpus-just-got-hacked">processors</a>. So when I saw that there was a <a href="https://www.amd.com/en/resources/product-security/bulletin/amd-sb-7055.html">security flaw in AMD&#8217;s Zen 5 processors</a> that made its random seed instruction, RDSEED, start returning far too many zeroes to be truly random, I knew I had to write about it. Plus, it&#8217;s a really great excuse to talk about one of my favorite topics in cryptography: why elliptic curve signatures are really hard to implement securely.</p><p>Generating true random numbers is really important for cryptography. Some algorithms don&#8217;t require particularly special random numbers, but others, like the elliptic curve digital signature algorithm (ECDSA), are incredibly sensitive to the &#8220;quality&#8221; of their randomness. If you&#8217;re using ECDSA, even slight biases in your random numbers can allow attackers to extract your private key, given enough signatures.</p><p>What makes a random number generator good enough for cryptography, then? Well, its output should be computationally indistinguishable from truly random numbers. That&#8217;s usually measured by <a href="https://en.wikipedia.org/wiki/Randomness_test">standard statistical tests</a>. Usually, the best random number generators are so-called true random number generators, or TRNGs, which use randomness extracted from the environment. But the specific flaw AMD disclosed lay within their implementation of the RDSEED instruction, which is supposed to generate random output using their TRNG. This should be the best source of randomness on the chip, but it suddenly started returning far more zeroes than expected. So what went wrong?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>What was wrong with AMD&#8217;s TRNG?</h1><p>Most TRNGs deployed in modern CPUs use some form of analog noise on the chip to generate truly random numbers. While AMD hasn&#8217;t publicly confirmed their TRNG&#8217;s randomness source, it&#8217;s most likely based on <a href="https://en.wikipedia.org/wiki/Hardware_random_number_generator#Free-running_oscillators-based_RNG">free-running ring oscillators</a>. The frequency of free-running ring oscillators changes randomly based on temperature, jitter, power supply noise, and all other sorts of random environmental factors. If you sample the state of that oscillator with your much-more-stable system clock, you can generate random bits.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eu3h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eu3h!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png 424w, https://substackcdn.com/image/fetch/$s_!eu3h!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png 848w, https://substackcdn.com/image/fetch/$s_!eu3h!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png 1272w, https://substackcdn.com/image/fetch/$s_!eu3h!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eu3h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png" width="600" height="379" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:379,&quot;width&quot;:600,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eu3h!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png 424w, https://substackcdn.com/image/fetch/$s_!eu3h!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png 848w, https://substackcdn.com/image/fetch/$s_!eu3h!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png 1272w, https://substackcdn.com/image/fetch/$s_!eu3h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9b6715-30d9-4b3a-b645-195be6e0b761_600x379.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://ieeexplore.ieee.org/document/4731825">Ring Oscillator TRNG from Wold et. al.</a></em></figcaption></figure></div><p>Unfortunately, the bitstream you generate from your ring oscillator is often biased in some way -- a specific chip may randomly sample more ones than zeroes, simply due to the process variation on that piece of silicon. So it&#8217;s common to use some form of <a href="https://en.wikipedia.org/wiki/Randomness_extractor">randomness extraction</a>, conditioning, and <a href="https://en.wikipedia.org/wiki/Whitening_transformation">whitening</a> algorithms to post-process your raw random bits into a set of uncorrelated random bits suitable for use in a cryptographic algorithm. Modern cryptographic TRNGs usually use a combination of <a href="https://en.wikipedia.org/wiki/Randomness_extractor#Von_Neumann_extractor">Von Neumann conditioning</a> and <a href="https://en.wikipedia.org/wiki/Randomness_extractor#Cryptographic_hash_function">hash based conditioning</a> to post-process their random source.</p><p>The Von Neumann conditioning algorithm is particularly simple and clever -- it groups output bits two-by-two, discards any pair where the bits are the same, and converts 01 into 0 and 10 into 1. This removes simple biases from the raw bitstream:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zlnp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zlnp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png 424w, https://substackcdn.com/image/fetch/$s_!zlnp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png 848w, https://substackcdn.com/image/fetch/$s_!zlnp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png 1272w, https://substackcdn.com/image/fetch/$s_!zlnp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zlnp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png" width="576" height="162" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:162,&quot;width&quot;:576,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zlnp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png 424w, https://substackcdn.com/image/fetch/$s_!zlnp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png 848w, https://substackcdn.com/image/fetch/$s_!zlnp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png 1272w, https://substackcdn.com/image/fetch/$s_!zlnp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79280ea3-463c-425d-8fb6-4f7ebd72e20a_576x162.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption"><em><a href="https://pit-claudel.fr/clement/blog/generating-uniformly-random-data-from-skewed-input-biased-coins-loaded-dice-skew-correction-and-the-von-neumann-extractor/">Von Neumann Extractor, from Clement Pit-Claudel</a></em></figcaption></figure></div><p>One big problem with this TRNG architecture is that it&#8217;s just a bit slow. Multiple ring oscillators are combined to return one raw random bit per cycle. The Von Neumann conditioning algorithm ends up discarding more than half of those bits to remove bias. Then, bits need to pass through a cryptographic hash function for additional post-processing before they can actually be used.</p><p>Because of this, modern CPUs don&#8217;t generate randomness on-demand, whenever a user invokes the RDSEED &#8212; it would be too slow. Instead, CPUs are constantly generating random numbers and storing them in a buffer, which the RDSEED instruction fetches random numbers from. But this means that, if a user invokes RDSEED too often, the buffer can run out of random numbers. In that case, the instruction is supposed to return with its carry flag, CF, set to 0, which indicates to software that the randomness buffer was not ready and that it should try again.</p><p>The issue with AMD&#8217;s Zen5 processors is that, for the 16-bit and 32-bit versions of the RDSEED instruction, CF gets set to 1 even if the randomness buffer is empty. And when the randomness buffer is empty, it returns zero, rather than a real random value.</p><p>But wait -- is it really that bad if a TRNG returns zero slightly more than it should? Well, it turns out that it can actually be extremely bad. In fact, even slight biases in random numbers can be catastrophic.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-amds-random-number-generator?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-amds-random-number-generator?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-did-amds-random-number-generator?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>ECDSA and the PlayStation 3</h1><p><a href="https://en.wikipedia.org/wiki/Elliptic-curve_cryptography">Elliptic curve cryptography</a> (ECC) replaced the <a href="https://en.wikipedia.org/wiki/RSA_cryptosystem">RSA cryptosystem</a> in the mid-2000s because of its ability to provide similar security levels with much smaller keys: <a href="https://www.ssl.com/article/comparing-ecdsa-vs-rsa-a-simple-guide/">a 256-bit ECDSA key has comparable security to a 3072-bit RSA key</a>. However, RSA-based signatures had one major advantage over the elliptic curve digital signature algorithm (ECDSA). Performing an ECDSA signature requires <a href="https://en.wikipedia.org/wiki/Elliptic_Curve_Digital_Signature_Algorithm#Signature_generation_algorithm">generating an ephemeral random number</a> called a <a href="https://en.wikipedia.org/wiki/Cryptographic_nonce">nonce</a>. If that nonce is reused, attackers can easily determine the private signing key, <a href="https://fahrplan.events.ccc.de/congress/2010/Fahrplan/attachments/1780_27c3_console_hacking_2010.pdf">as they famously did to extract the signing key for the PlayStation3</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UaYn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UaYn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png 424w, https://substackcdn.com/image/fetch/$s_!UaYn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png 848w, https://substackcdn.com/image/fetch/$s_!UaYn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png 1272w, https://substackcdn.com/image/fetch/$s_!UaYn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UaYn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png" width="891" height="565" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/14a49939-d3ba-408b-9489-a372f920e053_891x565.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:565,&quot;width&quot;:891,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UaYn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png 424w, https://substackcdn.com/image/fetch/$s_!UaYn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png 848w, https://substackcdn.com/image/fetch/$s_!UaYn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png 1272w, https://substackcdn.com/image/fetch/$s_!UaYn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a49939-d3ba-408b-9489-a372f920e053_891x565.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Don&#8217;t be like Sony <a href="https://fahrplan.events.ccc.de/congress/2010/Fahrplan/attachments/1780_27c3_console_hacking_2010.pdf">(fail0verflow, 2010)</a></em></figcaption></figure></div><p>Luckily, it&#8217;s pretty easy to avoid making this specific implementation mistake. Unfortunately, even if you change the nonce with every signature, slight biases in the RNG used to generate the nonce will still leak the private key! Each time an ECDSA key is used to generate a signature, if the nonce is biased, it leaks some information about that key; with enough signatures, an attacker can reconstruct the key. This happened in the real world when <a href="https://lists.debian.org/debian-security-announce/2008/msg00152.html">maintainers of the Debian-specific fork of the OpenSSL package modified their PRNG in a way that (hopefully accidentally) introduced significant bias.</a> If a software developer was using AMD&#8217;s RDSEED instruction to generate ECDSA nonces, they could end up signing multiple messages with a nonce of zero, or a nonce biased towards containing more zero bits in it, and compromising their private keys.</p><p>Thankfully, there are ways to work around the TRNG bug in AMD&#8217;s chips. The 64-bit version of RDSEED isn&#8217;t affected, so software developers can just use that instruction without concern. But it&#8217;s an interesting lesson in how TRNGs work and the ways very minor bugs can cause very major security flaws when they&#8217;re integrated into a full system.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why did Nvidia acqui-hire Groq?]]></title><description><![CDATA[And why not SambaNova, Cerebras, or d-Matrix?]]></description><link>https://www.zach.be/p/why-did-nvidia-acqui-hire-groq</link><guid isPermaLink="false">https://www.zach.be/p/why-did-nvidia-acqui-hire-groq</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Fri, 26 Dec 2025 22:19:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>On December 24th, the semiconductor industry got an early Christmas present with the announcement that <a href="https://www.cnbc.com/2025/12/24/nvidia-buying-ai-chip-startup-groq-for-about-20-billion-biggest-deal.html">Nvidia was acquiring Groq</a>. The CNBC article breaking the story originally read:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nKve!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nKve!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg 424w, https://substackcdn.com/image/fetch/$s_!nKve!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg 848w, https://substackcdn.com/image/fetch/$s_!nKve!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!nKve!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nKve!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg" width="680" height="284" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:284,&quot;width&quot;:680,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:54852,&quot;alt&quot;:&quot;Image&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Image" title="Image" srcset="https://substackcdn.com/image/fetch/$s_!nKve!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg 424w, https://substackcdn.com/image/fetch/$s_!nKve!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg 848w, https://substackcdn.com/image/fetch/$s_!nKve!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!nKve!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06bb159b-e024-4cce-8475-bc63478d3947_680x284.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Within the hour, Groq made their own announcement:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bqL1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bqL1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png 424w, https://substackcdn.com/image/fetch/$s_!bqL1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png 848w, https://substackcdn.com/image/fetch/$s_!bqL1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png 1272w, https://substackcdn.com/image/fetch/$s_!bqL1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bqL1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png" width="953" height="202" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e74a602e-9743-4787-9828-909c3b6b5c73_953x202.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:202,&quot;width&quot;:953,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:53788,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.zach.be/i/182655242?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bqL1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png 424w, https://substackcdn.com/image/fetch/$s_!bqL1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png 848w, https://substackcdn.com/image/fetch/$s_!bqL1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png 1272w, https://substackcdn.com/image/fetch/$s_!bqL1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe74a602e-9743-4787-9828-909c3b6b5c73_953x202.png 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><p>Instead of talking about an acquisition, Groq described a non-exclusive licensing agreement in which key Groq executives and engineers joined Nvidia, but Groq continued as an independent company. Shortly afterwards, the CNBC story was updated to reflect this reality.</p><p>Groq also didn&#8217;t mention the $20B number anywhere in their announcement, and Nvidia has yet to make any public comment. However, I&#8217;ve since gotten independent confirmation from an anonymous source with knowledge of the deal that it was, in fact, worth $20B.</p><p>It&#8217;s not a good look that startups are getting gutted in <a href="https://techcrunch.com/2025/07/14/cognition-maker-of-the-ai-coding-agent-devin-acquires-windsurf/">weird reverse-acqui-hire deals</a> that strip key talent and technology without necessarily bringing all of the employees along for the ride. But I&#8217;m not an expert on M&amp;A regulations, and this is Zach&#8217;s Tech Blog, not Zach&#8217;s Regulatory Compliance Blog, so I&#8217;m not going to discuss the specific deal structure here.</p><p>The important thing is that Nvidia did genuinely pay twenty billion dollars for Groq&#8217;s technology and key talent. From the outside, Groq is not worth $20B. Their <a href="https://www.investing.com/news/company-news/groq-slashes-2025-revenue-projections-to-500-million--the-information-93CH-4158309">projected revenue for 2025 was $500M</a>, so the price-to-sales ratio for this deal was an eye-watering 40 &#8212; 4x higher than the ratio of companies like Apple and Google. The first version of the LPU chip was taped out in 2020, making it woefully outdated compared to other companies&#8217; more recent chips. The TCO of a Groq cluster is unreasonably high, <a href="https://www.zach.be/p/why-is-everybody-talking-about-groq?utm_source=publication-search">requiring hundreds of chips worth millions of dollars in total to run relatively small open-source models like Llama70B</a>. Jonathan Ross and the other Groq executives don&#8217;t have a deep knowledge of Google&#8217;s TPU architecture, as some analysts are claiming; they left Google to found Groq after the first generation of the TPU, and Google is on the 8th generation by now. So&#8230; why do I think Nvidia spent $20B on Groq? Here are a few possibilities.</p><p><em><strong>Fair warning: this is all wild speculation. But speculation is fun!</strong></em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>The LPUv2 is something special.</h1><p>Dylan Patel of <a href="https://semianalysis.com/">SemiAnalysis</a> was quoted in <a href="https://www.theinformation.com/articles/nvidia-struck-20-billion-megadeal-groq">the Information</a> with this theory:</p><blockquote><p>&#8220;Groq&#8217;s first-generation chips were not competitive [with Nvidia&#8217;s chips], but there are two [more] generations coming back-to-back soon,&#8221; said Dylan Patel, chief analyst at chip consultancy SemiAnalysis. &#8220;Nvidia likely saw something they were scared of in those.&#8221;</p></blockquote><p>He&#8217;s right about the <a href="https://www.zach.be/p/why-is-everybody-talking-about-groq?utm_source=publication-search">LPUv1 being an uncompetitive chip</a>. As for the LPUv2, Groq has done a very good job keeping architectural details of their new chips from leaking. We do know that it&#8217;s being manufactured in <a href="https://www.kedglobal.com/korean-chipmakers/newsView/ked202308160014">Samsung&#8217;s 4nm process</a>, which may be important &#8212; more on that later.</p><p>But unless Groq has cooked up some <a href="https://sites.utexas.edu/CRL/files/2021/04/CICC_2021_3D_split_SRAM_slides.pdf">3D stacked SRAM architecture</a> that blows Nvidia out of the water, I&#8217;m personally doubtful that there&#8217;s something in the LPUv2 that makes it significantly better than Nvidia&#8217;s chips. Groq had <a href="https://www.teamblind.com/post/thinking-about-joining-groq-ncga4mvv">some major layoffs</a> and <a href="https://irrationalanalysis.substack.com/p/very-long-incoherent-writeup?utm_source=publication-search">lost a ton of great talent to attrition, including their old Chief Architect</a>. That&#8217;s not a great recipe for a company that&#8217;s suddenly able to out-perform Nvidia at scale.</p><h1>Groq has some unique partnership Nvidia wants.</h1><p>Maybe Groq has some <a href="https://x.com/IAmAdiFuchs/status/2004050560769839276">sales partnership or other business deal that Nvidia really wants access to</a>. Given the &#8220;non-exclusive licensing deal&#8221; structure, I&#8217;m not sure how that would work, but it&#8217;s certainly a possibility. For all of its technical weaknesses, <a href="https://www.zach.be/p/why-groq-might-struggle-as-an-api">Groq&#8217;s business development team has proven to be particularly shrewd</a>, turning mediocre and outdated silicon into multiple huge funding rounds. They were the first AI chip startup to pivot to selling inference tokens as their primary product, rather than trying to sell chips to hyperscalers &#8212; and most of their competitors have been following suit.</p><p>I&#8217;m not sure what partnership Groq has that Nvidia might want, though. The only major AI lab that Groq has a partnership with is Meta, <a href="https://groq.com/newsroom/meta-and-groq-collaborate-to-deliver-fast-inference-for-the-official-llama-api">which offers Groq&#8217;s cloud as an optional inference provider in the Llama4 API</a>. But that partnership isn&#8217;t unique &#8212; <a href="https://www.cerebras.ai/press-release/meta-collaborates-with-cerebras-to-drive-fast-inference-for-developers-in-new-llama-api">Cerebras is offered as another API inference provider on equal footing to Groq</a>.</p><p>A lot of Groq&#8217;s other big partnerships are with Middle Eastern countries and governments. They have partnerships with the <a href="https://groq.com/newsroom/saudi-arabia-announces-1-5-billion-expansion-to-fuel-ai-powered-economy-with-ai-tech-leader-groq?utm_source=chatgpt.com">Saudi Arabian government</a> as well as <a href="https://techafricanews.com/2025/12/10/zain-ksa-and-groq-introduce-high-performance-ai-platform-developed-in-saudi-arabia/?utm_source=chatgpt.com">various</a> <a href="https://www.humain.com/en/news/humain-and-groq-deliver?utm_source=chatgpt.com">Saudi</a> <a href="https://groq.com/newsroom/aramco-digital-and-groq-announce-progress-in-building-the-worlds-largest-inferencing-data-center-in-saudi-arabia-following-leap-mou-signing?utm_source=chatgpt.com">companies</a> to deploy Groq systems at scale. While these partnerships are valuable, they&#8217;re certainly not unique. <a href="https://www.zach.be/p/why-is-cerebras-trying-to-go-public?utm_source=publication-search">Cerebras also has a large number of Middle Eastern partnerships</a> &#8212; in their case, with the UAE.</p><p>So Groq doesn&#8217;t have any publicly announced partnerships that are so unique that they could justify a $20B price tag. There could be something that isn&#8217;t public yet, of course, but going purely based on public info, I think there&#8217;s only one thing that could make Groq this valuable: the fact that Groq&#8217;s chips could help Nvidia make their supply chain more resilient and less reliant on TSMC.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-nvidia-acqui-hire-groq?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-nvidia-acqui-hire-groq?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-did-nvidia-acqui-hire-groq?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>Supply chain resilience, and reducing reliance on TSMC.</h1><p>This is my personal pet theory, so I saved it for last. Of all of the different AI chip startups, Groq is unique in that its product just consists of a straightforward logic die manufactured in a Samsung process node. <a href="https://arxiv.org/abs/2405.07518v1">SambaNova relies on TSMC&#8217;s Chip-on-Wafer-on-Substrate (CoWoS) packaging</a> to leverage high-bandwidth memory (HBM), just like Nvidia&#8217;s chips. Furiosa, Etched, and MatX all use, or plan to use, HBM as well, so they also need a 2.5D packaging solution to package the HBM next to the logic die. And each 2.5D packaging solution is tied to a specific foundry and packaging provider &#8212; probably TSMC.</p><p>Other architectures without HBM also have foundry-specific features. <a href="https://d-matrix.ai/pdf/d-Matrix-WhitePaper-Technical-FINAL.pdf">d-Matrix uses TSMC-specific, custom SRAM bit-cells for their processing-in-memory architecture</a>, as well as an <a href="https://www.eetimes.com/d-matrix-targets-fast-llm-inference-for-real-world-scenarios/">organic interposer for their chiplet-based packaging solution</a>. And Cerebras&#8217;s wafer scale engine requires a ton of <a href="https://www.cerebras.ai/press-release/cerebras-systems-unveils-the-industrys-first-trillion-transistor-chip">specialized processing that was developed in collaboration with TSMC</a>.</p><p>Nvidia&#8217;s Blackwell B200 and upcoming Rubin architectures are already incredibly reliant on TSMC&#8217;s CoWoS packaging, which has limited production volume and is incredibly expensive. If Nvidia wants to make their supply chain more resilient and make themselves less dependent on TSMC, there are very few acquisitions they could make that would help them out. SambaNova, Furiosa, Etched, MatX, Cerebras, and d-Matrix all rely on some foundry-specific feature that would make their design extremely hard to port from one process node to another. That means that if TSMC suddenly runs out of production capacity, raises CoWoS prices, or even worse, gets invaded by mainland China, Nvidia would still be out of luck.</p><p>Groq&#8217;s chips, on the other hand, <a href="https://www.kedglobal.com/korean-chipmakers/newsView/ked202308160014">are already manufactured in a Samsung process</a>, and could easily be ported to an Intel process. By relying on a simple logic chip that can be manufactured by multiple foundries, the supply chain for Groq&#8217;s chips is extremely resilient, even in the case of surging demand for AI chips, HBM, and advanced packaging processes. This idea isn&#8217;t new; in the past, <a href="https://en.wikipedia.org/wiki/Apple_A9">Apple has manufactured the same chip on different process nodes</a> to ensure they had sufficient capacity for the immense demand of new iPhone launches.</p><p>By acquiring Groq&#8217;s assets, Nvidia can now sell more AI chips than TSMC has CoWoS production capacity to make for them. They can sell more AI chips even if they&#8217;ve gone through their entire HBM allocation. Groq&#8217;s chips are worse than Nvidia&#8217;s own Blackwell B200s, but Nvidia is already going to be able to sell every B200 they can possibly produce. The LPUs mean that Nvidia can keep selling chips even after the limited B200 stock runs dry.</p><div class="pullquote"><p><strong>By acquiring Groq&#8217;s assets, Nvidia can now sell more AI chips than TSMC has CoWoS production capacity to make for them.</strong></p></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The 2025 Zach's Tech Blog Holiday Roundup]]></title><description><![CDATA[We made it through another year!]]></description><link>https://www.zach.be/p/the-2025-zachs-tech-blog-holiday</link><guid isPermaLink="false">https://www.zach.be/p/the-2025-zachs-tech-blog-holiday</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Wed, 24 Dec 2025 20:46:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!0QHW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Merry Christmas, Happy New Year, Happy Hanukkah, and enjoy all of your other holidays as well! In between <a href="https://x.com/blip_tm/status/2002187991914000499">my quest to find the best hot buttered rum in New York City</a> and traveling back to my hometown of <a href="https://www.zach.be/p/my-favorite-cocktails-in-los-angeles">Los Angeles</a>, I&#8217;ve decided to revive the Zach&#8217;s Tech Blog holiday tradition of doing <a href="https://www.zach.be/p/the-zachs-tech-blog-holiday-roundup?utm_source=publication-search">a little roundup of all of my favorite things I&#8217;ve written this year!</a></p><p>I started 2025 with about 700 subscribers. This year, we&#8217;ve more than doubled that, to over 1,600! We&#8217;ve had some hits as well, getting over 8,000 views on <a href="https://www.zach.be/p/why-is-it-so-hard-for-startups-to">my most popular article this year</a>. I&#8217;ve gotten coffee and drinks with multiple people I&#8217;ve met through this blog, I&#8217;ve been <a href="https://www.zach.be/p/making-unconventional-computing-practical">invited to speak at conferences</a>, and overall, I&#8217;ve had an amazing year writing all these articles. And I wouldn&#8217;t be able to do it if there weren&#8217;t people reading!</p><p><strong>So&#8230; thank you so much for reading for another year!</strong></p><p>I really appreciate everybody who reads these articles, likes them, comments, or reaches out to tell me they enjoyed something I wrote. I read it all (really!) and seeing smart, cool, interesting people appreciate the work I do warms my heart. If you do feel inspired to message me on <a href="https://x.com/blip_tm">Twitter</a> on <a href="https://www.linkedin.com/in/zachary-belateche-68b298a5/">LinkedIn</a>, feel free to say hi!</p><p>Now on to my favorite articles of the year!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>My Hottest Take</h1><p>Most people think that large language models (LLMs) are pretty bad at writing Verilog, the special language we use to design digital chips. Unlike high-level programming languages like Python or JavaScript, the performance of a piece of Verilog code is highly dependent on the quality of the Verilog code itself, and LLMs generally write mediocre Verilog at best. There&#8217;s a reason why most startups using LLMs for chip design are <a href="https://www.zach.be/p/why-did-cadence-acquire-chipstack?utm_source=publication-search">building products for verifying chips</a>, rather than writing Verilog.</p><p><a href="https://www.zach.be/p/yc-is-wrong-about-llms-for-chip-design">I definitely agree that LLMs are a very, very long way away from writing the kind of high-quality Verilog</a> that you&#8217;d find in the core mathematics engine of something like an Nvidia GPU. But most chips have a lot of subsystems that are acquired from third-party vendors, called &#8220;<a href="https://en.wikipedia.org/wiki/Semiconductor_intellectual_property_core">IP cores</a>&#8221;. A large, modern chip will have a substantial part of its development budget allocated to IP cores. And a lot of these IP cores can fulfill their functions even if their implementation is mediocre. So I made a prediction:</p><div><hr></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;2f04f9a0-4624-4b77-b26c-57eb01ee80fd&quot;,&quot;caption&quot;:&quot;I&#8217;ve been notably skeptical of LLMs&#8217; ability to design chips. Verilog, the programming language most commonly used for chip design, is unforgiving, with minor differences in a block of code resulting in significantly worse performance. It&#8217;s difficult to get your hands on enough training data to fine-tune an LLM to generate Verilog, and even if you do, t&#8230;&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;lg&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;LLMs will power next-gen chip IP companies.&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:80506581,&quot;name&quot;:&quot;zach&quot;,&quot;bio&quot;:&quot;building chips, writing words&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f07352f5-7f47-4538-8448-1ecef96b606f_712x712.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2025-03-04T14:58:00.397Z&quot;,&quot;cover_image&quot;:null,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.zach.be/p/llms-will-power-next-gen-chip-ip&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:158317083,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:15,&quot;comment_count&quot;:6,&quot;publication_id&quot;:775919,&quot;publication_name&quot;:&quot;zach's tech blog&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!Qv3s!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div><hr></div><p>We haven&#8217;t actually seen an LLM-powered IP company be founded yet, but <a href="https://www.spheresemi.com/">Sphere Semi</a> is basically doing this exact thing for analog chips and analog IP. They&#8217;re developing <a href="https://www.zach.be/p/rl-isnt-the-silver-bullet-for-ai?utm_source=publication-search">reinforcement-learning powered AI tools</a> for analog chip design, but instead of selling those tools, they&#8217;re selling chips and IP cores that they develop in-house using those tools.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/the-2025-zachs-tech-blog-holiday?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/the-2025-zachs-tech-blog-holiday?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/the-2025-zachs-tech-blog-holiday?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>My Most Popular Article</h1><p>Every chip designer hates the oligopoly that <a href="https://www.cadence.com/en_US/home.html">Cadence</a>, <a href="https://www.synopsys.com/">Synopsys</a>, and <a href="https://eda.sw.siemens.com/en-US/">Siemens</a> have on the specialized software we use to design chips. What&#8217;s surprising, though, is that nobody&#8217;s successfully disrupted this oligopoly; after all, Silicon Valley startups are really good at making software capable of disrupting legacy incumbents. Those in the know, however, are aware of the inter-company politics that keep startups from successfully selling chip design software.</p><div><hr></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;6271af36-7452-4ffe-94c5-d72a97991faf&quot;,&quot;caption&quot;:&quot;If you talk to any chip designer, they&#8217;ll complain about Cadence&#8217;s chip design software. They&#8217;ll probably complain about Synopsys too, and maybe Mentor Graphics (which is now owned by Siemens, but everybody still calls them Mentor Graphics). These three companies have an oligopoly on chip design software, despite their software being slow, difficult to &#8230;&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;md&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;Why is it so hard for startups to compete with Cadence?&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:80506581,&quot;name&quot;:&quot;zach&quot;,&quot;bio&quot;:&quot;building chips, writing words&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f07352f5-7f47-4538-8448-1ecef96b606f_712x712.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2025-07-01T14:40:24.104Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/$s_!ueCk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.zach.be/p/why-is-it-so-hard-for-startups-to&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:167271180,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:25,&quot;comment_count&quot;:9,&quot;publication_id&quot;:775919,&quot;publication_name&quot;:&quot;zach's tech blog&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!Qv3s!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div><hr></div><p>I think that a lot of non-chip-designers probably appreciated getting a peek into the weird world the politics behind the chip design software industry. I also tried to end on a hopeful note: there are some startups that might be able to disrupt the industry, and I&#8217;m rooting for them!</p><p>This article also got me a bit of flak because I suggested a political marriage of an EDA startup founder to Morris Chang&#8217;s daughter as a solution to the Cadence / Synopsys / Siemens oligopoly. Though when everybody is talking about how Taiwan is just like Arrakis from Dune, I don&#8217;t think a political marriage should be totally out of the question.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0QHW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0QHW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg 424w, https://substackcdn.com/image/fetch/$s_!0QHW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg 848w, https://substackcdn.com/image/fetch/$s_!0QHW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!0QHW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0QHW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg" width="487" height="482.6711111111111" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1115,&quot;width&quot;:1125,&quot;resizeWidth&quot;:487,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;The spice must flow &#128640;&#128200; : r/wallstreetbets&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="The spice must flow &#128640;&#128200; : r/wallstreetbets" title="The spice must flow &#128640;&#128200; : r/wallstreetbets" srcset="https://substackcdn.com/image/fetch/$s_!0QHW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg 424w, https://substackcdn.com/image/fetch/$s_!0QHW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg 848w, https://substackcdn.com/image/fetch/$s_!0QHW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!0QHW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dee31b-4b09-433c-ad7d-a9ec968386e0_1125x1115.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>a16z American Dynamism is holding auditions for Paul Atreides</em></figcaption></figure></div><h1>My Favorite Article</h1><p><a href="https://www.zach.be/p/the-zachs-tech-blog-holiday-roundup?utm_source=publication-search">Just like last year</a>, my favorite article is about <a href="https://www.zach.be/p/approximate-computing-for-secure">side-channel attacks on AI chips</a>. Secure AI hardware is a really interesting mix of the two biggest focuses I&#8217;ve had in my career and in my writing: <a href="https://www.normalcomputing.com/">AI accelerators</a>, and <a href="https://www.btq.com/">secure hardware design</a>. Unfortunately, a lot of people see secure and tamper-proof AI hardware as unnecessary. Those people are wrong:</p><div><hr></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;4b6211f0-3dd0-48a4-adbb-71a3a5092ab5&quot;,&quot;caption&quot;:&quot;I&#8217;ve talked a lot about side-channel attacks on cryptographic systems. These sorts of attacks usually focus on compromising low-cost, commodity cryptographic hardware, like the chips you&#8217;d find in a credit card or a YubiKey. But more recently, industry leaders in the AI world, including OpenAI, have been calling for more focus on&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;md&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;Side channel attacks on AI chips are very real.&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:80506581,&quot;name&quot;:&quot;zach&quot;,&quot;bio&quot;:&quot;building chips, writing words&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f07352f5-7f47-4538-8448-1ecef96b606f_712x712.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2025-01-21T15:17:05.110Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/$s_!_MV4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F85303e92-d455-4f52-9233-a9ba9279a5b9_611x555.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.zach.be/p/side-channel-attacks-on-ai-chips&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:155290482,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:13,&quot;comment_count&quot;:4,&quot;publication_id&quot;:775919,&quot;publication_name&quot;:&quot;zach's tech blog&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!Qv3s!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div><hr></div><p>The BarraCUDA attack is an extremely important proof-of-concept for side-channel attacks on real-world AI systems. If you want to deploy an AI model on an edge device like a security camera, a drone, or a self-driving car, you have to assume that anybody with physical access to the device can steal your model weights. And if they have your model weights, they can find adversarial examples that make your model go haywire. For example, they could <a href="https://arxiv.org/pdf/1707.08945">install special graffiti on stop signs that makes your self-driving model think it&#8217;s actually a different kind of sign</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sewm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sewm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png 424w, https://substackcdn.com/image/fetch/$s_!sewm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png 848w, https://substackcdn.com/image/fetch/$s_!sewm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png 1272w, https://substackcdn.com/image/fetch/$s_!sewm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sewm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png" width="424" height="345.25714285714287" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:513,&quot;width&quot;:630,&quot;resizeWidth&quot;:424,&quot;bytes&quot;:134603,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.zach.be/i/182530227?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sewm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png 424w, https://substackcdn.com/image/fetch/$s_!sewm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png 848w, https://substackcdn.com/image/fetch/$s_!sewm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png 1272w, https://substackcdn.com/image/fetch/$s_!sewm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48414112-d1db-4ada-809a-f4d3343c1d1b_630x513.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://arxiv.org/pdf/1707.08945">Eykholt et al, 2018</a></em></figcaption></figure></div><p>On that slightly terrifying note, that&#8217;s all for this year! Happy holidays, thanks for reading, and I&#8217;ll catch you in the new year with some new articles!</p><p><em>- Zach from Zach&#8217;s Tech Blog</em></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why did Irreducible fail?]]></title><description><![CDATA[The ZKP chip market and the AI chip market aren&#8217;t the same.]]></description><link>https://www.zach.be/p/why-did-irreducible-fail</link><guid isPermaLink="false">https://www.zach.be/p/why-did-irreducible-fail</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 01 Dec 2025 15:54:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last month, Irredudible, the company founded to use field programmable gate array chips (FPGAs) to accelerate zero-knowledge proofs (ZKPs) <a href="https://www.irreducible.com/posts/irreducible-shutting-down">shut down</a>. This came only a couple months after they <a href="https://www.irreducible.com/posts/reinventing-irreducible">announced a pivot away from FPGAs and into high-performance software implementations of ZKPs</a>.</p><p>This surprised a lot of people in the industry. <a href="https://www.paradigm.xyz/2022/04/zk-hardware">The conventional wisdom was that zero-knowledge proofs deserve custom hardware</a>, just like AI workloads do. Both AI and ZKPs require computers to perform a lot of complex heavy-duty computations, but the result of all those computations should be valuable enough to make custom hardware worthwhile. In the case of AI, that result is a computer you can talk to. In the case of zero-knowledge proofs, <a href="https://en.wikipedia.org/wiki/Zero-knowledge_proof">it&#8217;s a mathematical way to guarantee you did a computation, without actually revealing the results</a>.</p><p>This has a ton of valuable use cases. You can, for example, <a href="https://arxiv.org/abs/2506.20915">use a ZKP to prove that an AI model was trained with a specific approved dataset</a>, without any adversarial examples secretly baked in. But by far, the largest use-case for ZKPs is in cryptocurrency. <a href="https://en.wikipedia.org/wiki/Zcash">Zcash</a> uses ZKPs to create a fully anonymous blockchain. <a href="https://www.coinbase.com/learn/crypto-glossary/what-are-zero-knowledge-zk-rollups">Zero-knowledge rollups</a> can improve the efficiency of blockchains by doing some computation off-chain, and only storing the proofs on-chain.</p><p>However, in the past year or so, we&#8217;ve been seeing the companies working on custom ZKP hardware struggle. <a href="https://www.ingonyama.com/">Ignoyama</a> started focusing <a href="https://www.ingonyama.com/post/icicle-v3-going-multi-platform">purely on software, rather than FPGA hardware</a>, and their core libraries started to <a href="https://www.ingonyama.com/post/icicle-v4-lattice-based-cryptography-and-better-devex">hedge into high-performance implementations of post-quantum cryptography algorithms</a>, rather than focusing purely on ZKPs. <a href="https://www.fabriccryptography.com/">Fabric Cryptography&#8217;s</a> ZKP chip still hasn&#8217;t hit the market after multiple years of promises. Irreducible shut down.</p><p>So clearly, AI workloads and ZKPs aren&#8217;t the same. AI chip companies are succeeding, while ZKP hardware companies are struggling. What are the key differences?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>ZKPs are much more diverse</h1><p>Most AI workloads are fairly similar. They consist of a lot of floating point matrix multiplications, and then a smaller amount of <a href="https://en.wikipedia.org/wiki/Activation_function">nonlinear activation functions</a>. Some workloads might need 32-bit floating point numbers, while some might only need 8-bit floating point numbers, and some workloads may require <a href="https://en.wikipedia.org/wiki/Softmax_function">specific nonlinearities that are difficult to implement efficiently</a> &#8212; but for the most part, most AI workloads don&#8217;t look wildly different from one another.</p><p>Different ZKP algorithms, on the other hand, can be extremely different from one another. Some ZKPs, like Binius, <a href="https://vitalik.eth.limo/general/2024/04/29/binius.html">operate on binary tower fields using bitwise operations</a>. <a href="https://polygon.technology/blog/plonky2-a-deep-dive">Plonky2 requires number-theoretic transforms</a>, which require significant memory bandwidth. Groth16 requires less memory bandwidth, <a href="https://alinush.github.io/groth16">but operates on 381-bit integers</a>. And other ZKP schemes like Halo2 and Bulletproofs have their own computational bottlenecks.</p><p>If you&#8217;re building hardware, this is a huge problem. For example, <a href="https://www.fabriccryptography.com/">Fabric Cryptography&#8217;s VPU features 384-bit ALUs</a>, which will be totally underutilized when running the bit-sliced operations required by Binius. And importantly, ZKPs aren&#8217;t standardized the way that <a href="https://csrc.nist.gov/projects/post-quantum-cryptography">conventional cryptography algorithms are</a>. This put Irreducible in a tough spot. They could spend their time developing the best hardware in the world for one specific ZKP, but if a new ZKP suddenly becomes popular, they have to go back to the drawing board.</p><p>This is part of why Irreducible <a href="https://www.irreducible.com/posts/announcing-binius64">tried to advocate for Binius as a proving scheme</a>. If they encouraged everybody to stop trying out new ZKPs and pick one algorithm that was well-suited to Irreducible&#8217;s hardware, that could solve the problem of needing to constantly support new algorithms. But there are already so many different L2 rollups and privacy-focused blockchains that already have non-Binius ZKPs baked into their protocols. By the time Irreducible shut down, there were no publicly deployed, production L2 rollups or privacy chains that used Binius.</p><p>Notably, conventional cryptographic hardware accelerators don&#8217;t have this problem, because <a href="https://csrc.nist.gov/projects/post-quantum-cryptography">conventional cryptography is standardized</a>. Most cryptographic accelerators only accelerate a single algorithm; <a href="https://www.rambus.com/security/crypto-accelerator-cores/">Rambus sells a </a><em><a href="https://www.rambus.com/security/crypto-accelerator-cores/">separate</a></em><a href="https://www.rambus.com/security/crypto-accelerator-cores/"> SHA-2, SHA3, and AES engine</a>! In the world of conventional cryptography, flexibility and <a href="https://en.wikipedia.org/wiki/Cryptographic_agility">cryptographic agility</a> is a value-add; only the best solutions on the market, like <a href="https://www.btq.com/blog/processing-in-memory-and-the-future-of-hardware-security">BTQ&#8217;s QCIM</a>, offer support for multiple algorithms, and inferior inflexible solutions still maintain substantial market share. In the world of ZKPs, though, support for multiple wildly different algorithms is a baseline requirement that companies like Irreducible and Fabric failed to meet.</p><p>But that&#8217;s still only half the answer. It&#8217;s extremely difficult, but it&#8217;s definitely possible to build flexible cryptographic hardware. It&#8217;s possible to partner with specific L2s to get them to use your specific ZKP algorithm. And yet Irreducible still failed. Why?</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-irreducible-fail?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-irreducible-fail?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-did-irreducible-fail?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>The ZKP market is much smaller than the AI market</h1><p>It&#8217;s obvious that the generative AI market is huge, even if a significant portion of that market is cheating on high school essays and generating slop. The same simply isn&#8217;t true of zero-knowledge proofs</p><p>Zcash may have a <a href="https://coinmarketcap.com/currencies/zcash/">$7B market cap</a>, but the <a href="https://www.theblock.co/post/378232/zcash-shielded-pool-climbs-23-supply-network-usage-surges?utm_source=chatgpt.com">vast majority of Zcash transactions don&#8217;t even use the private zero-knowledge proof transaction features</a>. Scroll, one of the largest ZK-rollups, has a <a href="https://www.coindesk.com/tech/2025/04/23/scrolls-euclid-upgrade-pushes-it-into-stage-1-decentralization-era">total value locked that fluctuates between $100M and $1B</a>. That&#8217;s the same size as a large seed round for an AI startup.</p><p>Ultimately, <em>most cryptocurrency users don&#8217;t care much about the privacy features ZKPs offer.</em> As the cryptocurrency industry has evolved, it&#8217;s become less focused cryptocurrency as a means of exchange, and more on cryptocurrency as an investment. And as large institutions invest in cryptocurrency, this only becomes more true.</p><p>Now, this isn&#8217;t necessarily a bad thing! More institutional investment activity could reduce volatility and help make cryptocurrency into a reliable part of a complete investment portfolio. But zero-knowledge proofs aren&#8217;t a key technology to enable that. The key features that ZKPs enable are private transactions and more efficient rollups with significantly lower transaction fees. These are valuable in the use-cases where cryptocurrency is transacted as a means of exchange, not used as an investment.</p><p>Again, this isn&#8217;t something that&#8217;s true of conventional cryptography hardware. There are over <a href="https://globalplatform.org/wp-content/uploads/2018/05/Introduction-to-Secure-Element-15May2018.pdf">50 billion secure elements in the world</a>, and that doesn&#8217;t include the cryptographic hardware accelerators inside of nearly every <a href="https://support.microsoft.com/en-us/windows/enable-tpm-2-0-on-your-pc-1fd5a332-360d-4f46-a1e7-ae6b0c90645c">laptop, desktop</a>, and <a href="https://support.apple.com/guide/security/secure-enclave-sec59b0b31ff/web">smartphone SoC</a> on the planet. Cryptography is necessary to guarantee basic privacy and security, so secure cryptography hardware is a requirement to comply with nearly every privacy-related regulation from any government in the world, creating massive demand.</p><p>Ironically, it&#8217;s the decentralized nature of cryptocurrency that reduces the demand for ZKP hardware. Investors care much more about cryptocurrency as an investment, rather than its ability to enable efficient or private transactions. And because there&#8217;s no central authority to enforce privacy regulations, ZKPs fall by the wayside. It&#8217;s unfortunate, because Irreducible&#8217;s technology was great&#8230; the crypto community just didn&#8217;t really want it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why did Cadence acquire ChipStack?]]></title><description><![CDATA[And what does it mean for other AI EDA startups?]]></description><link>https://www.zach.be/p/why-did-cadence-acquire-chipstack</link><guid isPermaLink="false">https://www.zach.be/p/why-did-cadence-acquire-chipstack</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Tue, 11 Nov 2025 17:00:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This week, Cadence Design Systems, one of the big 3 electronic design automation (EDA) companies that dominate the market for the software that engineers use to design microchips, <a href="https://www.chipstack.ai/blog/chipstack-is-joining-cadence">acquired a small startup called ChipStack</a>. ChipStack uses AI to speed up silicon verification, which is the process of testing chip designs in a simulation before sending them off to be manufactured. Cadence didn&#8217;t announce how much they spent buying ChipStack, which usually means it was a somewhat small acquisition &#8212; in the range of tens to hundreds of millions of dollars. But the implications for other startups developing AI-powered EDA software is massive.</p><p>For those of us who know the EDA industry well, this isn&#8217;t a surprise. As a matter of fact, <a href="https://www.zach.be/p/why-is-it-so-hard-for-startups-to">this is the standard modus operandi of the EDA industry</a>. Startups innovate, get acquired for $50-500M by the big players, and then get folded into their product portfolio. It&#8217;s much easier to sell EDA software when it&#8217;s bundled with existing Cadence products rather <a href="https://www.zach.be/p/llms-will-power-next-gen-chip-ip">than packaged as a standalone product</a>. That just leaves us to wonder: why did Cadence acquire ChipStack specifically? And what does this mean for all the other startups in the space?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Why ChipStack?</h1><p>From the outside, ChipStack doesn&#8217;t look all that different from their competitors, like <a href="https://chipagents.ai/">ChipAgents</a> and <a href="https://www.bronco.ai/">Bronco</a>. However, if you dig in a bit deeper, the differences get clearer. <a href="https://www.businesswire.com/news/home/20251021677325/en/ChipAgents-Raises-Oversubscribed-%2421M-Series-A-to-Redefine-AI-for-Chip-Design">ChipAgents just raised a $21M Series A</a>, while <a href="https://www.geekwire.com/2025/cadence-to-acquire-seattle-startup-chipstack-to-boost-chip-design-automation/">ChipStack has only raised about $7M to date</a> &#8212; that means that ChipStack&#8217;s valuation is lower, so they could be acquired for a much lower price. Bronco, on the other hand, <a href="https://www.ycombinator.com/companies/bronco-ai">is run by college dropouts</a>, and probably not a great acquisition target for a large, corporate company like Cadence.</p><p>A lot of the other AI-powered verification startups have similar stories; <a href="https://www.businesswire.com/news/home/20250515702098/en/Cognichip-Launches-out-of-Stealth-with-%2433M-in-Seed-Funding-to-Deliver-Artificial-Chip-Intelligence-ACI">Cognichip raised $33M</a>, so they probably would have been a lot more expensive to acquire than ChipStack. Normal Computing<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> offers an <a href="https://www.normalcomputing.com/solutions">AI-powered verification product</a>, but also develops <a href="https://www.zach.be/p/scaling-thermodynamic-computing">thermodynamic computing chips</a>, which are totally outside of Cadence&#8217;s wheelhouse. Ultimately, ChipStack was probably the most reasonably priced acquisition target that would fit well within Cadence as a parent company.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-cadence-acquire-chipstack?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-cadence-acquire-chipstack?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-did-cadence-acquire-chipstack?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>What about other AI chip design companies?</h1><p>For AI-powered chip verification companies building direct competitors to ChipStack, like <a href="https://chipagents.ai/">ChipAgents</a> and <a href="https://www.bronco.ai/">Bronco</a>, this acquisition is both a blessing and a curse. With Cadence clearly planting its flag in AI-powered verification software, the pressure is on for Cadence&#8217;s two biggest competitors, Synopsys and Siemens EDA, to make a move. This could mean that ChipStack&#8217;s competitors acquire their own AI-powered verification startups over the next few months.</p><p>However, there are a bunch of different AI-powered verification startups out there, and <a href="https://www.zach.be/p/why-is-it-so-hard-for-startups-to">the EDA industry is an oligopoly</a>. After Cadence, Synopsys, and Siemens all acquire AI-powered verification capabilities, there&#8217;s no other large player who could acquire any remaining competitors to ChipStack. That means that all those other startups either need to make it to IPO or get <a href="https://en.wikipedia.org/wiki/Acqui-hiring">acqui-hired</a> for far less than they&#8217;re worth &#8212; or they&#8217;ll go out of business.</p><p>Not all of the AI-powered chip design startups are competitors to ChipStack, though. <a href="https://www.silimate.com/">Silimate</a>, for example, is building an AI-powered tool to predict the performance of chip designs earlier on in the design process. <a href="https://silogy.io/">Silogy</a> is building an AI agent to debug test failures, rather than just writing tests using LLMs like ChipStack. These companies are in a much better position following this acquisition, because it signals that potential exit opportunities are opening up, without closing the door to a potential acquisition by Cadence.</p><p>Ultimately, the ChipStack acquisition is likely the first in a series of M&amp;A announcements we&#8217;ll hear over the next couple years, as the EDA oligopoly gobbles up the best deals on AI-powered EDA startups. And unfortunately, the companies that don&#8217;t get picked by the end of it may end up dying.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p><em>Disclosure: I used to work at Normal Computing as their silicon engineering lead. I am still an advisor there.</em></p></div></div>]]></content:encoded></item><item><title><![CDATA[So, I have to talk about Extropic.]]></title><description><![CDATA[It&#8217;s probabilistically nothing, right?]]></description><link>https://www.zach.be/p/so-i-have-to-talk-about-extropic</link><guid isPermaLink="false">https://www.zach.be/p/so-i-have-to-talk-about-extropic</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Tue, 04 Nov 2025 20:39:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Q-9w!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Disclosure before I jump into the article: I used to lead the silicon team at Normal Computing, a thermodynamic computing company that is considered a competitor to Extropic. I don&#8217;t work there anymore, but I&#8217;m still an advisor there.</em></p><div><hr></div><p>After years of doubt and claims that <a href="https://extropic.ai/">Extropic</a> was a total grift, Guillaume Verdon and his crew finally announced their first product, <a href="https://extropic.ai/writing/inside-x0-and-xtr-0">their XTR-0 devkit</a>. A lot of people are claiming that this vindicates all of Extropic&#8217;s claims and &#8220;proves the haters wrong&#8221;. But that&#8217;s not entirely true.</p><p>Yes, the folks who claimed that Extropic was a total scam that would die without ever shipping a piece of hardware were wrong. Extropic did design a chip that is capable of running certain small ML workloads with impressive efficiency. That&#8217;s cool, <a href="https://x.com/tenderizzation/status/1983898136373559709">but also most PhDs in chip design do at least one or two small ML chips as part of their graduate research</a>. Building an incredibly efficient chip that only works on MNIST is a bit of a punchline at this point.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q-9w!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q-9w!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png 424w, https://substackcdn.com/image/fetch/$s_!Q-9w!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png 848w, https://substackcdn.com/image/fetch/$s_!Q-9w!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png 1272w, https://substackcdn.com/image/fetch/$s_!Q-9w!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q-9w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png" width="226" height="274.86486486486484" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:740,&quot;resizeWidth&quot;:226,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Q-9w!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png 424w, https://substackcdn.com/image/fetch/$s_!Q-9w!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png 848w, https://substackcdn.com/image/fetch/$s_!Q-9w!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png 1272w, https://substackcdn.com/image/fetch/$s_!Q-9w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9dbcc62-6d0f-4d24-b826-16c0be063b9c_740x900.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Also, everything Extropic is doing is building off a large existing body of work on p-bits and probabilistic computing. I&#8217;ve been working in the unconventional computing space most of my career, and I&#8217;ve been plugged into what&#8217;s actually been going on in the world of p-bits, thermodynamic computers, and Ising machines. So today, we&#8217;re going to be talking about what Extropic is doing, the current state of the p-bit research landscape, and their novel contributions.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>The Current p-bit State of the Art</h1><p>Extropic makes one claim early on in their paper that I think is extremely unfair:</p><blockquote><p>Additionally, existing devices have relied on exotic components such as magnetic tunnel junctions as sources of intense thermal noise for random-number generation (RNG). These exotic components have not yet been tightly integrated with transistors in commercial CMOS processes and do not currently constitute a scalable solution.</p></blockquote><p>Put simply, this is false. <a href="https://arxiv.org/abs/2303.10728">Most state-of-the-art work on p-bits is implemented using digital logic gates</a>, which is highly scalable on FPGAs and commercial CMOS processes. These architectures use digital pseudorandom number generators (PRNGs), which offer an efficient way to generate numbers with sufficient randomness without relying on exotic materials, with the added bonus of reproducibility due to their pseudorandom nature.</p><p>There are also a large number of analog architectures out there that are scalable, manufacturable, and don&#8217;t use exotic technologies. Not only have researchers demonstrated analog p-bits <a href="https://www.nature.com/articles/s41598-025-93218-8">multiple</a> <a href="https://www.arxiv.org/pdf/2008.10578v3">times</a> over, but there&#8217;s <a href="https://arxiv.org/pdf/2508.15234">an equivalence</a> between <a href="https://www.nature.com/articles/s41467-022-33441-3">noise-injected Ising machines</a> and <a href="https://www.zach.be/p/boltzmann-sampling-on-coupled-oscillator">p-bit samplers</a>. That means that all of the analog Ising machine designs, from <a href="https://www.nature.com/articles/s41928-025-01393-3.epdf">ring oscillators</a> to <a href="https://ieeexplore.ieee.org/document/10558647">cross-coupled latches</a>, can also act as p-bits -- and they&#8217;re all manufactured in commercial CMOS processes.</p><p>Hell, if we&#8217;re less strict and include non-p-bit sampling architectures, <a href="https://link.springer.com/chapter/10.1007/978-1-4899-1331-9_4">analog</a> and <a href="https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=242756">digital CMOS implementations</a> of Boltzmann machines have been around since the 1990s.</p><p>Extropic also proposes other probabilistic bit structures that mirror existing work in literature. Extropic&#8217;s <a href="https://extropic.ai/hardware">p-dits and p-modes</a> are similar to current state-of-the-art work on <a href="https://arxiv.org/abs/2506.00269">p-dits</a> and <a href="https://arxiv.org/pdf/2312.04836">programmable Gaussian distributions</a>. To their credit, though, existing work on p-ints and p-dits are digital; if Exropic&#8217;s design is all-analog, that would represent a novel design for that specific probabilistic primitive.</p><p>Overall, there are a large number of competing designs for p-bits, all with different tradeoffs. The digital designs are often slightly larger and more power hungry, but are much easier to design, more scalable, and don&#8217;t suffer from issues due to process variability that plague analog designs. Analog designs may be smaller and more efficient, but often struggle to implement high-precision energy functions and may make computational errors due to process variations and parasitic effects introducing spurious biases and noise terms.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TmQu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TmQu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png 424w, https://substackcdn.com/image/fetch/$s_!TmQu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png 848w, https://substackcdn.com/image/fetch/$s_!TmQu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png 1272w, https://substackcdn.com/image/fetch/$s_!TmQu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TmQu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png" width="550" height="317.86570743405275" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a2db71ca-3a61-458d-8903-3158c9835718_834x482.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:482,&quot;width&quot;:834,&quot;resizeWidth&quot;:550,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TmQu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png 424w, https://substackcdn.com/image/fetch/$s_!TmQu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png 848w, https://substackcdn.com/image/fetch/$s_!TmQu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png 1272w, https://substackcdn.com/image/fetch/$s_!TmQu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2db71ca-3a61-458d-8903-3158c9835718_834x482.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://www.nature.com/articles/s41598-025-93218-8">Process variability can affect p-bit sampling</a></em></figcaption></figure></div><p>It&#8217;s unclear how much of Extropic&#8217;s architecture is analog and how much is digital. It&#8217;s also hard to tell how, if at all, they&#8217;re addressing issues of process variations in their design. Ultimately, these are the practical issues that plague most analog implementations of Boltzmann machines, and represent the reason why digital implementations are currently dominant in the literature. We&#8217;ll see if Extropic has a novel solution to this problem or not.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/so-i-have-to-talk-about-extropic?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/so-i-have-to-talk-about-extropic?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/so-i-have-to-talk-about-extropic?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>Comparing to GPUs</h1><p>Extropic also makes the bold claim that their chips are 10,000x more efficient than a GPU. Unfortunately, these headline numbers are fairly rough estimates for both their hardware, and for GPUs. Because Extropic&#8217;s target task is so simple, there isn&#8217;t an optimized GPU state-of-the-art model to compare it to. Instead, their GPU estimates measure the power consumption of very small models on large, powerful hardware; this likely results in extremely poor utilization of the GPU and a lot of wasted energy.</p><p>At the same time, their analysis of their hardware is fairly optimistic. They only consider the energy consumed by their RNG, resistor network, clock tree, and local wire capacitance. The Extropic team does admit that this is a very rough estimate, but also raises one big question: if this chip actually implements a full Boltzmann machine, why not just measure its power consumption and report it? It could be the case that Extropic&#8217;s test chip has p-bit test structures, but lacks the resistor network between them to implement an entire Boltzmann machine on-chip.</p><p>One other major concern of mine is that <a href="https://arxiv.org/pdf/2510.23972">Extropic&#8217;s hardware implements bipartite graphs</a>. Not only does this limit your hardware to natively accelerating sparse models, but it also makes your hardware a lot less competitive against GPUs.</p><p>The biggest reason why Gibbs sampling is really slow on GPUs is its serial nature. If you try to update multiple states at the same time, you run into weird issues where <a href="https://arxiv.org/abs/2110.02481">&#8220;frustrated&#8221; connections start to cause oscillations</a> in the system state and errors in the result. <a href="https://www.gatsby.ucl.ac.uk/~gretton/papers/GonLowGreGue11.pdf">This is bad for GPUs</a>, which are natively parallel processing units. But when you start working with bipartite graphs or other sparse graphs, this problem gets a lot easier, and <a href="https://www.gatsby.ucl.ac.uk/~gretton/papers/GonLowGreGue11.pdf">you can update all the nodes in each half of the graph in parallel</a>. <a href="https://en.wikipedia.org/wiki/Restricted_Boltzmann_machine">Restricted Boltzmann Machines</a>, a kind of EBM that relies on sparse graphs, <a href="https://robotics.stanford.edu/~ang/papers/icml09-LargeScaleUnsupervisedDeepLearningGPU.pdf">were partially popularized because they can run so much more quickly on GPUs!</a> If you&#8217;re only focusing on bipartite graphs, GPUs are actually not all that bad at sampling.</p><h1>So, what&#8217;s new here?</h1><p>Even though Extropic&#8217;s paper makes some questionable comments about the state-of-the-art, they do make two novel contributions to p-bit research. Specifically, they propose a new kind of energy-based model, the Denoising Thermodynamic Model, or DTM, with potentially greater stability and efficiency than conventional EBMs. They also propose a more efficient RNG for their p-bits.</p><p>Conventional monolithic EBMs draw samples from a single distribution, often parameterized by a neural network. Extropic correctly points out that this often results in distributions that are extremely hard to sample from. Their DTMs instead perform multiple de-noising steps, similar to a diffusion model, where each step is relatively easier to sample from. This idea isn&#8217;t totally novel; it&#8217;s similar to other techniques to improve sampling, like <a href="https://en.wikipedia.org/wiki/Diffusion_model">conventional diffusion models</a> and <a href="https://www.nature.com/articles/ncomms12999?utm_source=chatgpt.com">counterdiabatic driving</a>, but it&#8217;s optimized for the specific properties of Extropic&#8217;s p-bit computer.</p><p>Their p-bits also use a novel RNG architecture. Current digital implementations of p-bits currently use digital PRNGs combined with LUTs to create a stochastic sigmoid activation function. The most recent, non-state-of-the-art result I found uses <a href="https://escholarship.org/content/qt7n5016w1/qt7n5016w1_noSplash_2d66d9d76fa8fc4652b489f3b78a21d2.pdf">42 look-up tables and 33 registers in an FPGA</a> to represent the p-bit PRNG and activation function. Depending on the process node, an ASIC implementation of the same logic could take up hundreds of square microns of silicon area. <a href="https://arxiv.org/pdf/2510.23972">Extropic&#8217;s RNG is only nine square microns</a>, which is a meaningful improvement!</p><p>However, it&#8217;s unclear how robust the sigmoid transfer function of their RNG is to process variations, as the only information on process variation provided relates to the energy and lag time. One of the major advantages of the digital PRNG designs is that they get to ignore mismatch entirely. Also, digital PRNG designs are much, much faster than Extropic&#8217;s analog RNG. Extropic&#8217;s RNG produces random bits at a rate of 10MHz, while digital PRNGs can often hit frequencies approaching 1GHz. Hopefully Extropic&#8217;s upcoming physical review letters paper sheds some additional light here.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>My take on Extropic</h1><p>Extropic is doing some genuinely interesting research on p-bits and energy-based models. If their RNGs are actually more efficient than the PRNGs used by existing p-bit architectures, that could offer a meaningful improvement in the number of p-bits you can fit on a single chip, as well as their power consumption.</p><p>Also, their results seem to genuinely show that their new Denoising Thermodynamic Models genuinely out-perform more naive implementations of EBMs. That&#8217;s genuine research progress! And if their new models are also more performant than conventional diffusion models in a more fair head-to-head comparison with a GPU, that could be an interesting first step towards scaling up EBMs on probabilistic computers.</p><p>Ultimately, though, the idea of p-bits isn&#8217;t novel. The idea of analog p-bits isn&#8217;t novel. Running EBMs using p-bits isn&#8217;t novel. Extropic is specifically proposing a new p-bit RNG and a new kind of EBM that could improve the performance and scalability of probabilistic computers. It&#8217;s interesting research progress, and would make for a very impressive PhD thesis or two, but it&#8217;s not &#8220;climbing the Kardashev scale through worship of the Thermodynamic God&#8221;, or whatever Gill has been tweeting about these days.</p><p>I will give Extropic one thing, though. The limiting factor for p-bit adoption is the lack of interested users and worthwhile algorithms. The sheer amount of hype and interest they&#8217;ve generated may bring p-bits from an obscure academic interest for unconventional computing nerds, to something that a much larger crowd of researchers are interested in tinkering with. There&#8217;s a chance that turns into a real algorithmic breakthrough that could enable their hardware to run more meaningful, large-scale workloads than just Fasion-MNIST.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/so-i-have-to-talk-about-extropic?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/so-i-have-to-talk-about-extropic?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/so-i-have-to-talk-about-extropic?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div>]]></content:encoded></item><item><title><![CDATA[Making Unconventional Computing Practical]]></title><description><![CDATA[Lessons learned from building weird chips for years.]]></description><link>https://www.zach.be/p/making-unconventional-computing-practical</link><guid isPermaLink="false">https://www.zach.be/p/making-unconventional-computing-practical</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Wed, 22 Oct 2025 15:22:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!lSNK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This blog post is based off a talk that I&#8217;ve given a couple times: first at <a href="https://homebrew.nyc/">Homebrew</a>, then at <a href="https://www.betaworks.com/">Betaworks</a>, and finally at <a href="https://foresitelabs.com/">Foresite Labs&#8217;</a> Reinventing the Semiconductor Future Symposium this past week. It&#8217;s an interesting topic -- if unconventional computing is <em>really</em> so unconventional, how does it have any chance of becoming practical or mainstream?</p><p>Well, unconventional computing is an umbrella term. It can range from the extremely weird, like <a href="https://www.mediamatic.net/en/page/16716/slime-mold-computing">slime mold computing</a>, to the relatively mainstream, like processing-in-memory. Broadly, unconventional computing refers to all methods of doing computation that differ from the <a href="https://en.wikipedia.org/wiki/Von_Neumann_architecture">standard Von Neumann architecture</a>, which forms the basis of modern CPUs and, to some degree, GPUs. For decades, the academic community has done research on novel methods of computing, but recently, driven by the end of <a href="https://en.wikipedia.org/wiki/Moore%27s_law">Moore&#8217;s Law transistor scaling</a> and the rising computational demands of AI, some startups and large corporations have been trying to commercialize these unconventional computing technologies.</p><p>I&#8217;ve worked at some of those startups, and I know the stories of many more. It turns out that taking research ideas, which are designed to maximize novelty and land in impressive journals, and trying to commercialize them is difficult. This blog post, like my talk, is about the lessons learned from trying to move three different unconventional computing ideas out of the lab and into industry.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Processing-in-Memory</h1><p>When I first encountered processing-in-memory, the dream was simple, but exciting. If we could store data as analog values in a resistive crossbar, we could perform entire vector-matrix multiplications at once. It was incredibly alluring. This sort of architecture would offer incredibly dense data storage, significantly reduce data movement, and take an iterative algorithm and reduce it to a single atomic operation.</p><p>There were a couple companies trying to commercialize this sort of architecture. The most well known is <a href="https://mythic.ai/">Mythic AI</a>, which used flash memory cells as the resistive element in their crossbar. Another startup, <a href="https://www.syntiant.com/">Syntiant</a>, proposed doing a similar thing, but was targeting much lower-power applications. Then a large number of research labs were working on using novel resistive memory devices, like <a href="https://en.wikipedia.org/wiki/Resistive_random-access_memory">resistive RAM (RRAM)</a> and <a href="https://en.wikipedia.org/wiki/Magnetoresistive_RAM">magnetoresistive RAM (MRAM)</a>, to implement these crossbars. RRAM and MRAM are a lot less mature than flash memory, but offer increased linearity and density.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lSNK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lSNK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png 424w, https://substackcdn.com/image/fetch/$s_!lSNK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png 848w, https://substackcdn.com/image/fetch/$s_!lSNK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png 1272w, https://substackcdn.com/image/fetch/$s_!lSNK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lSNK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png" width="514" height="599" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:599,&quot;width&quot;:514,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:90063,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lSNK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png 424w, https://substackcdn.com/image/fetch/$s_!lSNK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png 848w, https://substackcdn.com/image/fetch/$s_!lSNK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png 1272w, https://substackcdn.com/image/fetch/$s_!lSNK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a717660-d2e0-45ab-80a5-bdf977e73386_514x599.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>(source: WikiChip)</em></figcaption></figure></div><p>Unfortunately, that original alluring PIM dream is dead. <a href="https://www.theregister.com/2022/11/09/mythic_analog_ai_chips/">Mythic famously nearly ran out of capital</a> before <a href="https://techcrunch.com/2023/03/09/ai-chip-startup-mythic-rises-from-the-ashes-with-13m-new-ceo/">raising a small down round and replacing their CEO</a>. Syntiant <a href="https://www.eenewseurope.com/en/ceo-interview-with-kurt-busch-of-always-on-startup-syntiant/">quietly dropped their analog architecture in favor of a conventional digital chip</a>, which has been fairly successful. Many of the research labs working on RRAM and MRAM have struggled to get their chips to scale past simple image classification workloads, owing to device-level mismatch and variability over process, voltage and temperature variations.</p><p>&#8220;But Zach&#8221;, you might object, &#8220;haven&#8217;t there been successful processing-in-memory companies?&#8221; And the answer is actually, yes! My previous startup, Radical Semiconductor, built best-in-class cryptography accelerators using PIM technology. We were <a href="https://www.prnewswire.com/news-releases/btq-technologies-completes-acquisition-of-radical-semiconductors-processing-in-memory-technology-portfolio-advancing-post-quantum-cryptography-capabilities-302256705.html">acquired by BTQ</a>, where our architecture forms the basis of <a href="https://www.btq.com/products/qcim">BTQ&#8217;s QCIM offering</a>. <a href="https://www.d-matrix.ai/">d-Matrix</a> is selling fast and efficient LLM accelerators using PIM technology as well. But neither BTQ nor d-Matrix are working on that original analog PIM dream.</p><p>Analog PIM technology offers a number of advantages: reduced data movement, dense storage, and one-shot vector-matrix multiplication. But those advantages are not created equal: just reducing data movement offers a massive performance and efficiency boost on its own. Doing an 8-bit addition operation consumes about 0.3pJ of energy, but fetching data from memory costs upwards of 10pJ. If we develop PIM technology that gives up on multi-level memories and analog one-shot operation, we can still get a 10-100x performance improvement just by reducing data movement.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZFa9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZFa9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png 424w, https://substackcdn.com/image/fetch/$s_!ZFa9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png 848w, https://substackcdn.com/image/fetch/$s_!ZFa9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png 1272w, https://substackcdn.com/image/fetch/$s_!ZFa9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZFa9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png" width="694" height="233" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4c85a753-deed-45e7-9904-e24e49511714_694x233.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:233,&quot;width&quot;:694,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZFa9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png 424w, https://substackcdn.com/image/fetch/$s_!ZFa9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png 848w, https://substackcdn.com/image/fetch/$s_!ZFa9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png 1272w, https://substackcdn.com/image/fetch/$s_!ZFa9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c85a753-deed-45e7-9904-e24e49511714_694x233.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption"><em>(from Dr. Mark Horowitz, Stanford)</em></figcaption></figure></div><p>d-Matrix is developing digital processing-in-memory for matrix-vector operations; by integrating custom digital adder trees directly into SRAM memory arrays, they can still get huge performance advantages. The memory isn&#8217;t nearly as dense as the analog arrays proposed by Mythic or the RRAM labs, but they still reduce data movement massively. BTQ&#8217;s QCIM focuses on the vector-vector operations at the heart of the number theoretic transform, which powers NIST standard cryptographic algorithms like ML-KEM and ML-DSA.</p><p>By narrowly focusing on the biggest advantage PIM offers -- reducing data movement -- and ignoring the secondary benefits of all-analog architectures, startups have built legitimately successful products. And the story is pretty similar for another field that&#8217;s traditionally used analog architecture: neuromorphic computing.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/making-unconventional-computing-practical?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/making-unconventional-computing-practical?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/making-unconventional-computing-practical?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>Neuromorphic Computing</h1><p>The original dream of neuromorphic computing, cooked up by <a href="https://en.wikipedia.org/wiki/Carver_Mead">Carver Mead</a> in the 1980s, was to build ultra-efficient analog circuits that operated similarly to neurons in the brain. Neurons communicate with electrical pulses, called spikes, that are integrated as charge across the cell membrane of the neuron. The most traditional neuromorphic systems operate the same way, with spikes of voltages integrated onto capacitors inside of silicon neurons. Some neuromorphic systems, like <a href="https://en.wikipedia.org/wiki/Neurogrid">Neurogrid</a>, take this to an extreme, attempting to model all of the complex ion channels in human neurons as accurately as possible using silicon.</p><p>But when researchers tried to use these neuromorphic computers to accelerate meaningful machine learning models, they hit a ton of roadblocks. Because neuromorphic architectures use analog circuits, they struggle with <a href="https://web.stanford.edu/group/brainsinsilicon/documents/ReidPinT_DATE.pdf">temperature changes</a> and <a href="https://web.stanford.edu/group/brainsinsilicon/documents/abrams1.pdf">manufacturing variability</a>. But also, the spiking neural networks, or SNNs, native to neuromorphic computers aren&#8217;t compatible with conventional machine learning techniques. Gradient-based training methods, like backpropagation, <a href="https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=10242251">don&#8217;t support the non-differentiable spiking activations inside of SNNs</a>. This makes commercialization of neuromorphic computing technologies difficult. Customers want hardware that can run the models they already know and love.</p><p>There are some successful efforts in the realm of neuromorphic computing, though! But just like d-Matrix and BTQ did for processing-in-memory, they focus on achieving the key advantages of neuromorphic computing in the most practical way. In the case of neuromorphic computing, that advantage is <em>sparsity.</em> Neuromorphic computing systems are so efficient because they operate using sparse trains of spikes, rather than dense vectors of activations. By focusing on sparsity specifically, some companies have been able to build commercially viable neuromorphic-inspired chips.</p><p>The team at <a href="https://femto.ai/">femtoAI</a> (formerly Femtosense) are building neuromorphic-inspired ultra-power-efficient edge AI processors tailored to sparse networks. However, their chips use digital rather than analog circuits, and operate on sparse conventional neural networks, rather than the esoteric and unwieldy SNNs used by more traditional neuromorphic systems. Their focus on the most impactful benefit of neuromorphic computing ultimately allowed them to get meaningful commercial traction in the <a href="https://www.eenewseurope.com/en/partnership-brings-energy-saving-ai-technology-to-home-appliances/">smart home</a> and <a href="https://hearingreview.com/hearing-products/hearing-aids/otc/newsound-ai-driven-otc-hearing-aids-feature-femtosense-technology">hearing aid</a> markets.</p><h1>Thermodynamic / Probabilistic Computing</h1><p>Last but not least is thermodynamic and probabilistic computing, two related fields both focusing on leveraging specialized architectures to accelerate stochastic differential equations (SDEs) and stochastic random sampling workloads. I ran the silicon team at <a href="https://www.normalcomputing.com/">Normal Computing</a> for a year, and before that worked on <a href="https://www.zach.be/p/solving-optimization-problems-with">DIMPLE, the largest open-source fully-connected Ising machine ever demonstrated</a> -- so I know the space pretty intimately.</p><p>The story of thermodynamic and probabilistic computing starts off similarly to that of processing-in-memory and neuromorphic computing: a big dream and a lot of cool ideas for new circuits and new materials. <a href="https://www.purdue.edu/p-bit/">Supriyo Datta&#8217;s group at Purdue</a> proposed building probabilistic bits (p-bits) using stochastic magnetic tunnel junctions, which required new materials and fab processes. Normal Computing proposed <a href="https://arxiv.org/abs/2302.06584">analog circuits with thermal noise</a> to act as unit cells that could be coupled together. And <a href="https://www.zach.be/p/whats-the-difference-between-extropic">Extropic proposed superconducting circuits</a> to implement their p-bits. All of these were exciting ideas. None were particularly practical.</p><p>Currently, the largest published result demonstrating a p-bit system comes from <a href="https://opus.ece.ucsb.edu/">Karem Camsari&#8217;s lab</a>. They don&#8217;t use magnetic tunnel junctions or analog circuits. Instead, they use a digital random number generator to generate noise, and digital logic to implement couplings between p-bits. Normal Computing is <a href="https://www.zach.be/p/scaling-thermodynamic-computing">also leveraging a digital architecture</a> to accelerate the SDEs at the heart of diffusion models. And Extropic has ditched their esoteric superconductors in favor of mixed-signal CMOS circuits.</p><p>It took a little while, but so far, it&#8217;s seemed like the innovators in the world of thermodynamic computing have learned the lessons from other kinds of unconventional computing. Often, analog circuits are expensive and difficult to implement, even though they may be elegant and exciting. Digital architectures are reliable, efficient, and easy to manufacture and sell at scale. And so when unconventional computing technologies move from academic labs into industry, which cares less about publishing novel papers and more about selling chips, scalable digital architectures designed to maximize the key benefits of a new technology, while avoiding its biggest downsides, often win out.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[RL isn’t the silver bullet for AI-powered chip design]]></title><description><![CDATA[But new EDA tools could change that&#8230;]]></description><link>https://www.zach.be/p/rl-isnt-the-silver-bullet-for-ai</link><guid isPermaLink="false">https://www.zach.be/p/rl-isnt-the-silver-bullet-for-ai</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Thu, 25 Sep 2025 15:56:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>My most popular post of all time was about <a href="https://www.zach.be/p/yc-is-wrong-about-llms-for-chip-design">why Y Combinator is wrong about LLMs for chip design</a>. Put simply, I don&#8217;t think LLMs are able to generate high-performance or high-efficiency chip designs, because the process of designing high-performance chips is incredibly unforgiving. The entire &#8220;vibe coding&#8221; trend is about generating a large quantity of mediocre code, while the challenge of chip design is writing a relatively small amount of extremely high-quality, extremely performance-sensitive code in a specialized language called Verilog. And so far, I&#8217;ve been proven right. By and large, LLMs kinda suck at writing Verilog.</p><p>Never fear, though! Reinforcement learning (aka RL) will come to the rescue, right? RL has proven invaluable when it comes to making normally error-prone AI agents solve unforgiving tasks, from playing board games to constructing complex mathematical proofs. Why can&#8217;t we leverage RL systems to train AI models to write the sort of high-quality, performance-sensitive Verilog that is required to deliver high-performance chips?</p><p>Well, it turns out that it&#8217;s really, really hard, but not for the reasons you might think. Put simply, existing chip design tools are just too slow to allow any RL system to learn anything useful in a reasonable amount of time.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>How would RL help?</h1><p>Reinforcement learning is a training methodology for AI models that lets them improve at a specific task through trial-and-error. This is particularly useful for tasks that are not well-represented in the training set, but can be quick for a model to try and fail at. A great example is mathematical proofs. Google&#8217;s AlphaProof was able to learn to solve IMO math problems by repeatedly trying solutions and seeing if they were logically consistent.</p><p>One of the key pieces of the RL training process is the &#8220;RL environment&#8221;. These environments are sandboxes that allow AI models to perform tasks and see how well they do. In the case of AlphaProof, the environment was an automated theorem prover that could check if the model&#8217;s generated proof was correct. For coding models, the environment could measure how many unit tests a model&#8217;s code passes.</p><p>In the world of chip design, the environment seems pretty obvious. Most industry standard EDA tools can measure a design&#8217;s correctness, speed, size, and power consumption. So we should be able to train a model to write Verilog code, floorplan a chip, design SPICE netlists, or lay out analog circuits, and then measure the performance of those circuits using these industry standard chip design and analysis tools. But in practice, this is extremely difficult to actually pull off.</p><h1>Why is it so hard to use RL for chip design?</h1><p>If RL allows AI models to effectively learn from a reward signal, and chip design tools have such clear reward signals, why hasn&#8217;t RL resulted in an AI chip design breakthrough? Well, the problem is that chip design tools run extremely slowly. The process of going from Verilog to a set of logic gates, called <em>synthesis</em>, can take hours for a large design. Actually placing those logic gates on a chip layout and wiring them all up, called <em>place and route</em>, can take even longer. And then analyzing the timing and power information from that final design takes a while, too.</p><p>The same problem exists in the analog design world. Analog designs feature fewer transistors, but they need to be laid out and simulated with far more precision than digital logic gates. Circuit simulators need to solve complicated differential equations and take manufacturing variability into account, or a circuit that works in simulation may not work when it&#8217;s actually baked into a physical chip. Often, signing off on an analog circuit design requires hours or even days worth of simulation.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/rl-isnt-the-silver-bullet-for-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/rl-isnt-the-silver-bullet-for-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/rl-isnt-the-silver-bullet-for-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>How do we go faster?</h1><p>If the biggest challenge stopping RL from being useful for chip design is how slow the tools are, why not just make the tools faster? Well, a handful of new tools may make that possible.</p><p><a href="https://www.silimate.com/">Silimate</a> offers an AI-powered PPA prediction tool that can estimate a design&#8217;s power consumption and timing characteristics, without having to go through the conventional synthesis and place-and-route process. Their methodology is far faster than conventional methods, and could form the basis of an RL environment that can learn at a fast enough pace to actually get good at writing Verilog. However, having an AI-powered reward function is also risky; the AI model being trained could end up <a href="https://lilianweng.github.io/posts/2024-11-28-reward-hacking/">reward hacking</a> by writing code that the PPA prediction tool thinks will perform well, but would actually perform poorly when passed through the conventional, slower synthesis and place-and-route tools.</p><p>There are also companies like <a href="https://partcl.com/">Partcl</a> who are tackling the problem of accelerating silicon tooling more directly. By accelerating place-and-route and static timing analysis using GPUs, they can speed up key aspects of a chip design RL environment by orders of magnitude. However, <a href="https://www.zach.be/p/why-is-it-so-hard-for-startups-to">building a complete set of chip design tools with full compatibility with modern PDKs is really, really hard</a> -- and making it fast enough to enable RL is even harder. On the analog design side, there&#8217;s <a href="https://dashcrystal.com/">Dash Crystal</a>, who are also building GPU-accelerated simulation tools.</p><p>Ultimately, I have faith that eventually, we&#8217;ll get fast enough chip design tools to enable RL. Whether or not the engineers at these startups are even thinking about RL or not, faster simulation and analysis tools for chip design are valuable in and of themselves. So while it may take a lot of time and effort to develop the full suite of tools necessary for a high-speed, high-performance RL environment for chip design, all of the intermediate steps are still valuable to us human chip designers by making our current workflows that much faster.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How did AMD’s secure enclave get hacked?]]></title><description><![CDATA[Trusting TEEs is harder than you&#8217;d think.]]></description><link>https://www.zach.be/p/how-did-amds-secure-enclave-get-hacked</link><guid isPermaLink="false">https://www.zach.be/p/how-did-amds-secure-enclave-get-hacked</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 18 Aug 2025 14:21:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!lZrJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In the past couple of years, large swaths of the security community have been getting excited about trusted execution environments (TEEs) and secure enclaves (SEVs). Compared to other privacy-enhancing technologies like <a href="https://en.wikipedia.org/wiki/Homomorphic_encryption">homomorphic encryption</a> and <a href="https://en.wikipedia.org/wiki/Zero-knowledge_proof">zero-knowledge proofs</a>, TEEs represent a much faster, cheaper, and more efficient way to secure data. This makes TEEs a viable solution for securing large, complex workloads. People have proposed using TEEs to <a href="https://www.zach.be/p/openais-secure-infrastructure-isnt">secure AI workloads</a>, to <a href="https://writings.flashbots.net/introducing-rollup-boost">accelerate blockchain rollups</a>, and <a href="https://www.anjuna.io/">enable data cleanrooms</a>. But TEEs aren&#8217;t as secure as we might want to believe.</p><p>Unlike cryptographic privacy-enhancing technologies, TEEs and SEVs aren&#8217;t mathematically guaranteed to be secure. They&#8217;re only secure because a bunch of hardware security engineers worked hard to try and figure out every vulnerability and protect against it&#8230; but sometimes they miss things. Earlier this year, we talked about <a href="https://www.zach.be/p/hacking-trusted-execution-environments">BadRAM</a>, which modifies DRAM memory modules to enable unauthorized access to secure memory. There are a large number of <a href="https://www.zach.be/p/apples-m-series-cpus-just-got-hacked">speculative execution vulnerabilities</a>, which leverage performance-enhancing processor features to leak data access patterns. And that&#8217;s not even mentioning <a href="https://www.zach.be/p/side-channel-attacks-on-ai-chips">side-channel attacks</a>, which require physical access to a chip but can <a href="https://www.zach.be/p/the-most-secure-chip-in-the-world">easily leak cryptographic keys</a>.</p><p>Well, there&#8217;s a new TEE vulnerability in town. The <a href="https://heracles-attack.github.io/">Heracles attack</a> on AMD&#8217;s SEV technology enables a malicious hypervisor to read secure memory from a user process. Essentially, this means that a malicious cloud provider could access secret customer data running in confidential TEEs on their hardware. Today, we&#8217;re going to walk through the specifics of the attack, how to stop it, and what it means for TEE design in the future.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>How do TEEs encrypt memory?</h1><p>Trusted execution environments let user processes define different memory regions as <em>enclaves</em>, which can only be accessed by that specific process. This prevents other processes on the system from accessing or tampering with a user&#8217;s secret data. Most TEEs also feature both hardware and software isolation features to prevent processes from accessing each others&#8217; memory regions. But encrypting large amounts of active memory <a href="https://en.wikipedia.org/wiki/Disk_encryption_theory">is a difficult problem</a>.</p><p>To make a TEE work efficiently, large blocks of data need to be securely encrypted without significantly impacting the latency of accessing and using that data. At the same time, a user needs to be able to access and decrypt any chunk of data from anywhere on the disk without too much overhead. Many common block cipher modes, like <a href="https://en.wikipedia.org/wiki/Block_cipher_mode_of_operation">cipher block chaining</a>, encrypt data in blocks such that each block depends on the previous block. This means that a user would need to read the whole disk to decrypt any block. There are workarounds, but they require adding a kind of metadata called <em>initialization vectors</em> to each sector of the disk.</p><p>AMD SEV uses a block cipher mode called <a href="https://en.wikipedia.org/wiki/Xor%E2%80%93encrypt%E2%80%93xor">Xor-Encrypt-Xor (XEX)</a> for memory encryption. XEX is designed from the ground up to be well-suited to encrypting large chunks of memory. Both XEX encryption and decryption have independent block processing, which means that decryption can be parallelized to significantly improve latency. XEX also avoids initialization vectors by <em>tweaking one</em> of the encryption keys based on the address of the data in memory. That way, a user can read any chunk of memory they want, independently of other blocks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lZrJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lZrJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png 424w, https://substackcdn.com/image/fetch/$s_!lZrJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png 848w, https://substackcdn.com/image/fetch/$s_!lZrJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png 1272w, https://substackcdn.com/image/fetch/$s_!lZrJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lZrJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png" width="841" height="323" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:323,&quot;width&quot;:841,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lZrJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png 424w, https://substackcdn.com/image/fetch/$s_!lZrJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png 848w, https://substackcdn.com/image/fetch/$s_!lZrJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png 1272w, https://substackcdn.com/image/fetch/$s_!lZrJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb33485d8-9dc2-4ec3-9be5-5dcde06d5923_841x323.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>(<a href="https://en.wikipedia.org/wiki/Disk_encryption_theory">from Wikipedia</a>)</em></figcaption></figure></div><p>This address-based key tweak is important, as it prevents an adversary from copying a user&#8217;s data to a different chunk of memory and then decrypting it. Because the copied data is in a different address, it will decrypt with a different key, and turn into gibberish. But the Heracles attack breaks that assumption using some clever memory-swapping tricks.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/how-did-amds-secure-enclave-get-hacked?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/how-did-amds-secure-enclave-get-hacked?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/how-did-amds-secure-enclave-get-hacked?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>How does Heracles work?</h1><p>Heracles relies on the fact that the <a href="https://en.wikipedia.org/wiki/Hypervisor">hypervisor</a>, which is the piece of software that manages virtualization on a computer, is capable of moving user data around in memory. When the hypervisor does this, the memory controller re-encrypts the memory with a new key, based on the new memory address. But while the encryption keys vary from address to address, they don&#8217;t change over time.</p><p>To perform the attack, a malicious hypervisor sets up two chunks (pages) of memory: one controlled by the victim, and one controlled by the hypervisor. The attacker observes the encrypted value stored in the user&#8217;s memory. Then, he writes data into his memory, and swaps his memory with the user&#8217;s. Finally, the attacker observes the encrypted value now stored in his memory, which is located in the same address where the user&#8217;s memory just was. If the encrypted value of the attacker&#8217;s memory matches the old encrypted value of the user&#8217;s memory, they must also have had the same plaintext. If the attacker continues on this way block-by-block, they can reconstruct the entire chunk of secure user memory.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3O-u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3O-u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png 424w, https://substackcdn.com/image/fetch/$s_!3O-u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png 848w, https://substackcdn.com/image/fetch/$s_!3O-u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png 1272w, https://substackcdn.com/image/fetch/$s_!3O-u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3O-u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png" width="593" height="424" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e2b87c19-cb20-4823-a324-87e2642559f9_593x424.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:424,&quot;width&quot;:593,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3O-u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png 424w, https://substackcdn.com/image/fetch/$s_!3O-u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png 848w, https://substackcdn.com/image/fetch/$s_!3O-u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png 1272w, https://substackcdn.com/image/fetch/$s_!3O-u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b87c19-cb20-4823-a324-87e2642559f9_593x424.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Unfortunately, this is very slow. AES uses 16-byte blocks, so an adversary would need to guess 2<sup>128</sup> different plaintexts for every block. Luckily, the Heracles team has a clever workaround.</p><p>Specifically, the Heracles team tailors their attack to work in situations where the attacker only needs to guess a single byte. This narrows down the search space from 2<sup>128</sup> down to 2<sup>8</sup> -- several orders of magnitude faster. And it turns out that there are a lot of situations where such a byte-by-byte attack makes sense. For example, when a user is inputting their password while running the <a href="https://en.wikipedia.org/wiki/Sudo">sudo command</a>, that password is written to memory one byte at a time, so Heracles can easily leak it.</p><p>Using similar tricks to restrict an attack to single-byte guessing, the Heracles team can hijack a user&#8217;s session on a webserver, and they can even leak a user&#8217;s cryptographic keys. This makes Heracles a powerful and flexible attack, even with its performance limitations.</p><h1>So what do we do now?</h1><p>Luckily, there are mitigations to the Heracles attack. AMD already added support for a policy that disables the ability of the hypervisor to swap guest pages, but this may negatively affect memory utilization and overall system performance. And as far as we know, this group of ETH Zurich researchers are the first to have discovered this attack and disclosed it responsibly, so AMD was able to deploy a fix before anybody was able to leverage Heracles in the wild.</p><p>More generally, though, modern TEEs have some fundamental problems. Firstly, they&#8217;re proprietary, poorly documented, and often have their security features partially obfuscated by the companies that design them. Because the security community isn&#8217;t able to audit these TEEs, there could always be another vulnerability lurking around the corner, waiting to be discovered. That&#8217;s part of the reason I&#8217;m excited about <a href="https://writings.flashbots.net/ZTEE">open-source TEEs</a> and their ability to restore trust into the world of secure hardware.</p><p>But also, building secure TEEs faces some fundamental challenges. Modern TEEs like AMD SEV and Intel SGX often act as a set of additional features crudely bolted on to the side of an existing, insecure processor. Building a truly secure processor might require a fundamentally different architecture. For example, many modern CPUs run <a href="https://en.wikipedia.org/wiki/Instruction_pipelining">deep out-of-order pipelines</a> to increase performance, but these often <a href="https://en.wikipedia.org/wiki/Spectre_(security_vulnerability)">introduce side-channel attacks</a>. A security-oriented high-performance processor might use a large number of parallel compute cores without any pipelining to prevent these sorts of attacks. But that sort of design wouldn&#8217;t be possible inside of the current paradigm of adding additional security instructions to conventional processor cores. I hope one day, the demand for secure processing grows to the point where engineers can build hardware that&#8217;s actually secure from the ground up.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why did Tesla Dojo fail?]]></title><description><![CDATA[Building training chips is a unique challenge.]]></description><link>https://www.zach.be/p/why-did-tesla-dojo-fail</link><guid isPermaLink="false">https://www.zach.be/p/why-did-tesla-dojo-fail</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 11 Aug 2025 16:54:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xG2N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This week, news broke that Elon Musk is <a href="https://techcrunch.com/2025/08/07/tesla-shuts-down-dojo-the-ai-training-supercomputer-that-musk-said-would-be-key-to-full-self-driving/">shutting down the Dojo supercomputer</a> it was using to train its full self-driving models. Part of the reason for the shutdown was the departure of a number of key executives, who <a href="https://www.bloomberg.com/news/articles/2025-08-06/former-tesla-executives-start-automotive-ai-company-densityai">left to found a startup called DensityAI</a>. However, I think there are clear technical reasons why it didn&#8217;t make sense for Tesla to keep developing their own training chips in house.</p><p>And Tesla isn&#8217;t the first company to have given up on training. <a href="https://www.zach.be/p/why-is-sambanova-giving-up-on-ai">Groq, Cerebras, and SambaNova all used to be targeting training as well as inference</a>, but have pivoted, one by one, to a pure inference focus. This isn&#8217;t surprising; not only are there unique technical challenges required to build training chips, but also, moving training runs to new hardware is an immense financial risk for any company. Ultimately, I think Tesla is making the right decision to focus its custom silicon efforts on inference and leave training to Nvidia silicon.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>The risk of failed training runs</h1><p>Training AI models is hard. It&#8217;s also usually very, very expensive. Training GPT-4 scale large language models can take many months and <a href="https://en.wikipedia.org/wiki/GPT-4#:~:text=Sam%20Altman%20stated%20that%20the,4%20had%201%20trillion%20parameters.">cost over $100 million dollars</a>. But training isn&#8217;t just a matter of spending a lot of money on compute and waiting until the model is ready. During the process of training, engineers monitor the <a href="https://machine-learning.paperspace.com/wiki/accuracy-and-loss#loss">network&#8217;s accuracy</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> to ensure that the model is continuing to learn as new data is being fed into the training process. If the network accuracy suddenly starts dropping despite the model getting more data, the model may be failing to converge, which could put the entire training run at risk.</p><p>There are a lot of potential causes for convergence failure when training large models, but numerical stability is often a key reason a training run may go wrong. As models are trained at <a href="https://www.zach.be/p/how-do-nvidia-blackwell-gpus-train">lower and lower precisions</a>, the specific way low-precision numbers are handled in hardware starts to affect how well a training run converges. And unfortunately, <a href="https://www.zach.be/p/how-do-nvidia-blackwell-gpus-train">this is not something that is constant across different pieces of hardware</a>. Nvidia GPUs have automatic mixed-precision features that will automatically tune mixed-precision models to maximize tensor core utilization, scale the loss to avoid non-convergence, and handle issues with out-of-bounds gradient values.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xG2N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xG2N!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png 424w, https://substackcdn.com/image/fetch/$s_!xG2N!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png 848w, https://substackcdn.com/image/fetch/$s_!xG2N!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png 1272w, https://substackcdn.com/image/fetch/$s_!xG2N!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xG2N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png" width="707" height="421" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:421,&quot;width&quot;:707,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xG2N!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png 424w, https://substackcdn.com/image/fetch/$s_!xG2N!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png 848w, https://substackcdn.com/image/fetch/$s_!xG2N!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png 1272w, https://substackcdn.com/image/fetch/$s_!xG2N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20e4168f-c983-4ad5-93fb-942ec3ae6387_707x421.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Mixed precision training with loss scaling <em>(<a href="https://docs.nvidia.com/deeplearning/performance/mixed-precision-training/index.html">Nvidia Docs</a>)</em></figcaption></figure></div><p>Existing training pipelines for large models rely on these hardware-specific features. If you want to shift your training process to a completely new piece of hardware that handles mixed-precision arithmetic differently, like the <a href="https://cdn.motor1.com/pdf-files/535242876-tesla-dojo-technology.pdf">Tesla Dojo</a>, you&#8217;ll also need to re-develop your training pipeline. And if you re-develop your training pipeline, there&#8217;s a good chance you run into non-convergence issues the first few times you try to actually train models. For large models, this is a hugely expensive risk to run.</p><p>Notably, this isn&#8217;t an issue for inference. While models do need to be optimized for inference on specific pieces of hardware, that optimization process is much faster and easier. Instead of developing a new hardware-specific training pipeline and having to wait weeks to ensure that it&#8217;s numerically stable, an inference implementation can be benchmarked in minutes by simply running forward passes through the network. <a href="https://www.zach.be/p/why-is-sambanova-giving-up-on-ai">This is part of the reason why we see much more hardware diversity in inference than in training</a>; put simply, it&#8217;s much easier and less risky to port an inference pipeline to a new chip than a training pipeline.</p><p>Thus far, Tesla&#8217;s training process for their FSD models has <a href="https://techcrunch.com/2025/08/07/tesla-shuts-down-dojo-the-ai-training-supercomputer-that-musk-said-would-be-key-to-full-self-driving/">leveraged both Tesla Dojo chips and Nvidia chips</a>, which means that Tesla has had to maintain two complex training recipes for large models that each account for the specific quirks of two different pieces of hardware. I believe that simplifying to training purely on Nvidia hardware, <a href="https://www.reuters.com/business/autos-transportation/tesla-streamline-its-ai-chip-design-work-musk-says-2025-08-07/">while using its custom AI5 and AI6 silicon for inference</a>, makes a lot of sense for the company.</p><p>But there&#8217;s another piece of this story: all the executives from Dojo who left the company to found a startup called DensityAI. If training on new hardware is so difficult, why are they starting a company to do just that?</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-tesla-dojo-fail?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-tesla-dojo-fail?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-did-tesla-dojo-fail?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>What is DensityAI, anyways?</h1><p><a href="https://www.densityai.com/">DensityAI</a> has been pretty secretive thus far. <a href="https://www.bloomberg.com/news/articles/2025-08-06/former-tesla-executives-start-automotive-ai-company-densityai">According to Bloomberg</a>, they&#8217;re a hardware company focused on industries like automotive and robotics. <a href="https://tally.so/r/wk0kbj">They&#8217;re hiring for roles focused on the datacenter</a>, not on the edge -- so they&#8217;re not building inference chips to go in cars or robots. Instead, they&#8217;re building infrastructure for training models. But a key clue comes from some of their job postings which specifically mention<a href="https://tally.so/r/wz01jE"> GPU programming</a> and <a href="https://tally.so/r/wA6Dr0">CUDA programming</a>. At the same time, their only open hardware roles are focused on <a href="https://tally.so/r/w41vQO">packaging, PCB design, and thermal management.</a> To me, this indicates that DensityAI isn&#8217;t trying to build new chips to replace Nvidia GPUs. Instead, they&#8217;re trying to build new datacenters optimized for large-scale training using existing GPUs.</p><p>On one hand, this makes some sense. Most automotive manufacturers and robotics companies don&#8217;t have the resources or talent to maintain large datacenters for training state-of-the-art models for autonomous systems. DensityAI could make significant inroads with companies that want access to the best AI models for autonomy, but don&#8217;t have the capacity to train those models themselves.</p><p>However, it&#8217;s also unclear to me why autonomous vehicles and robots require specialized datacenters built from the ground up. Many hyperscalers like Amazon, Google, and Microsoft already have compelling cloud datacenter offerings, which I would assume can handle training what is essentially a very large and complex computer vision model. It could be the case that certain automotive reliability and security standards, like <a href="https://en.wikipedia.org/wiki/ISO_26262">ISO 26262</a> or <a href="https://en.wikipedia.org/wiki/ISO_26262">ISO/SAE 21434</a> could require specialized, secure training for autonomous driving models that would make the hyperscalar cloud solutions non-viable -- but to answer that question, we&#8217;ll have to wait for DensityAI to fully come out of stealth.</p><p>Ultimately, it seems like the DensityAI team has learned from Tesla&#8217;s failures and stayed focused on building the best training systems they can using existing GPU hardware, building off of existing state-of-the-art training pipelines, rather than taking the massive risk of building new training hardware from scratch and sinking months and millions of dollars into training runs that fail to converge.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Technically, they monitor the training loss, validation loss, and perplexity, as well as other task-specific metrics.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Sabotaging AI models with GPUHammer]]></title><description><![CDATA[AI chips continue to be insecure.]]></description><link>https://www.zach.be/p/sabotaging-ai-models-with-gpuhammer</link><guid isPermaLink="false">https://www.zach.be/p/sabotaging-ai-models-with-gpuhammer</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Tue, 22 Jul 2025 15:40:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bdf5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>GPUs and AI chips continue to be a huge area of investment, from <a href="https://www.zach.be/p/why-hasnt-there-been-a-big-winner">edge chips like the Jetson</a> to <a href="https://www.zach.be/p/how-do-nvidia-blackwell-gpus-train">datacenter behemoths like the B200</a>. But one area of research that&#8217;s been continuously overlooked is the security of these systems. If we want GPU-powered systems to be making medical decisions or piloting drones, they should be secure. I&#8217;m not the only person saying this: since mid-2024, <a href="https://www.zach.be/p/openais-secure-infrastructure-isnt">OpenAI has been pushing for secure and trusted GPUs to protect their models</a>. But what OpenAI is proposing simply <a href="https://www.zach.be/p/openais-secure-infrastructure-isnt">isn&#8217;t secure enough</a>.</p><p>Specifically, OpenAI and Nvidia are advocating for <a href="https://www.nvidia.com/en-us/data-center/solutions/confidential-computing/">trusted execution environments</a>, which keep AI weights and inputs encrypted until they&#8217;re securely inside the GPU ready for processing. This is an important security feature, but it has its limitations. Firstly, <a href="https://www.zach.be/p/hacking-trusted-execution-environments">TEE implementations are often flawed</a>, which can undermine the security they offer. But more importantly, TEEs can&#8217;t protect against <a href="https://www.zach.be/p/the-basics-of-side-channel-analysis">side-channel attacks</a> like power analysis and electromagnetic analysis -- <a href="https://www.zach.be/p/side-channel-attacks-on-ai-chips">attacks that GPUs are vulnerable to</a>. Luckily, most side-channel attacks on GPUs require physical access to the chip to attach power or EM probes, making them only suitable for attacking edge devices.</p><p>Recently, though, researchers have identified a GPU hardware vulnerability that can be carried out fully remotely on a single shared piece of hardware. It&#8217;s called <a href="https://gpuhammer.com/">GPUHammer</a>, and it&#8217;s an adaption of the <a href="https://en.wikipedia.org/wiki/Row_hammer">classic rowhammer attack</a> to GPU memories. When multiple processes are sharing the same GPU, a malicious process can mount an attack to flip bits in rows of memory it doesn&#8217;t have access to, and sabotage the performance of machine learning models running in a different process. Today, we&#8217;re going to talk about how it works, what it means, and how it should affect GPU design going forwards.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>What is rowhammer, anyways?</h1><p><a href="https://en.wikipedia.org/wiki/Row_hammer">Rowhammer</a> is an attack on DRAM that manipulates the underlying physics of the memory cells to flip bits in rows of memory that an attacker shouldn&#8217;t have access to. In DRAM, each memory cell is implemented using one capacitor and one transistor, in a design called 1T1C memory. All of these capacitors are packed together so densely in modern DRAM that there&#8217;s a meaningful electrical coupling between capacitors in different rows of the memory.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bdf5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bdf5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png 424w, https://substackcdn.com/image/fetch/$s_!bdf5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png 848w, https://substackcdn.com/image/fetch/$s_!bdf5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png 1272w, https://substackcdn.com/image/fetch/$s_!bdf5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bdf5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png" width="985" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:985,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bdf5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png 424w, https://substackcdn.com/image/fetch/$s_!bdf5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png 848w, https://substackcdn.com/image/fetch/$s_!bdf5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png 1272w, https://substackcdn.com/image/fetch/$s_!bdf5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59c45d83-d2ff-4d8d-a221-1c403a03e8a0_985x460.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://users.ece.cmu.edu/~yoonguk/papers/kim-isca14.pdf">If a malicious process repeatedly reads from the same row in memory</a>, it can disturb the values <a href="https://www.mdpi.com/1424-8220/24/2/592">stored in the memory cells next to that row</a>. Not only can this corrupt data that the malicious process doesn&#8217;t have access to, but it can even be used to <a href="https://googleprojectzero.blogspot.com/2015/03/exploiting-dram-rowhammer-bug-to-gain.html">perform sophisticated attacks like privilege escalation</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kQu7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kQu7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png 424w, https://substackcdn.com/image/fetch/$s_!kQu7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png 848w, https://substackcdn.com/image/fetch/$s_!kQu7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png 1272w, https://substackcdn.com/image/fetch/$s_!kQu7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kQu7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png" width="1456" height="395" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:395,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kQu7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png 424w, https://substackcdn.com/image/fetch/$s_!kQu7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png 848w, https://substackcdn.com/image/fetch/$s_!kQu7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png 1272w, https://substackcdn.com/image/fetch/$s_!kQu7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd3be9ee-1c02-4961-a191-34d69ab6c8fa_1600x434.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>DRAM manufacturers have been working to mitigate rowhammer using error correcting codes, automatic refreshing of possible victim rows, and memory scrambling. At the same time, security researchers have reverse engineering scrambling algorithms and finding more complex hammering patterns that can outwit DRAM mitigations. But thus far, all of the research around rowhammer has been focused on CPU memories.</p><h1>How is GPUHammer different?</h1><p>While GPUs also use DRAM, they don&#8217;t use the same DDR and LPDDR memory that CPUs do. Instead, they use a special kind of DRAM optimized for GPUs, called GDDR memory. GDDR has both a faster refresh rate and a higher latency than DDR and LPDDR memory, which makes hammering harder. At the same time, nobody has taken the time to reverse engineer memory scrambling for GPUs. So while rowhammer should conceptually work on a GPU, the GPUHammer authors are the first researchers to put in the legwork to overcome those obstacles and actually develop a real GPU rowhammer attack.</p><p>GPU memory scrambling is uniquely difficult, because Nvidia GPUs never expose physical memory addresses to user-level CUDA code. However, the GPUHammer authors used <a href="https://www.usenix.org/system/files/conference/usenixsecurity16/sec16_paper_pessl.pdf">memory latency patterns</a> to reverse engineer the physical memory addresses from the virtual memory addresses.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NZUF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NZUF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png 424w, https://substackcdn.com/image/fetch/$s_!NZUF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png 848w, https://substackcdn.com/image/fetch/$s_!NZUF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png 1272w, https://substackcdn.com/image/fetch/$s_!NZUF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NZUF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png" width="1456" height="611" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:611,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NZUF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png 424w, https://substackcdn.com/image/fetch/$s_!NZUF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png 848w, https://substackcdn.com/image/fetch/$s_!NZUF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png 1272w, https://substackcdn.com/image/fetch/$s_!NZUF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4df4e0-9c24-40f8-902a-090033798ffd_1600x671.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Memory access latency versus sub-address, by bank</em></figcaption></figure></div><p>Another big challenge is that <a href="https://chipsandcheese.com/p/measuring-gpu-memory-latency">GPU memory accesses are slower than CPUs</a>, which makes hammering much harder. If you try to have a single threaded process repeatedly access the same memory, there won&#8217;t be enough memory accesses fast enough to cause a rowhammer effect. However, by leveraging <a href="https://developer.nvidia.com/blog/using-cuda-warp-level-primitives/">warp-level parallelism</a>, multiple warps can be issuing memory reads simultaneously, and cause enough hammering to actually flip bits in the victim rows.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wqiZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wqiZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png 424w, https://substackcdn.com/image/fetch/$s_!wqiZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png 848w, https://substackcdn.com/image/fetch/$s_!wqiZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png 1272w, https://substackcdn.com/image/fetch/$s_!wqiZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wqiZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png" width="1456" height="208" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:208,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wqiZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png 424w, https://substackcdn.com/image/fetch/$s_!wqiZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png 848w, https://substackcdn.com/image/fetch/$s_!wqiZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png 1272w, https://substackcdn.com/image/fetch/$s_!wqiZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb75d626d-fa2b-49c5-8b59-678aaa8d3709_1600x229.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption"><em>Single-threaded, multi-threaded, and multi-warp hammering.</em></figcaption></figure></div><p>These techniques let the researchers isolate which memory addresses to target with a rowhammer attack, and hammer those addresses hard enough to flip bits in adjacent rows. To prove how devastating this attack could be, they used the attack to flip the sign bit in the weights of a machine learning model running in a different process on the shared GPU. This reduced the accuracy of several machine learning models by over 80%.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SkqN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SkqN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png 424w, https://substackcdn.com/image/fetch/$s_!SkqN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png 848w, https://substackcdn.com/image/fetch/$s_!SkqN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png 1272w, https://substackcdn.com/image/fetch/$s_!SkqN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SkqN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png" width="1456" height="435" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:435,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SkqN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png 424w, https://substackcdn.com/image/fetch/$s_!SkqN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png 848w, https://substackcdn.com/image/fetch/$s_!SkqN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png 1272w, https://substackcdn.com/image/fetch/$s_!SkqN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe02d121a-9a71-453c-8eb9-fc7322119d56_1600x478.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Essentially, if an attacker can successfully carry out a GPUHammer attack, they can totally sabotage the output of an AI model running on a shared GPU.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/sabotaging-ai-models-with-gpuhammer?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/sabotaging-ai-models-with-gpuhammer?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/sabotaging-ai-models-with-gpuhammer?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>Does this matter?</h1><p>As AI becomes more and more important and entrenched in our lives, the ability to tamper-proof AI models also becomes important. Thankfully, though, GPUHammer isn&#8217;t going to suddenly break ChatGPT. Firstly, GPUHammer has thus far only been demonstrated on A6000 GPUs with GPDDR6. The large cloud H100 and B200 cloud GPUs used by state-of-the-art AI models leverage HBM that has stronger on-die error correction and rowhammer prevention. More importantly, though, most cloud services don&#8217;t offer shared GPU workloads; instead, each user gets a certain number of dedicated GPUs, which would mitigate this sort of attack.</p><p>However, there are cases where single GPUs run multiple processes concurrently! <a href="https://en.wikipedia.org/wiki/WebGPU">WebGPU</a> lets websites leverage GPU capabilities of a user&#8217;s computer. In theory, a website with WebGPU could perform rowhammer attacks when opened, corrupting data for other concurrent GPU processes including AI models. More generally, whenever multiple processes are sharing the same GPU, that could open the door for a GPUHammer vulnerability.</p><h1>What do we do now?</h1><p>GPUHammer just highlights the lack of attention that hardware security has gotten in the world of GPUs and AI accelerators. Most CPUs have supported complex trusted execution environments, secure boot protocols, inline memory encryption, and other security features for years. Security conscious edge microcontrollers for applications in payments, crypto wallets, defense, and critical industry come with <a href="https://www.zach.be/p/the-most-secure-chip-in-the-world">even more protections against complex side-channel attacks</a>. CPU DDR memory has been engineered with rowhammer mitigations for almost a decade. But outside of <a href="https://www.zach.be/p/approximate-computing-for-secure">a small number of research papers</a>, building secure AI chips is not a major industry focus.</p><p>I think that&#8217;s a major problem. As AI starts to influence more and more aspects of our lives, we need to make sure that the hardware it&#8217;s running on is secure from first principles. AI chips need <a href="https://www.zach.be/p/novel-cryptography-needs-novel-hardware">post-quantum secure boot</a>, <a href="https://www.zach.be/p/approximate-computing-for-secure">side-channel protections</a>, secure memory, and other key features that are implemented in other high-value, security conscious chips. If we can put that level of security into the chips powering <a href="https://www.zach.be/p/the-most-secure-chip-in-the-world">the credit card you use to buy a burrito</a>, we should be able to put it into the chips running AI models for the defense industry.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why did Tenstorrent just buy Blue Cheetah?]]></title><description><![CDATA[Is it 5D chess, or just a good price?]]></description><link>https://www.zach.be/p/why-did-tenstorrent-just-buy-blue</link><guid isPermaLink="false">https://www.zach.be/p/why-did-tenstorrent-just-buy-blue</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 07 Jul 2025 14:53:04 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oDkZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>News broke last week that <a href="https://tenstorrent.com/">Tenstorrent</a>, the AI startup run by the legendary CPU architect <a href="https://en.wikipedia.org/wiki/Jim_Keller_(engineer)">Jim Keller</a>, was acquiring the <a href="https://tenstorrent.com/en/vision/tenstorrent-acquires-blue-cheetah-analog-design">analog design startup Blue Cheetah</a>. I wasn&#8217;t surprised to see Blue Cheetah take a bit of a <a href="https://daslee.me/quick-thoughts-on-acquihiressoft-landings/">soft landing</a>; while their product portfolio is genuinely impressive, the <a href="https://www.zach.be/p/chiplets-the-future-of-semiconductor?utm_source=publication-search">chiplet market</a> simply isn&#8217;t mature enough to build a huge company there yet. I was much more surprised to see Tenstorrent as the buyer.</p><p>As recently as 2024, most of Tenstorrent&#8217;s sales haven&#8217;t come from selling their Wormhole and Blackhole PCIe cards, or selling chiplets, but instead have <a href="https://www.eetimes.com/tenstorrent-raises-693-million-series-d/">come from licensing their RISC-V-based AI chip designs to other companies</a>, especially in the automotive and robotics industries. So if most of Tenstorrent&#8217;s revenue comes from licensing IP, and Blue Cheetah&#8217;s IP isn&#8217;t valuable enough due to the immature chiplet market to meaningfully add to Tenstorrent&#8217;s portfolio, why would they buy Blue Cheetah?</p><p>More broadly, I know a lot of people have questions about Tenstorrent&#8217;s business strategy. They have so many products: RISC-V IP, AI IP, two different PCIe cards, two different versions of a desktop workstation, and both datacenter and cloud product offerings as well. And this is in a world where most other AI chip startups are simply <a href="https://www.zach.be/p/why-groq-might-struggle-as-an-api">selling LLM inference by the token</a>! What&#8217;s even going on here?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!oDkZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!oDkZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png 424w, https://substackcdn.com/image/fetch/$s_!oDkZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png 848w, https://substackcdn.com/image/fetch/$s_!oDkZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png 1272w, https://substackcdn.com/image/fetch/$s_!oDkZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!oDkZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png" width="1040" height="574" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:574,&quot;width&quot;:1040,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!oDkZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png 424w, https://substackcdn.com/image/fetch/$s_!oDkZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png 848w, https://substackcdn.com/image/fetch/$s_!oDkZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png 1272w, https://substackcdn.com/image/fetch/$s_!oDkZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f4f4d7-1615-42f4-b0b1-f832e46bbaf2_1040x574.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Part of this product confusion has to do with Tenstorrent&#8217;s complicated corporate history -- Jim Keller, the current CEO, was brought in as CTO years after Tenstorrent was founded, and re-molded the company based on his vision. But it&#8217;s unclear how Blue Cheetah&#8217;s IP portfolio will fit into Tenstorrent&#8217;s product portfolio, or if they simply made the acquisition opportunistically because Blue Cheetah was a company with great technology going through financial struggles.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Why was Blue Cheetah struggling?</h1><p><a href="https://www.bcanalog.com/about/">Blue Cheetah Analog Design</a> was founded in 2018 by Dr. Elad Alon and Dr. Eric Chang, two UC Berkeley researchers who developed the <a href="https://bag3-readthedocs.readthedocs.io/en/latest/">Berkley Analog Generator</a>, or BAG. Normally, analog design is a highly manual process, with engineers picking the size and shape of every transistor in a design by hand. BAG allowed designers to quickly produce custom analog designs without all of the manual effort.</p><p>But instead of trying to sell BAG as a tool, Blue Cheetah decided to build a company that used BAG internally to build analog blocks, and sell the analog designs that were produced. This sort of business model meant that they could deliver products that semiconductor companies were already comfortable licensing and integrating into their designs, <a href="https://www.zach.be/p/llms-will-power-next-gen-chip-ip">while maintaining much lower engineering costs</a>.</p><p>The problem arose when they decided which kind of analog designs they wanted to target BAG at. Blue Cheetah decided to primarily focus on die-to-die interfaces for <a href="https://www.zach.be/p/chiplets-the-future-of-semiconductor?utm_source=publication-search">chiplets</a>. I&#8217;m entirely speculating here, but I think they probably focused on chiplet interfaces due to the lack of competing IP in the chiplet space. If they were selling more conventional off-die interface IP, they&#8217;d be competing with giants like Broadcom and Cadence.</p><p>At the time, many people thought <a href="https://www.zach.be/p/chiplets-the-future-of-semiconductor?utm_source=publication-search">a chiplet revolution was coming soon</a>, but it hasn&#8217;t entirely played out that way. Chiplet-based architectures are very common, but they usually consist of multiple chiplets designed by a single company, rather than a heterogeneous set of chiplets connected like LEGOs. <a href="https://en.wikipedia.org/wiki/High_Bandwidth_Memory">High bandwidth memory</a> is the one major exception, but HBM uses a <a href="https://en.wikipedia.org/wiki/High_Bandwidth_Memory#Interface">completely different die-to-die interface</a> than the <a href="https://opencomputeproject.github.io/ODSA-BoW/bow_specification.html">Bunch-of-Wires</a> and <a href="https://www.uciexpress.org/">UCIe</a> interfaces that chiplet IP vendors like Blue Cheetah were selling.</p><p>There are a lot of reasons why we don&#8217;t see the chiplet market growing significantly. There are concerns about which manufacturer is liable if one chiplet fails. Companies are worried about giving up some of their unique competitive advantages if they sell chiplets on the open market. The number of buyers for chiplets is still small. These are all structural challenges with the market that Blue Cheetah wasn&#8217;t equipped to solve, even though they had good technology.</p><p>I&#8217;m personally still optimistic about both automatic analog circuit generation and chiplets, and I&#8217;m especially optimistic about the idea of <a href="https://www.zach.be/p/llms-will-power-next-gen-chip-ip">building silicon IP companies using internal tooling to supercharge your margin structure</a>. Unfortunately, Blue Cheetah tied their fate to the success of the chiplet ecosystem before it was developed enough to support a company of their scale.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-tenstorrent-just-buy-blue?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-tenstorrent-just-buy-blue?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-did-tenstorrent-just-buy-blue?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>Now, what about Tenstorrent?</h1><p>Tenstorrent has had a complicated history. <a href="https://www.eetimes.com/jim-keller-steps-into-ceo-role-at-tenstorrent/">Originally founded in 2016 by Ljubisa Bajic</a>, they were one of the &#8220;first wave&#8221; of AI chip startups, along with Groq, Cerebras, and SambaNova, to recognize that specialized compute for AI would be a large market. Ljubisa and the original Tenstorrent team architected <a href="https://www.techradar.com/pro/firm-headed-by-legendary-chip-architect-behind-amd-zen-finally-releases-first-hardware-days-after-being-selected-to-build-the-future-of-ai-in-japan-tenstorrent-unveils-grayskull-its-risc-v-answer-to-gpus">Grayskull</a>, their first chip designed for AI. Grayskull, as well as its successors <a href="https://tenstorrent.com/hardware/wormhole">Wormhole</a> and <a href="https://tenstorrent.com/hardware/blackhole">Blackhole</a>, leveraged <a href="https://www.eetimes.com/jim-keller-steps-into-ceo-role-at-tenstorrent/">SiFive X280 RISC-V AI CPUs</a> as control processors.</p><p>However, when <a href="https://www.prnewswire.com/news-releases/jim-keller-joins-tenstorrent-as-president-and-cto-301201505.html">Jim Keller joined Tenstorrent as CTO</a>, he started pushing the company towards developing its own RISC-V cores in-house, rather than relying on RISC-V cores licensed from <a href="https://www.sifive.com/">SiFive</a>, an external vendor. This marked a clear departure from Tenstorrent&#8217;s original goals of building AI chips. And that departure became even more marked when <a href="https://tenstorrent.com/ip/risc-v-cpu">Tenstorrent started licensing their RISC-V core externally</a>.</p><p>This was a confusing move. Usually, companies don&#8217;t want to both license chip IP cores and also sell chips. When a company makes both chips and IP, the chips they make could end up competing with the chips their licensees make, which can scare companies away from becoming potential licensees. But because Tenstorrent was just <em>so passionate about RISC-V</em>, they did it anyway. <a href="https://tenstorrent.com/ip/tensor-neo">They also licensed their AI cores externally as well</a>.</p><p>While this move was confusing, it did seem to pay off for the company. When they closed a <a href="https://www.eetimes.com/tenstorrent-raises-693-million-series-d/">$600M+ fundraise in 2024</a>, most of their large deals were <a href="https://www.eetimes.com/tenstorrent-raises-693-million-series-d/">IP licensing agreements</a>, rather than sales of physical chips. And apparently, a big reason Tenstorrent claims they were able to close these licensing deals was because of their <a href="https://tenstorrent.com/software/tt-metalium">open source software stack</a>. This, in turn, makes Tenstorrent&#8217;s foray into <a href="https://tenstorrent.com/hardware/tt-quietbox">desktop workstations</a> make a little bit more sense, at least.</p><p>If Tenstorrent&#8217;s IP licensees want an open-source compiler, there need to be open-source developers to work on that compiler. And what do those developers need? <a href="https://tenstorrent.com/hardware/tt-quietbox">Workstations</a>. Do most of them want workstations that cost <em><a href="https://tenstorrent.com/hardware/tt-quietbox">twelve thousand dollars</a>?</em> Probably not, but apparently Tenstorrent thinks that this is the best way to get their hardware into the hands of open-source developers.</p><p>Tenstorrent seems to be running a weird strategy, but at least it makes some sense: make PCIe cards, desktop workstations, and an <a href="https://tenstorrent.com/hardware/cloud">evaluation cloud</a> to get developers to support your open source ecosystem, and then use that ecosystem to sell RISC-V and AI IP to customers in the <a href="https://tenstorrent.com/vision/rapidus-and-tenstorrent-partner-to-accelerate-development-of-ai-edge-device-domain-based-on-2nm-logi">automotive industry in Japan</a>. It&#8217;s unclear if this strategy will pay off, but this pivot from dedicated chips to RISC-V IP may have been one of the reasons why Ljubisa Bajic and some of the original Tenstorrent founding team left to go found <a href="https://taalas.com/">Taalas</a>, an absolutely wild AI chip startup that could be the subject of an entire article on its own.</p><p>But even if Tenstorrent&#8217;s strategy pays out, it&#8217;s unclear how Blue Cheetah factors into that strategy.</p><h1>Does Blue Cheetah add anything to Tenstorrent?</h1><p>As far as I can tell, Tenstorrent&#8217;s actual PCIe chips aren&#8217;t all that impressive, and mostly serve as a platform for developers to contribute to the open-source compiler ecosystem that drives Tenstorrent&#8217;s IP sales. So if Tenstorrent primarily acquired Blue Cheetah to put Blue Cheetah IP inside of Tenstorrent&#8217;s chips, it doesn't make all that much sense, as chip sales aren&#8217;t a major driver of Tenstorrent&#8217;s revenue.</p><p>But it also doesn&#8217;t make sense for Tenstorrent to acquire Blue Cheetah just to continue licensing Blue Cheetah&#8217;s IP. As I just discussed, Blue Cheetah&#8217;s IP, despite being high-quality, isn&#8217;t driving enough revenue to support the company due to the immature chiplet market. So why did Tenstorrent buy Blue Cheetah? Was it just such a good deal due to Blue Cheetah&#8217;s financial struggles that they couldn&#8217;t say no?</p><p>Well, maybe. But also, Blue Cheetah&#8217;s technology could factor into Tenstorrent&#8217;s plans to sell chiplets rather than just selling IP. They&#8217;re working on selling off-the-shelf chiplets for both their <a href="https://tenstorrent.com/vision/tenstorrent-risc-v-and-chiplet-technology-selected-to-build-the-future-of-ai-in-japan">RISC-V</a> and <a href="https://tenstorrent.com/vision/tenstorrent-selects-samsung-foundry-to-manufacture-next-generation-ai-chiplet">AI cores</a> that can be integrated into larger systems for automotive and robotics applications. These chiplets need die-to-die interfaces in multiple different process nodes including <a href="https://tenstorrent.com/vision/tenstorrent-selects-samsung-foundry-to-manufacture-next-generation-ai-chiplet">Samsung&#8217;s SF4X</a> and <a href="https://tenstorrent.com/vision/tenstorrent-risc-v-and-chiplet-technology-selected-to-build-the-future-of-ai-in-japan">Rapidus&#8217;s new 2nm node</a>. Blue Cheetah&#8217;s analog design automation technology may be a key way for Tenstorrent to deliver these chiplet-based designs.</p><p>Could Tenstorrent just license Blue Cheetah&#8217;s IP to build their chiplets? Sure. But when Blue Cheetah is going through financial struggles, it may be easier to just buy the whole company than some of their products. It&#8217;s a weird reality of startups, but I think it&#8217;s the most likely motivation for this acquisition.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why is it so hard for startups to compete with Cadence?]]></title><description><![CDATA[And is Morris Chang&#8217;s daughter single?]]></description><link>https://www.zach.be/p/why-is-it-so-hard-for-startups-to</link><guid isPermaLink="false">https://www.zach.be/p/why-is-it-so-hard-for-startups-to</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Tue, 01 Jul 2025 14:40:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ueCk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>If you talk to any chip designer, they&#8217;ll complain about <a href="https://www.cadence.com/en_US/home.html">Cadence&#8217;s chip design software</a>. They&#8217;ll probably complain about <a href="https://www.synopsys.com/">Synopsys</a> too, and maybe <a href="https://eda.sw.siemens.com/en-US/">Mentor Graphics</a> (which is now owned by Siemens, but everybody still calls them Mentor Graphics). These three companies have an oligopoly on chip design software, despite their software being slow, difficult to use, segfaulting constantly, and having user interfaces that look like they were designed in 2003. But if everybody hates these electronic design automation (EDA) tools so much, why hasn&#8217;t a cool new startup disrupted the industry?</p><p>Well, startups have been trying for years. Way back in 2007, <a href="https://www.eetimes.com/synopsys-vet-heads-sim-startup-xoomsys/">Xoomsys</a> developed parallel Spice simulators to greatly speed up analog and mixed signal simulators, but failed to ever reach mainstream adoption. In 2011, <a href="http://www.eejournal.com/article/a-brief-and-personal-history-of-eda-part-5-the-acquisition-era/">Extreme DA was acquired by Synopsys</a> for their multi-core static timing analysis tools. <a href="http://www.eejournal.com/article/a-brief-and-personal-history-of-eda-part-5-the-acquisition-era/">Rocketick Technologies met a similar fate in 2016</a>, being acquired by Cadence for their multi-core RTL simulator. And in 2025, <a href="https://partcl.com/">Partcl</a> is building GPU accelerated chip placement and signoff tools.</p><p>In the past, all of these startups either failed or got acquired for $100-500M by the big players like Cadence or Synopsys, and folded into their product portfolio. None of them grew to the size and scale where they could actually break up the Cadence-Synopsys-Mentor oligopoly. The biggest culprit has nothing to do with technology, and everything to do with the unique relationship Cadence, Synopsys, and Mentor have with bleeding-edge fabs like TSMC and Samsung.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>The EDA <s>Alliance</s> Axis</h1><p>TSMC is probably the most important fab in the world. The most valuable chips, from Apple&#8217;s M4 CPUs to NVidia B200 GPUs, are made in their foundries in Taiwan. And TSMC maintains close relationships with a small number of EDA vendors, dubbed the &#8220;<a href="https://www.tsmc.com/english/dedicatedFoundry/oip/eda_alliance">EDA Alliance</a>&#8221;. Some of these vendors have their tools officially certified to work with TSMC processes. For example, here are the certifications for TSMC&#8217;s 3nm process:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ueCk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ueCk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png 424w, https://substackcdn.com/image/fetch/$s_!ueCk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png 848w, https://substackcdn.com/image/fetch/$s_!ueCk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png 1272w, https://substackcdn.com/image/fetch/$s_!ueCk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ueCk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png" width="1198" height="680" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:680,&quot;width&quot;:1198,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ueCk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png 424w, https://substackcdn.com/image/fetch/$s_!ueCk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png 848w, https://substackcdn.com/image/fetch/$s_!ueCk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png 1272w, https://substackcdn.com/image/fetch/$s_!ueCk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd29926e-9a43-4504-b7fc-87f5ace0fff0_1198x680.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Even the third largest EDA vendor in the world, Siemens, doesn&#8217;t have complete certification coverage! It&#8217;s incredibly hard for a startup to get a single tool qualified, let alone multiple. And when chip designers working in modern process nodes spend millions of dollars on mask sets and manufacturing runs, they are very unlikely to take a risk on a tool that isn&#8217;t officially certified and supported by TSMC.</p><p>TSMC is a huge company, without a major incentive to work with startups to certify their tools. When people ask me about competing with Cadence, my tongue-in-cheek response would be that the best first step would be to marry one of the daughters of <a href="https://en.wikipedia.org/wiki/Morris_Chang">Morris Chang, TSMC&#8217;s legendary founder</a>. It&#8217;s a joke, but it rings a bit true -- to make waves in the EDA industry, you need your tools to get certified, and for that, you need to somehow curry favor with TSMC.</p><p>This problem is exacerbated by the sheer complexity of modern process nodes. Building static timing analysis tools for the 28nm process node back in 2010 was much, much easier than building the same tool for the modern 2nm process. Modern tools need massive parallelism to fit large designs, and require complex features to incorporate effects of local statistical variation on the behavior of every individual transistor in the design. It&#8217;s simply a lot harder to build a good tool to support a modern process node.</p><p>Some tools developed by startups have had some success though! How did they pull it off? And more importantly, how can new startups replicate that same success?</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-is-it-so-hard-for-startups-to?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-is-it-so-hard-for-startups-to?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-is-it-so-hard-for-startups-to?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>The Startup Rebels</h1><p>If developing a fully certified and qualified tool as a startup is so hard, how&#8217;s a startup to take on the EDA Alliance? Well, even if every chip designer wants their final signoff flow to rely on certified and qualified tools, that doesn&#8217;t mean that they can <em>only</em> use those tools.</p><p>Most larger chip companies maintain a redundant suite of EDA tools, ranging from simulators to sign-off tools. If your tool is faster, but not certified, it may find a niche in a certain part of the chip design flow. For example, an incredibly fast synthesis and timing analysis tool may enable much faster design space exploration, enabling chip designers to estimate whether their design will meet timing without having to leverage the expensive and slow but fully-qualified tools from Cadence and Synopsys.</p><p>Some companies are taking this idea even further. <a href="https://www.silimate.com/">Silimate</a> offers an AI-powered PPA prediction tool that can provide rough estimates of a chip&#8217;s power consumption and timing characteristics far faster than a conventional synthesis and signoff tool suite can. Obviously a prediction tool will never be qualified the same way that a signoff tool could be, but it proves the value of fast iteration times independent of signoff-ready certification. Perhaps tools like Partcl could end up in this &#8220;complimentary tools&#8221; category at first.</p><p>This strategy has two major downsides, though. The primary buyers of these sorts of tools are companies that already maintain a suite of redundant EDA tools, and those buyers are usually large companies. Chip design startups simply don&#8217;t have the budget to own multiple versions of the same tool. That means that EDA startups selling these &#8220;complimentary&#8221; tools won&#8217;t be able to reap the benefits of <a href="https://www.zach.be/p/why-do-so-many-startups-sell-to-other">selling to other startups</a>. It also could paint a target on their back that&#8217;s clearly visible to Cadence and Synopsys. There&#8217;s a reason why so many EDA startups either <a href="http://www.eejournal.com/article/a-brief-and-personal-history-of-eda-part-5-the-acquisition-era/">get acquired for less than $500M</a>, or get sued into oblivion by major EDA companies.</p><p>Ultimately, if a startup wants to actually take down Cadence and Synopsys, they&#8217;ll need to get certified by TSMC, Samsung, GlobalFoundries, and all the other major fabs. But in the meantime, they can build market share by focusing less on replacing the existing qualified tools, but by complimenting them. And if they can withstand the IP lawsuits and enticing acquisition offers, they may even be able to meaningfully compete with the big EDA companies. And all that without marrying Morris Chang&#8217;s daughter.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why did Untether AI fail?]]></title><description><![CDATA[Another one bites the dust.]]></description><link>https://www.zach.be/p/why-did-untether-ai-fail</link><guid isPermaLink="false">https://www.zach.be/p/why-did-untether-ai-fail</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 16 Jun 2025 13:50:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last week, news broke that <a href="https://www.untether.ai/">Untether AI</a>, an AI inference chip startup, is shutting down. <a href="https://www.eetimes.com/untether-ai-shuts-down-engineering-team-joins-amd/">Their engineering team is joining AMD</a>, but their core product lines of AI inference chips and software aren&#8217;t coming with them to their acquirer. Obviously, I hope this is a good outcome for the engineers at Untether, but I can&#8217;t say I&#8217;m particularly surprised. While Untether had good technology and <a href="https://www.eetimes.com/amd-and-untether-take-on-nvidia-in-mlperf-benchmarks/">impressive performance for small neural networks</a>, they ran into the same pitfall that many AI chip startups have in the past few years. If a company isn&#8217;t focused on large, generative models like diffusion models or LLMs, <a href="https://www.zach.be/p/why-hasnt-there-been-a-big-winner">there&#8217;s not a huge market for their chip</a>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Untether&#8217;s First Chip</h1><p>Untether was <a href="https://www.untether.ai/about/">founded in 2018</a> -- notably, before the generative AI boom kicked off by ChatGPT and StableDiffusion. Their original goal was to <a href="https://www.untether.ai/about/">&#8220;untether&#8221; AI from datacenters</a> and enable AI inference at the <a href="https://www.fortinet.com/resources/cyberglossary/network-edge">network edge</a>. Notably, this is a bit different from what some other <a href="https://www.zach.be/p/why-hasnt-there-been-a-big-winner">edge AI startups were working on at the time</a>; Untether didn&#8217;t want to put their AI chips in edge devices like laptops and ultra-low-power sensors. Instead they wanted to bring edge AI to distributed servers at the periphery of a network. For datacenter AI workloads like <a href="https://ai.meta.com/blog/dlrm-an-advanced-open-source-deep-learning-recommendation-model/">recommender models</a>, avoiding routing requests to the network core could reduce latency and improve reliability.</p><p>Over time, though, that mission expanded. <a href="https://www.untether.ai/untether-ai-raises-13-million-from-intel-capital-and-other-investors-to-accelerate-ai-innovation/">Their 2019 fundraising announcement</a> focused on automotive and traditional cloud use-cases, alongside a mention of more conventional battery-powered edge devices. At the time, most AI chip startups were either focusing purely on high-power datacenter chips, or ultra-low-power chips for always-on sensors, so it made sense for Untether to focus on niches that other players were avoiding. And their first chip, runAI, followed through on their promises, <a href="https://www.untether.ai/untether-ai-ushers-in-the-petaops-era-with-at-memory-computation-for-ai-inference-workloads/">delivering up to 2 PetaOps per accelerator card and as much as 8 TOPs/watt.</a></p><h1>The rise of GenAI</h1><p>Untether&#8217;s second chip, speedAI, was released in 2022, delivering a significant efficiency improvement -- <a href="https://www.untether.ai/untether-ai-unveils-its-second-generation-at-memory-compute-architecture-at-hot-chips-2022/">30 TOPs/watt compared to runAI&#8217;s 8 TOPs/ watt</a>. But it was also released at an unfortunate time. SpeedAI was announced just months before the release of ChatGPT, and the refocusing of the entire AI industry around large language models and generative AI. Without HBM, mediocre chip-to-chip connectivity, and limited on- and off-chip memory, speedAI was doomed to be an also-ran in the LLM inference world.</p><p>So, Untether doubled down on vision inference, a market that their chips could still succeed in. <a href="https://www.untether.ai/untether-ai-and-general-motors-to-develop-next-generation-autonomous-vehicle-perception-systems/">They partnered with GM</a> to work on autonomous vehicle systems. They released reference applications for their tsn200 accelerator card <a href="https://www.untether.ai/untether-ai-ships-the-tsunaimi-tsn200-accelerator-card-delivering-high-performance-inference-beyond-the-datacenter/">focused on &#8220;smart city&#8221; use cases like pedestrian and vehicle detection</a>. And they kept doubling down on vision and sensing applications, <a href="https://www.untether.ai/untether-ai-announces-collaboration-with-arm-to-deliver-high-performance-energy-efficient-solutions/">collaborating with ARM to develop automotive solutions for driver assistance and self-driving cars</a>, and taking aim at markets in <a href="https://www.untether.ai/untether-ai-announces-speedai-accelerator-cards-as-worlds-highest-performing-most-energy-efficient-ai-accelerators-according-to-mlperf-benchmarks/">robotics, surveillance, agriculture, and machine inspection</a>.</p><p>But while Untether was focusing on vision applications for surveillance cameras, the world was moving on. Large language models became the darlings of the AI world, while computer vision became an afterthought, and for good reason. AI-powered surveillance cameras and cars were still a relatively small market, and I still don&#8217;t fully understand what a &#8220;smart city&#8221; even is. On the other hand, LLMs were clearly disrupting huge industries, and quickly. Because of this, AI chip startups were starting to focus less on vision and convolutional networks, and more on accelerating LLMs. And building LLM accelerators requires <a href="https://www.zach.be/p/most-ai-chips-arent-cost-effective">different considerations than vision accelerators</a>. Running an LLM quickly simply requires building a large, fast matrix multiply, and keeping it supplied with as much data as possible. Untether&#8217;s chips were good at matrix multiplication, but didn&#8217;t have the memory capacity, memory hierarchy, or chip-to-chip connectivity to actually tackle LLMs.</p><p>When Untether finally announced MLPerf results in 2024, <a href="https://www.eetimes.com/amd-and-untether-take-on-nvidia-in-mlperf-benchmarks/">they focused on ResNet-50 inference performance</a>. While their competitors were releasing impressive tokens-per-second results for LLama-70B, Untether was bragging about power efficiency on a <a href="https://arxiv.org/pdf/1512.03385">model from 2015</a>.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-untether-ai-fail?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-did-untether-ai-fail?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-did-untether-ai-fail?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>AI chip startups need to see the future</h1><p>Untether had good technology. They had a fantastic team -- there&#8217;s a reason AMD hired all of them. But they missed one of the biggest developments in the AI world: the rise of large generative models. If they had started a couple years later, and hadn&#8217;t taped out their second-generation silicon before the launch of ChatGPT, things could have easily turned out differently.</p><p>This is one of the reasons why I think one of the most important things for an AI chip startup to do is keep an ear to the ground for future developments in AI, and to build chips that are flexible enough to run whatever might be coming down the pipeline. Untether AI called their shot that vision models at the network edge were the future, and so their chips weren&#8217;t flexible enough to run transformers. <a href="https://www.zach.be/p/stop-trying-to-make-etched-happen">Maybe the companies focusing exclusively on transformers</a> will meet a similar fate soon enough.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why is SambaNova giving up on AI training?]]></title><description><![CDATA[The one where Zach gets sued.]]></description><link>https://www.zach.be/p/why-is-sambanova-giving-up-on-ai</link><guid isPermaLink="false">https://www.zach.be/p/why-is-sambanova-giving-up-on-ai</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 05 May 2025 20:42:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Qv3s!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc9753e-da17-4a33-8e22-b4ef3f3d9e07_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Disclaimer: I worked at SambaNova Systems between 2019 and 2021, and so I am going to be somewhat careful about discussing sensitive information. Despite the tongue-in-cheek subtitle, I don&#8217;t want to get in legal trouble. Please don&#8217;t sue me!</em></p><p>In late April, one of the most well-funded AI chip startups out there, SambaNova Systems, <a href="https://www.eetimes.com/sambanova-lays-off-15-of-workforce-to-refocus-on-inference/">pivoted away significantly from their original goal</a>. Like many other AI chip startups, SambaNova wanted to offer a unified architecture for both training and inference. But, as of this year, they&#8217;ve given up on their training ambitions, laid off 15% of their workforce, and are focusing entirely on AI inference. And they&#8217;re not the first company to be making this pivot.</p><p><a href="https://www.electronicdesign.com/markets/automation/article/21805844/groq-outlines-potential-power-of-artificial-intelligence-chip">In 2017, Groq was bragging about their training performance</a>, but by <a href="https://www.prnewswire.com/news-releases/groq-first-to-announce-performance-advantage-results-with-stac-ml-markets-inference-benchmark-meeting-needs-of-financial-services-industry-301664184.html">2022, they were entirely focused on inference benchmarks</a>. The <a href="https://www.nextplatform.com/2024/09/10/the-battle-begins-for-ai-inference-compute-in-the-datacenter/">Cerebras CS-1 was originally sold primarily for training workloads</a>, but the <a href="https://www.nextplatform.com/2024/09/10/the-battle-begins-for-ai-inference-compute-in-the-datacenter/">CS-2 and later shifted their focus to inference</a>. SambaNova seemed to be the last holdout from that first generation of AI chip startups to still seriously focus on training, but that&#8217;s finally changed. So, why are all of these startups pivoting from training to inference? Luckily, as somebody who worked at SambaNova, I have a bit of an insider&#8217;s perspective.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Training was always part of the plan</h1><p>SambaNova was very serious about training models on their hardware. They put out articles about <a href="https://sambanova.ai/blog/training-long-sequence-size-models-on-sambanova">how to train on their hardware</a>, <a href="https://sambanova.ai/blog/world-record-large-language-models-training-performance-sambanova-dataflow-as-a-service-gpt">bragged about their training performance</a>, and <a href="https://docs.sambanova.ai/sambastudio/latest/training.html">addressed training in their official documentation</a>. A lot of analysts and outside observers, including me, saw the ability to tackle both the inference and training markets with one chip as a unique edge that SambaNova had over competitors like Groq, <a href="https://www.forbes.com/sites/karlfreund/2021/02/25/the-cambrian-ai-landscape-groq/">which was one of the earliest startups to pivot to inference</a>.</p><p>SambaNova also invested significant time and effort into enabling efficient training. When I was at the company between 2019 and 2021, I spent a considerable amount of time implementing a kernel for the <a href="https://keras.io/api/optimizers/Nadam/">NAdam optimizer</a>, a momentum-based optimizer commonly used to train large neural networks. We had hardware and software features designed and optimized for training, and both internal and external messaging indicated that support for training was a key part of our value proposition.</p><p>Now, all of a sudden, SambaNova is essentially scrapping most of that work to focus entirely on inference. I think that they&#8217;re doing this for three main reasons: the fact that inference is an easier problem to tackle, the fact that inference may represent a larger market than training, and Nvidia&#8217;s total dominance in the world of AI training chips.</p><h1>Inference is an easier, larger market.</h1><p>Many analysts believe that the market size for AI inference <a href="https://www.nextplatform.com/2024/09/10/the-battle-begins-for-ai-inference-compute-in-the-datacenter/">could be ten times bigger than the market for AI training</a>. Intuitively, this makes sense. Normally, you only train a model once, and then perform inference using that model many, many, many times. Each time you run inference, it costs far, far less than the entire training process for a model &#8212; but if you run inference using the same model enough times, it becomes the dominant cost when serving that model. If the future of AI is a small number of large models, each with significant inference volume, the <a href="https://www.zach.be/p/why-is-everybody-talking-about-groq">inference market will dwarf the training market</a>. But if many organizations end up training their own bespoke models, <a href="https://www.zach.be/p/why-is-everybody-talking-about-groq">this future may not come to pass</a>.</p><p>But even if inference doesn&#8217;t pan out to be a much larger market than training, there are technical reasons why inference is easier to tackle for AI chip startups. When training a model, you need to run a bunch of training data though that model, collect gradient information during the model&#8217;s operation, and use those gradients to update the model&#8217;s weights. This process is what allows the model to learn. It&#8217;s also extremely memory-intensive, as you need to cache all of those gradients as well as other values, like the model&#8217;s activations.</p><p>So, to efficiently perform training, you need a complex memory hierarchy with on-die SRAM, in-package HBM, and off-chip DDR. <a href="https://www.nextplatform.com/2024/11/06/we-cant-get-enough-hbm-or-stack-it-up-high-enough/">It&#8217;s hard for AI startups to get their hands on HBM</a>, and <a href="https://www.zach.be/p/ai-chip-startups-dont-need-to-be">hard to integrate HBM into a high-performance system</a> -- so many AI chips like <a href="https://www.zach.be/p/why-is-everybody-talking-about-groq">Groq</a> and <a href="https://www.zach.be/p/most-ai-chips-arent-cost-effective">d-Matrix</a> just don&#8217;t feature the necessary HBM or DDR capacity or bandwidth to efficiently train large models. Inference doesn&#8217;t have this problem. During the inference process, gradients don&#8217;t need to be stored, and activations can be discarded after they&#8217;re used. This vastly reduces the memory footprint of inference as a workload, and reduces the complexity of the memory hierarchy inference-only chips need.</p><p>Another challenge is inter-chip networking. All of those gradients generated during training need to be synchronized across every chip used in the training process. That means you need a <a href="https://www.zach.be/p/softbank-is-building-an-ai-datacenter">large, complex, all-to-all network</a> to efficiently run training. Inference, on the other hand, is a feed-forward operation, with each chip only talking to the next chip in the inference pipeline.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> <a href="https://www.zach.be/p/why-is-everybody-talking-about-groq">Many startups&#8217; AI chips have limited networking capabilities</a>, which makes them poorly suited for the all-to-all connectivity that training training, but sufficient for inference workloads. Nvidia, on the other hand, has addressed both the memory and networking challenges for AI training extremely well.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-is-sambanova-giving-up-on-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-is-sambanova-giving-up-on-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-is-sambanova-giving-up-on-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>Nvidia is extremely good at training.</h1><p>Nvidia has been the hardware of choice for both inference and training since the <a href="https://en.wikipedia.org/wiki/AlexNet">AlexNet days in 2012.</a> Because of the versatility CUDA grants GPUs, they&#8217;re capable of performing all of the necessary operations for both training and inference. And in the past decade, Nvidia has been focused not just on <a href="http://www.zach.be/p/stop-trying-to-make-etched-happen">building hyper-optimized chips for machine learning workloads</a>, they&#8217;ve also been <a href="https://www.zach.be/p/softbank-is-building-an-ai-datacenter">optimizing their entire memory and networking stack for large scale training and inference</a>.</p><p>With access to <a href="https://www.zach.be/p/most-ai-chips-arent-cost-effective">significant amounts of HBM on every die</a>, Nvidia hardware is able to easily and efficiently cache all of the gradient updates generated by each training step. And with scale-up technologies like <a href="https://en.wikipedia.org/wiki/NVLink">NVLink</a> and scale-out technologies like <a href="https://en.wikipedia.org/wiki/InfiniBand">Infiniband</a>, Nvidia hardware is able to handle the all-to-all networking required to update all of the weights of a large neural network after each training step is complete. Inference-only competitors like <a href="https://www.zach.be/p/why-is-everybody-talking-about-groq?utm_source=publication-search">Groq</a> and <a href="https://www.zach.be/p/most-ai-chips-arent-cost-effective">d-Matrix</a> simply lack the memory and networking capabilities to compete with Nvidia on training.</p><p>But <a href="https://www.eetimes.com/sambanova-adds-hbm-for-llm-inference-chip/">SambaNova chips do have HBM</a>. SambaNova chips <a href="https://sambanova.ai/blog/sn40l-chip-best-inference-solution">have a peer-to-peer network</a> at both the server-level and rack-level. Why can&#8217;t they tackle training the way that Nvidia can?</p><p>Well, it turns out that Nvidia has more than just HBM and networking to give them a leg up on training performance. <a href="https://www.zach.be/p/how-do-nvidia-blackwell-gpus-train">They&#8217;ve put significant effort into low-precision training</a>, and top AI labs, in turn, have put significant effort into tuning algorithm hyper-parameters to work well with the specific intricacies of Nvidia&#8217;s low-precision training hardware. Shifting from Nvidia to SambaNova chips for training requires changing extremely sensitive training code to run on entirely new hardware with an entirely new set of pitfalls. The cost and risk of doing that for a large, GPT-4 scale model is immense.</p><p>SambaNova&#8217;s pivot to inference is proof that, even if an AI chip startup manages to offer competitive memory and networking capabilities to Nvidia, it&#8217;s not enough to take on the green giant in the training market. If a startup wants to challenge Nvidia on training, they need to offer such impressive training performance, that they can overcome Nvidia&#8217;s inertia in the training market. And so far, nobody&#8217;s been able to pull that off.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>This is an oversimplification, but serves to illustrate the different networking requirements for training and inference.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Why is it taking so long to build new IP cores?]]></title><description><![CDATA[Standards are getting more complex and harder to satisfy.]]></description><link>https://www.zach.be/p/why-is-it-taking-so-long-to-build</link><guid isPermaLink="false">https://www.zach.be/p/why-is-it-taking-so-long-to-build</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Mon, 28 Apr 2025 14:50:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!FmEa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The USB4 specification was <a href="https://en.wikipedia.org/wiki/USB4">first released in 2019</a>. It took a year for the <a href="https://news.synopsys.com/2020-06-03-Synopsys-Introduces-Industrys-First-Complete-USB4-IP-Solution">first USB4 IP cores to hit the market</a>. Even worse, it wasn&#8217;t until 2025 that the <a href="https://www.synopsys.com/blogs/chip-design/usb4-device-ip-edge-ai.html">first USB4 IP was fully certified by the USB implementers forum</a>. And USB4 isn&#8217;t the only connectivity protocol where there was a major lag between the specification being finalized and the first IP being available.</p><p><a href="https://web.archive.org/web/20170207150136/https://www.techrepublic.com/article/new-pci-express-4-0-delay-may-empower-next-gen-alternatives/">PCIe 4.0 was notably delayed</a>; the PCIe 4.0 standard was <a href="https://en.wikipedia.org/wiki/PCI_Express">released in 2017</a> but it took AMD and Intel until 2019 and 2020, respectively, to ship CPUs with PCIe 4.0 support. It got to a point where <a href="https://en.wikipedia.org/wiki/Gen-Z_(consortium)#:~:text=The%20effort%20followed%20years%20of,formed%20to%20work%20on%20an">a new industry consortium spun up to find alternatives</a> to the delayed standard. Similar stories exist for DisplayPort 2.0: it was released in 2019, but major vendors like Synopsys <a href="https://www.synopsys.com/designware-ip/interface-ip/displayport.html#products">don&#8217;t even offer DisplayPort 2.0 cores at all</a> in 2025.</p><p>And it&#8217;s not just connectivity IP cores that often experience major delays from specification to commercial release. The same is true for video codecs: <a href="https://en.wikipedia.org/wiki/High_Efficiency_Video_Coding">HEVC H.265 was released in January 2013</a>, but IP <a href="https://www.design-reuse.com/news/33893/hardware-hevc-h-265-encoder-ip.html">didn&#8217;t hit the market</a> until <a href="https://www.prnewswire.com/news-releases/squid-systems-video-codec-hardware-ip-available-for-licensing-270007291.html">2014</a>, and only from small, specialized video IP vendors. AV1 was <a href="https://en.wikipedia.org/wiki/AV1">standardized in 2018</a>, and it took <a href="https://www.prnewswire.com/news-releases/av1-encoder-and-decoder-hardware-ips-available-from-allegro-dvt-and-embedded-in-products-by-end-of-2020-301038104.html">until late 2020 for IP to hit the market</a>. VVC H.266 was <a href="https://en.wikipedia.org/wiki/Versatile_Video_Coding">standardized in 2020</a>, and IP wasn&#8217;t released <a href="https://www.allegrodvt.com/news/allegro-dvt-industrys-real-time-vvc-h-266-encoder-ip/">until 2024</a>.</p><p>In the last decade or so, we&#8217;ve seen more and more examples of commercial IP releases lagging significantly behind the release of corresponding specifications and standards. This is becoming a major problem; widely available IP cores are necessary for an ecosystem to grow around a new standard. For example, if DisplayPort 2.0 IP isn&#8217;t available, it becomes harder for CPU, GPU, and display vendors to justify building DP 2.0 capabilities into their devices, because other devices consumers own won&#8217;t support DP 2.0. The same is true for video codecs, cryptographic algorithms, and memory interfaces.</p><p>But why is it taking so much longer to build new IP cores?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Standards are getting more complex</h1><p>First off, the standards that IP vendors are expected to implement are getting more and more complex over time. Let&#8217;s take a look at a state machine from <a href="https://ieeemilestones.ethw.org/File:USB_1.0_Specification.pdf">the USB 1.0 specification</a>, released in 1996. The entire spec is a little over 250 pages long, and includes everything you need to know about USB 1.0, from the logical protocol to the electrical specifications to the mechanical drawings of the connector. This state machine handles &#8220;bit stuffing&#8221;, which ensures that the data doesn&#8217;t have too many consecutive 1&#8217;s or 0&#8217;s.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FmEa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FmEa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png 424w, https://substackcdn.com/image/fetch/$s_!FmEa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png 848w, https://substackcdn.com/image/fetch/$s_!FmEa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png 1272w, https://substackcdn.com/image/fetch/$s_!FmEa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FmEa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png" width="378" height="788" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:788,&quot;width&quot;:378,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FmEa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png 424w, https://substackcdn.com/image/fetch/$s_!FmEa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png 848w, https://substackcdn.com/image/fetch/$s_!FmEa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png 1272w, https://substackcdn.com/image/fetch/$s_!FmEa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F102c4a13-57b3-4854-943a-7f4f93d37d61_378x788.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://ieeemilestones.ethw.org/w/images/4/44/USB_1.0_Specification.pdf">USB Bit Stuffing Diagram</a></figcaption></figure></div><p>This single diagram includes basically all of the information you need to implement bit stuffing on USB 1.0. It&#8217;s a simple and effective solution for the data transfer speeds that USB 1.0 runs at.</p><p>USB 4.0, on the other hand, <a href="https://www.usb.org/document-library/usb4r-specification-v20">is a much more complex specification</a>. The core spec is over 800 pages long, and that doesn&#8217;t include many of the mechanical specifications of the connector. USB 4.0 also has a method to prevent long runs of consecutive 1&#8217;s or 0&#8217;s, but it&#8217;s much more complex. It uses a <a href="https://en.wikipedia.org/wiki/Scrambler">scrambler</a>, which randomly shuffles data according to a pseudorandom number generator, or PRNG. There are many pages of specifications for this scrambler, as both the transmitter and receiver need to have their PRNGs synchronized to ensure the message gets properly encoded and decoded. Below is a diagram of how USB 4.0 handles re-synchronizing the transmitter and receiver&#8217;s scramblers, depending on what mode it&#8217;s operating in.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lBAR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lBAR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png 424w, https://substackcdn.com/image/fetch/$s_!lBAR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png 848w, https://substackcdn.com/image/fetch/$s_!lBAR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png 1272w, https://substackcdn.com/image/fetch/$s_!lBAR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lBAR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png" width="874" height="894" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:894,&quot;width&quot;:874,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lBAR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png 424w, https://substackcdn.com/image/fetch/$s_!lBAR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png 848w, https://substackcdn.com/image/fetch/$s_!lBAR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png 1272w, https://substackcdn.com/image/fetch/$s_!lBAR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e78ea1b-2dd2-45ff-ada4-46ab102ad6c8_874x894.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://www.usb.org/document-library/usb4r-specification-v20">USB 4.0 Scrambler</a></figcaption></figure></div><p>As you can see, USB 4.0 is already a lot more complex, and this is just a single diagram that explains a single part of the scrambler resynchronization process. There are pages and pages of additional information on how to properly implement this feature, and it&#8217;s one of many, many complex features in the USB 4.0 specification.</p><p>USB 4.0 also has entirely new features, like <a href="https://www.edn.com/what-is-link-training-and-when-should-i-use-it/">link training</a>. This is a complex handshake between the transmitter and the receiver to tune their equalization settings to make the communication link more reliable. USB 1.0 didn&#8217;t need this at all, as it was simple and low-speed enough to get away with straightforward, constant equalization methods. But USB 4.0 has multiple different link training methods, depending on whether you&#8217;re using the USB 4.0 bus to transmit DisplayPort data or regular USB3.0-style serial data.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!THP9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!THP9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png 424w, https://substackcdn.com/image/fetch/$s_!THP9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png 848w, https://substackcdn.com/image/fetch/$s_!THP9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png 1272w, https://substackcdn.com/image/fetch/$s_!THP9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!THP9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png" width="863" height="543" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:543,&quot;width&quot;:863,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!THP9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png 424w, https://substackcdn.com/image/fetch/$s_!THP9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png 848w, https://substackcdn.com/image/fetch/$s_!THP9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png 1272w, https://substackcdn.com/image/fetch/$s_!THP9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7cf0651e-367c-42c5-a94e-eaef0447f339_863x543.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://www.usb.org/document-library/usb4r-specification-v20">USB 4 DisplayPort Link Training</a></figcaption></figure></div><p>All of these features are really important, though. USB 4.0 can run at up to <a href="https://en.wikipedia.org/wiki/USB4">80 Gbit/s</a>, while USB 1.0 only ran at <a href="https://en.wikipedia.org/wiki/USB#USB_1.x">12 Mbit/s</a>. That means that USB 4.0 is over 6000 times faster than USB 1.0, and those performance gains had to come from somewhere. To reliably transmit data so much faster, much more complex transmitter and receiver architectures are necessary. But that makes the process of designing IP cores for these new standards that much harder. And it gets even harder when IP core vendors have to support so many different IP cores at once.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-is-it-taking-so-long-to-build?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/why-is-it-taking-so-long-to-build?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/why-is-it-taking-so-long-to-build?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>IP vendors are stretched thin</h1><p>IP vendors have large, complex product portfolios they need to maintain. Even though USB 3.0 and 4.0 have been released, some low-speed, low-cost devices still use USB 2.0 IP cores, and IP vendors are often expected to support those legacy protocols alongside more modern ones. Maintaining those cores isn&#8217;t free, though. Often, customers will want tweaks to customize legacy IP cores, which requires engineering effort that could otherwise be spent developing new IP cores to meet new standards.</p><p>As an example, Synopsys has an <a href="https://www.synopsys.com/designware-ip.html">incredibly wide portfolio of IP cores</a>, and <a href="https://en.wikipedia.org/wiki/Synopsys">about 20,000 employees</a>. Of course, some of those employees work on Synopsys&#8217; EDA products, but a good portion of them work on IP development. The overhead of managing that many employees to maintain that many products is huge, and prevents Synopsys from moving quickly to assign their best and brightest employees to tackle IP core development for cutting-edge standards.</p><p>But this also creates opportunities for startups, who don&#8217;t have that overhead, to find niches where they can outcompete those big players like Synopsys.</p><h1>Opportunities for disruption</h1><p>As new standards for connectivity, cryptography, video compression, and memory get released, new opportunities for startups are created. As I mentioned, large IP companies like Cadence, Synopsys, and Rambus are primarily focused on maintaining a large, broad product portfolio. Startups, unencumbered by the need to maintain existing product lines, can focus on out-executing incumbents on one particular new standard.</p><p>My last startup, Radical Semiconductor, <a href="https://www.zach.be/p/novel-cryptography-needs-novel-hardware">did this for post-quantum cryptography</a>, and <a href="https://www.prnewswire.com/news-releases/btq-technologies-completes-acquisition-of-radical-semiconductors-processing-in-memory-technology-portfolio-advancing-post-quantum-cryptography-capabilities-302256705.html">we were acquired in 2024</a>. And we&#8217;re not the only example. NGCodec was one of the first providers of FPGA H.265 IP, <a href="https://www.streamingmedia.com/Articles/News/Online-Video-News/Xilinx-Acquires-NGCodec-for-Cloud-Video-Encoding-IP-and-Talent-132841.aspx">and was acquired by Xilinx in 2019</a>. PLDA beat many major IP providers to silicon-ready IP for PCIe 5.0 and CXL 2.0, <a href="https://www.rambus.com/plda/">and were acquired by Rambus in 2021</a>.</p><p>Most of these IP core startups end up being acquired by an existing IP vendor, rather than becoming a successful IP company in their own right. But potentially, with the <a href="https://www.zach.be/p/yc-is-wrong-about-llms-for-chip-design">advent of AI-powered chip design</a>, a startup could stay at the forefront of new standards and specifications across multiple product lines, and <a href="https://www.zach.be/p/llms-will-power-next-gen-chip-ip">become a legitimate rival to IP giants like Synopsys and Cadence</a>. Either way, regardless of a founder's ambitions, new standards and specifications have historically been great opportunities for startups to disrupt incumbents, and I suspect that trend will continue into the future.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Scaling Thermodynamic Computing]]></title><description><![CDATA[Or: What the hell is Normal Computing actually making?]]></description><link>https://www.zach.be/p/scaling-thermodynamic-computing</link><guid isPermaLink="false">https://www.zach.be/p/scaling-thermodynamic-computing</guid><dc:creator><![CDATA[zach]]></dc:creator><pubDate>Wed, 16 Apr 2025 12:28:25 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Note:</em> Since this post was published, I&#8217;ve transitioned to an advisory role at Normal Computing, and taken a full-time role leading hardware security at <a href="https://www.btq.com/">BTQ Technologies</a>, which <a href="https://www.zach.be/p/novel-cryptography-needs-novel-hardware">acquired my previous startup, Radical Semiconductor, in 2024</a>. This post may not reflect the current architecture or plans of Normal Computing whenever you&#8217;re reading this.</p><div><hr></div><p><em>This is a crosspost between my personal blog, <a href="http://zach.be">zach.be</a>, and the official Normal Computing blog (<a href="https://www.normalcomputing.com/blog">https://www.normalcomputing.com/blog</a>)!</em></p><p>Last week, I gave a talk at Normal Computing&#8217;s event at<a href="https://nyc-deep-tech-week.com/"> NYC Deep Tech Week</a>. This was a pretty exciting time &#8211; not only did we have a full slate of excellent speakers and panels, but I also got to publicly share some more details &#8211; for the first time ever &#8211; about the chip that we&#8217;re actually building here at Normal. Specifically, I went over how our silicon architecture enables unprecedented scalability for thermodynamic computing, with a focus on large-scale generative design, scientific computing, and probabilistic reasoning algorithms.</p><p>Because not everybody is fortunate enough to live in NYC, I wanted to also write a short blog post summing up what I presented in that talk. If this sort of thing piques your interest, I&#8217;d recommend<a href="https://jobs.ashbyhq.com/Normal%20Computing%20AI"> applying to work at Normal</a>. We have a world-class team in SF, NYC, London, and Copenhagen, and we&#8217;re entering an important scaling phase now. Our work is<a href="https://www.prnewswire.com/news-releases/normal-computing-selected-for-arias-50m-scaling-compute-programme-to-revolutionize-ai-hardware-costs-302293266.html"> in significant part funded</a> by the ambitious Advanced Research + Invention Agency (ARIA) Scaling Compute Program.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Scaling laws: today and looking ahead</h1><p>AI workloads are getting bigger, and more than<a href="https://www.google.com/search?q=semianalysis+300%2C000+gpu+clusters&amp;oq=semianalysis+300%2C000+gpu+clusters&amp;gs_lcrp=EgZjaHJvbWUyBggAEEUYOTIHCAEQIRirAjIHCAIQIRirAjIHCAMQIRirAtIBCDYzMzdqMGoxqAIAsAIA&amp;sourceid=chrome&amp;ie=UTF-8"> 300,000 GPU clusters</a> are in the works in 2025. Frontier labs are starting to do their first<a href="https://semianalysis.com/2024/09/04/multi-datacenter-training-openais/"> multi-data-center distributed training runs</a>. And in the next 5 years we&#8217;ll need gigawatt data centers, to scale these runs by <a href="https://epoch.ai/blog/can-ai-scaling-continue-through-2030#the-current-trend-of-ai-power-demand">10,000x</a>.</p><p>The last few years of AI development have been focused on <em>scaling laws</em>. For LLM pre-training, these laws tell us that with more data and compute, we can predictably improve the quality of our AI models. There are clear paths to getting more data; for example, <a href="https://semianalysis.com/2024/12/11/scaling-laws-o1-pro-architecture-reasoning-training-infrastructure-orion-and-claude-3-5-opus-failures/">online video data likely contains trillions of additional tokens to train on</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-duj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-duj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png 424w, https://substackcdn.com/image/fetch/$s_!-duj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png 848w, https://substackcdn.com/image/fetch/$s_!-duj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png 1272w, https://substackcdn.com/image/fetch/$s_!-duj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-duj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png" width="1456" height="818" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:818,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-duj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png 424w, https://substackcdn.com/image/fetch/$s_!-duj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png 848w, https://substackcdn.com/image/fetch/$s_!-duj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png 1272w, https://substackcdn.com/image/fetch/$s_!-duj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec89e7ea-9a72-43c4-97e2-a69e4a81595c_1600x899.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://openai.com/index/learning-to-reason-with-llms/">OpenAI, 2024</a></em></figcaption></figure></div><p>These scaling laws demand ever increasing amounts of compute, and chip startups have popped up to meet that demand. Some chips are trying to deliver more compute through <a href="https://www.zach.be/p/stop-trying-to-make-etched-happen">specialization</a>; others are leveraging emerging technologies like <a href="https://www.zach.be/p/most-ai-chips-arent-cost-effective">processing-in-memory</a> to break through key system-level bottlenecks. These methods may give 5-10x performance improvements over GPUs for some use-cases, at the expense of <a href="https://www.zach.be/p/why-etched-probably-wont-beat-nvidia">technical risk or market risk</a>. But even these new chip architectures won&#8217;t be sufficient as AI models continue to scale. Experts believe that, within five years, AI data centers will start <a href="https://epoch.ai/blog/data-movement-bottlenecks-scaling-past-1e28-flop">hitting fundamental memory and compute bottlenecks.</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4rDZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4rDZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png 424w, https://substackcdn.com/image/fetch/$s_!4rDZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png 848w, https://substackcdn.com/image/fetch/$s_!4rDZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png 1272w, https://substackcdn.com/image/fetch/$s_!4rDZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4rDZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png" width="1400" height="715" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:715,&quot;width&quot;:1400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4rDZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png 424w, https://substackcdn.com/image/fetch/$s_!4rDZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png 848w, https://substackcdn.com/image/fetch/$s_!4rDZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png 1272w, https://substackcdn.com/image/fetch/$s_!4rDZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0792765-3a7c-4ba9-82e0-1f13c5ee88e1_1400x715.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://arxiv.org/abs/2403.14123">Gholami et. al, 2024</a></em></figcaption></figure></div><p>Current LLMs may be good enough for some tasks, but as we scale to more complex workloads, like <a href="https://lilianweng.github.io/posts/2024-04-12-diffusion-video/">video diffusion</a> or complex reasoning with uncertainties, we&#8217;ll start to hit those bottlenecks even faster. Ultimately, the race to build AI is not just so we can do homework faster or draft emails. From<a href="https://semianalysis.com/2025/03/19/nvidia-gtc-2025-built-for-reasoning-vera-rubin-kyber-cpo-dynamo-inference-jensen-math-feynman/"> Jensen</a> to the <a href="https://www.linkedin.com/posts/yann-lecun_jensen-huang-at-gtc-the-future-of-ai-is-activity-7307813334974713856-zfoj/">AI godfathers</a>, experts are in agreement that the next frontier for intelligence is in advancing the capabilities to <em><strong>reason about the physical world.</strong></em></p><p>These applications are the ones we typically find ourselves most excited about: from redesigning hardware and chips themselves, to achieving autonomy in transportation, to scaling higher fidelity forms of planning and reasoning compared to the heuristics we have today.</p><h1>Simulating the physical world</h1><p>What&#8217;s interesting about many of these algorithms, which can natively reason about the world, is there is a nice framework to write them down <em>&#8211;</em> though it is not matrix multiplication. From diffusion, to transition path sampling, to stochastic linear algebra, we can view these algorithms as simulating thermodynamics. Specifically, we can view them<em> as<a href="https://en.wikipedia.org/wiki/Langevin_equation"> Langevin dynamics</a></em>: a general<a href="https://en.wikipedia.org/wiki/Stochastic_differential_equation"> stochastic differential equation</a>, or SDE, that represents how a system evolves under both deterministic and random forces.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sF2t!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sF2t!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png 424w, https://substackcdn.com/image/fetch/$s_!sF2t!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png 848w, https://substackcdn.com/image/fetch/$s_!sF2t!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png 1272w, https://substackcdn.com/image/fetch/$s_!sF2t!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sF2t!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png" width="522" height="74" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:74,&quot;width&quot;:522,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sF2t!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png 424w, https://substackcdn.com/image/fetch/$s_!sF2t!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png 848w, https://substackcdn.com/image/fetch/$s_!sF2t!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png 1272w, https://substackcdn.com/image/fetch/$s_!sF2t!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F338d7e50-2d90-4e82-982d-5c260eff03a3_522x74.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>We&#8217;re focused on accelerating algorithms that can be expressed as these special SDEs. Similar to how quantum computers<a href="https://permalink.lanl.gov/object/tr?what=info:lanl-repo/lareport/LA-UR-02-4969-02"> emulate</a> Schrodinger's equation, thermodynamic computers emulate Langevin's equations, the natural regime for most of the pertinent physical world: this is a broad class of algorithms that often relate to simulating real-world systems. For example, SDEs are commonly used to simulate problems in<a href="https://en.wikipedia.org/wiki/Molecular_dynamics"> molecular dynamics</a>,<a href="https://en.wikipedia.org/wiki/Computational_fluid_dynamics"> fluid dynamics</a>, and<a href="https://en.wikipedia.org/wiki/Computational_materials_science"> materials science</a>.</p><p>But these SDEs aren&#8217;t just useful for simulating physics! They also have valuable applications in machine learning. Most straightforwardly,<a href="https://en.wikipedia.org/wiki/Diffusion_model">diffusion ML models</a> are based on stochastic differential equations. But there are also many other ML applications that can be expressed as SDEs; for example, the probabilistic sampling required for reasoning with<a href="https://www.cs.toronto.edu/~duvenaud/distill_bayes_net/public/">Bayesian neural networks</a> and energy-based models can be formulated as the simulation of a Langevin equation.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6XXV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6XXV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png 424w, https://substackcdn.com/image/fetch/$s_!6XXV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png 848w, https://substackcdn.com/image/fetch/$s_!6XXV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png 1272w, https://substackcdn.com/image/fetch/$s_!6XXV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6XXV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png" width="561" height="443.4828296703297" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1151,&quot;width&quot;:1456,&quot;resizeWidth&quot;:561,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6XXV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png 424w, https://substackcdn.com/image/fetch/$s_!6XXV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png 848w, https://substackcdn.com/image/fetch/$s_!6XXV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png 1272w, https://substackcdn.com/image/fetch/$s_!6XXV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7e0a82e-89b4-4ce4-8aac-2b46f3e324d1_1600x1265.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://arxiv.org/abs/2304.08818">Blattmann et al, 2021</a></em></figcaption></figure></div><p>Algorithms that have to simulate and reason about the physical world, whether they&#8217;re powered by classical simulation techniques, by deep learning, or by a combination of both, have the potential to create<a href="https://blogs.nvidia.com/blog/nvidia-keynote-at-gtc-2025-ai-news-live-updates/#:~:text=Physical%20AI%20for%20industrial%20and,Cosmos%20platforms%20leading%20the%20way."> $50T+</a> value. But to actually deliver on that promise, they need to break through key computational bottlenecks. Sampling from the trajectory of an SDE is difficult on current hardware. At Normal Computing, we&#8217;re building new hardware to accelerate sampling, and building new algorithms that can reason about the physical world.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/scaling-thermodynamic-computing?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/p/scaling-thermodynamic-computing?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.zach.be/p/scaling-thermodynamic-computing?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>In practice: current hardware for sampling is limited</h1><p>Back in the early days of computational methods, engineers would <a href="https://www.columbia.edu/~mh2078/MonteCarlo/MCS_SDEs.pdf">simulate SDEs</a> and run sampling algorithms, like <a href="https://en.wikipedia.org/wiki/Markov_chain_Monte_Carlo">Markov Chain Monte Carlo</a>, on CPUs. These algorithms can accurately approximate continuous state spaces, and could get asymptotically close to arbitrary probability distributions with enough samples. But these SDE simulations and Markov chain algorithms are fundamentally serial, drawing one sample at a time from a chain. This limits their performance and scalability, especially as the scale and complexity of sampling problems grows.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-1ao!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-1ao!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png 424w, https://substackcdn.com/image/fetch/$s_!-1ao!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png 848w, https://substackcdn.com/image/fetch/$s_!-1ao!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png 1272w, https://substackcdn.com/image/fetch/$s_!-1ao!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-1ao!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png" width="1456" height="296" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:296,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-1ao!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png 424w, https://substackcdn.com/image/fetch/$s_!-1ao!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png 848w, https://substackcdn.com/image/fetch/$s_!-1ao!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png 1272w, https://substackcdn.com/image/fetch/$s_!-1ao!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc49ac0ee-e76a-46a7-84e9-26e84d46f304_1600x325.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption"><em>Sampling: CPUs, GPUs, p-bits, analog s-units</em></figcaption></figure></div><p>Eventually, people started applying GPU acceleration to sampling algorithms and to SDEs, inspired by the unreasonable effectiveness of GPU acceleration on other workloads, like conventional deep learning. GPUs offer <a href="https://arxiv.org/abs/2411.04260v1">better performance through parallelism</a>, but the effectiveness of that parallelism is limited on these sorts of algorithms. Running multiple parallel Markov chains generates more samples faster, but <a href="https://link.springer.com/article/10.1007/s11222-022-10116-z#:~:text=While%20separate%20MCMC%20chains%20can,each%20chain%20would%20not%20change.">doesn&#8217;t eliminate</a> <a href="https://synergy.cs.vt.edu/pubs/papers/wanye-hybrid-sbp-icpp2022.pdf">burn-in time</a> or <a href="https://proceedings.mlr.press/v151/de-souza22a/de-souza22a.pdf">other statistical artifacts</a>; that can only be improved by running longer chains, rather than more chains.</p><p>To really break through performance bottlenecks, we would need an ASIC for sampling. Since 2019, researchers at Purdue and UCSB have been working on probabilistic bits, or p-bits. Each p-bit takes on a probabilistic value between 0 and 1, and interacts with other p-bits through a sparse matrix of interaction terms. These systems can sample from <a href="https://arxiv.org/abs/2110.02481">certain classes of distributions extremely quickly</a>, outputting a valid sample every single clock cycle.</p><p>However, they have to make a lot of sacrifices to achieve that performance. First of all, they only sample from distributions with binary state variables. While a clever programmer can <a href="https://ieeexplore.ieee.org/abstract/document/10873478">construct more complex states from binary state variables</a>, doing so requires significant overhead. Constructing 32-bit state variables from single-bit p-bits requires <a href="https://ieeexplore.ieee.org/abstract/document/10873478">1024x as many coupling terms</a>. Also, non-sparse interaction matrices can cause frustrated sampling, and significantly degrade sample quality. p-bits are well suited to some problems, including simulating spin glasses and binary optimization problems, but I don&#8217;t believe them to be well-suited for general sampling problems. <a href="https://www.ludwigcomputing.com/">Multiple</a> startups have been <a href="https://www.cobi.tech/">founded</a> to try to commercialize p-bit technology, but all seem to be lagging behind Camsari&#8217;s group in terms of practical scalability, so I&#8217;m primarily citing his group&#8217;s work, which represents the state-of-the-art in p-bits.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0gWf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0gWf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png 424w, https://substackcdn.com/image/fetch/$s_!0gWf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png 848w, https://substackcdn.com/image/fetch/$s_!0gWf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png 1272w, https://substackcdn.com/image/fetch/$s_!0gWf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0gWf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png" width="1456" height="389" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:389,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0gWf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png 424w, https://substackcdn.com/image/fetch/$s_!0gWf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png 848w, https://substackcdn.com/image/fetch/$s_!0gWf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png 1272w, https://substackcdn.com/image/fetch/$s_!0gWf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5d203a7-8335-4880-9072-c8db27bb5637_1600x428.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Table: CPUs, GPUs, p-bits, analog s-units</em></figcaption></figure></div><p>At Normal, we&#8217;re working on building sampling hardware that&#8217;s expressive, scalable, and reconfigurable. <a href="https://arxiv.org/abs/2312.04836v1">Our initial proof-of-principle was all-analog</a>, and proves that custom sampling hardware with full-precision states and fully parallel updates is possible. The big challenge with purely analog sampling hardware is its lack of reconfigurability; that first prototype could only sample from Gaussian distributions, which limited its ability to support general sampling workloads. Our new architecture takes the key advantages of that first prototype &#8211; namely, its support for parallel updates of full-precision state variables &#8211; and develops it into a reconfigurable design that can scale up in silicon.</p><h1>Normal Computing: Precise, expressive, and scalable</h1><p>Like I mentioned earlier, we want to be able to sample from a complicated, nonlinear distribution, with a high-dimensional, continuous state space. This means that our sampling hardware needs to fulfill three key properties:</p><ul><li><p>Express this state space precisely, even for continuous variables.</p></li><li><p>Support expressive, parameterizable distributions.</p></li><li><p>Scale to large numbers of state variables.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!F5Uy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!F5Uy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png 424w, https://substackcdn.com/image/fetch/$s_!F5Uy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png 848w, https://substackcdn.com/image/fetch/$s_!F5Uy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png 1272w, https://substackcdn.com/image/fetch/$s_!F5Uy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!F5Uy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png" width="1456" height="752" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:752,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!F5Uy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png 424w, https://substackcdn.com/image/fetch/$s_!F5Uy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png 848w, https://substackcdn.com/image/fetch/$s_!F5Uy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png 1272w, https://substackcdn.com/image/fetch/$s_!F5Uy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F889a21e7-c3ff-4fa0-a994-fd3d4be4830f_1600x826.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Precise, Expressive, Scalable</em></figcaption></figure></div><p>Our architecture, Carnot, is designed from the ground up to solve these three challenges.</p><p>Its state variables are inherently multi-level, enabling us to accurately represent 32-bit numbers without an explosion in the number of required interaction terms. And instead of limiting our nonlinear functions to just sigmoid or ReLU, we&#8217;re implementing reconfigurable nonlinear functions, parameterized as the linear combination of an expressive set of basis functions. Notably, this means that we need to store and compute on our state variables digitally. In this architecture, the states of s-units are no longer <a href="https://arxiv.org/abs/2312.04836">analog voltages on capacitors</a>, but instead 32-bit numbers in digital registers.</p><p>Finally, we&#8217;re implementing a scalable and efficient interaction matrix using some clever architectural techniques. This is one of the most important aspects when it comes to scaling up a system, as an N-dimensional system requires N<sup>2</sup> interaction terms. That means that large interaction terms implemented as <a href="https://arxiv.org/pdf/2110.02481">binary multipliers</a> will struggle to scale for dense interaction matrices; it also means that many all-analog approaches will struggle due to <a href="https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;arnumber=10558647">parasitic and mismatch effects</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H309!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H309!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png 424w, https://substackcdn.com/image/fetch/$s_!H309!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png 848w, https://substackcdn.com/image/fetch/$s_!H309!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png 1272w, https://substackcdn.com/image/fetch/$s_!H309!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H309!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png" width="1456" height="705" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:705,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H309!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png 424w, https://substackcdn.com/image/fetch/$s_!H309!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png 848w, https://substackcdn.com/image/fetch/$s_!H309!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png 1272w, https://substackcdn.com/image/fetch/$s_!H309!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd6c97e3-0f43-43fb-8e26-bb0136fb0025_1600x775.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Significant power and area efficiency that will lead to high scalability</em></figcaption></figure></div><p>More concretely, the Carnot architecture will feature a set of compute tiles, connected together with a network-on-chip (NoC). Each tile will contain a set of digital s-units to store state, a set of reconfigurable nonlinear function units, and an efficient and compact interaction matrix,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> to compute the interactions between states as the system updates. Mapping a sampling problem to this chip entails <a href="https://en.wikipedia.org/wiki/Fokker%E2%80%93Planck_equation">describing the desired distribution with a stochastic differential equation</a> (SDE), and using the chip to evolve that SDE over time.</p><p>By combining these three techniques, we can build a system that solves the three biggest challenges of designing efficient sampling hardware:</p><ul><li><p>Multi-bit s-units allow a single tile to quickly compute complicated quantities, like a 64x64 linear system solution to 32-bit precision.</p></li><li><p>By expressing nonlinearities as a linear combination of basis functions, we can support many different kinds of distributions.</p></li><li><p>Our interaction matrix achieves high TOPS/(Wum<sup>2</sup>), representing significant power and area efficiency that will lead to high scalability.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!soqC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!soqC!,w_424,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif 424w, https://substackcdn.com/image/fetch/$s_!soqC!,w_848,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif 848w, https://substackcdn.com/image/fetch/$s_!soqC!,w_1272,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif 1272w, https://substackcdn.com/image/fetch/$s_!soqC!,w_1456,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!soqC!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif" width="543" height="434.4" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:800,&quot;width&quot;:1000,&quot;resizeWidth&quot;:543,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!soqC!,w_424,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif 424w, https://substackcdn.com/image/fetch/$s_!soqC!,w_848,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif 848w, https://substackcdn.com/image/fetch/$s_!soqC!,w_1272,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif 1272w, https://substackcdn.com/image/fetch/$s_!soqC!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e08cae-8430-45ab-99a4-a061e7da5904_1000x800.gif 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Transition path sampling: emulations of Normal silicon</em></figcaption></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.zach.be/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading zach's tech blog! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Our silicon roadmap</h1><p>And as part of building this technology, we have taken the great responsibility to not do it alone. We currently are partnered with a half dozen of the most important institutions in the world. We&#8217;ve deployed our EDA and other physical (world) AI products at some of these companies. And together we are building a roadmap to establish standards for this compute paradigm.</p><p>Our chip architecture is fundamentally scalable. But there are still a significant number of engineering challenges that need to be solved to actually scale this system to production.</p><p>We&#8217;ll be demonstrating our first chip, which we&#8217;re dubbing Carnot CN101, this year. With four compute tiles containing 64 state variables each, and a reconfigurable NoC to enable tile-to-tile communication, this chip will be able to support 256-dimensional problems with 32-bit state variables. We&#8217;ll de-risk key engineering challenges, including the asynchronous parallel state updates that enable our architecture&#8217;s high performance. CN101 is still a test chip, with relatively low-speed I/O interfaces, but will serve as the first demonstration that scalable thermodynamic computing is a reality.</p><p>In 2026, we want to make key architectural and circuit-level improvements to the Carnot architecture. We&#8217;ll have a more compact interaction matrix, a network-on-chip capable of supporting many more tiles, and support for additional vector math operations. But more importantly, we&#8217;ll introduce on-chip control and caching hardware so that the chip is capable of sampling from multiple different distributions sequentially without off-chip data transfer bottlenecks. We could even sample from one distribution parameterized by samples from a different distribution!</p><p>Finally, to scale up our system to production scale, we need to contend with all of the practical engineering challenges of developing reticle-size chips in modern process nodes with modern interface IP. This is a huge and intensive engineering challenge, but once we solve it, we&#8217;ll be able to deploy thermodynamic computers at data center scale. Each chip will be able to support tens of thousands of state variables, and sample from complex distributions orders of magnitude more efficiently than GPUs. That means that the computational bottlenecks in key algorithms, including physical simulation workloads and probabilistic machine learning, will get solved. This will, in turn, enable new sorts of AI models to reason about the physical world, and reason with uncertainty.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>All performance numbers for conventional approaches are normalized to the process node we&#8217;re manufacturing our first test chips in.</p></div></div>]]></content:encoded></item></channel></rss>