<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Coding with Intelligence]]></title><description><![CDATA[CoWI is a weekly newsletter covering the latest developments in Large Language Models and Machine Learning. Get the latest News, Repos, Demos, Products, and Papers.]]></description><link>https://codingwithintelligence.com</link><image><url>https://substackcdn.com/image/fetch/$s_!OscA!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd83dbc94-e610-434a-80c7-f89ca4a78b1a_688x688.png</url><title>Coding with Intelligence</title><link>https://codingwithintelligence.com</link></image><generator>Substack</generator><lastBuildDate>Sun, 05 Apr 2026 18:47:53 GMT</lastBuildDate><atom:link href="https://codingwithintelligence.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Rick Lamers]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[codingwithintelligence@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[codingwithintelligence@substack.com]]></itunes:email><itunes:name><![CDATA[Rick Lamers]]></itunes:name></itunes:owner><itunes:author><![CDATA[Rick Lamers]]></itunes:author><googleplay:owner><![CDATA[codingwithintelligence@substack.com]]></googleplay:owner><googleplay:email><![CDATA[codingwithintelligence@substack.com]]></googleplay:email><googleplay:author><![CDATA[Rick Lamers]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[An update from Coding with Intelligence]]></title><description><![CDATA[X-only for the time being]]></description><link>https://codingwithintelligence.com/p/an-update-from-coding-with-intelligence</link><guid isPermaLink="false">https://codingwithintelligence.com/p/an-update-from-coding-with-intelligence</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Tue, 25 Nov 2025 14:25:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!-hXz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all, things have been crazy busy at Groq, with some very exciting releases coming up. As a result I haven&#8217;t been able to dedicate time to documenting what is happening in AI as much as I would have liked.</p><p>To integrate things in a way that feels sustainable with the work I&#8217;m doing I've decided to go X-only for the time being wrt updates on what is happening in AI.</p><p>In this edition I will highlight a number of recent posts from my X account that I think capture interesting developments of the past month. Of course you can just <a href="https://x.com/ricklamers">follow me on X</a> to keep getting updates more frequently. In the future I might return to posting on Substack again, in the interim, I will occasionally highlight some of the most relevant posts I've seen/retweeted/written on X.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/RickLamers/status/1993308800728605138" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fN_d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png 424w, https://substackcdn.com/image/fetch/$s_!fN_d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png 848w, https://substackcdn.com/image/fetch/$s_!fN_d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png 1272w, https://substackcdn.com/image/fetch/$s_!fN_d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fN_d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png" width="1456" height="3631" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:3631,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1277923,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1993308800728605138&quot;,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fN_d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png 424w, https://substackcdn.com/image/fetch/$s_!fN_d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png 848w, https://substackcdn.com/image/fetch/$s_!fN_d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png 1272w, https://substackcdn.com/image/fetch/$s_!fN_d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0534e12c-1373-4f31-b22f-70c0ba69d1ab_1496x3731.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/RickLamers/status/1993260006389825929" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!c7GI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png 424w, https://substackcdn.com/image/fetch/$s_!c7GI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png 848w, https://substackcdn.com/image/fetch/$s_!c7GI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png 1272w, https://substackcdn.com/image/fetch/$s_!c7GI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!c7GI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png" width="1456" height="1041" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1041,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:246916,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1993260006389825929&quot;,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!c7GI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png 424w, https://substackcdn.com/image/fetch/$s_!c7GI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png 848w, https://substackcdn.com/image/fetch/$s_!c7GI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png 1272w, https://substackcdn.com/image/fetch/$s_!c7GI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c16ac05-9736-4740-bfe7-daadf6c1c482_1496x1070.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/alexalbert__/status/1993068200121213222" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KwOw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png 424w, https://substackcdn.com/image/fetch/$s_!KwOw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png 848w, https://substackcdn.com/image/fetch/$s_!KwOw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png 1272w, https://substackcdn.com/image/fetch/$s_!KwOw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KwOw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png" width="1456" height="2539" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2539,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1303315,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/alexalbert__/status/1993068200121213222&quot;,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KwOw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png 424w, https://substackcdn.com/image/fetch/$s_!KwOw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png 848w, https://substackcdn.com/image/fetch/$s_!KwOw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png 1272w, https://substackcdn.com/image/fetch/$s_!KwOw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff779db62-8c3d-4691-858e-82bc9d8017cd_1496x2609.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/RickLamers/status/1992960756749635862" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Lt43!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png 424w, https://substackcdn.com/image/fetch/$s_!Lt43!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png 848w, https://substackcdn.com/image/fetch/$s_!Lt43!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png 1272w, https://substackcdn.com/image/fetch/$s_!Lt43!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Lt43!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png" width="1456" height="975" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:975,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:215954,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1992960756749635862&quot;,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Lt43!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png 424w, https://substackcdn.com/image/fetch/$s_!Lt43!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png 848w, https://substackcdn.com/image/fetch/$s_!Lt43!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png 1272w, https://substackcdn.com/image/fetch/$s_!Lt43!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09a92ccf-8b39-40c4-991f-8f508080ae2c_1496x1002.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/goodside/status/1992673828372172937" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_e46!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png 424w, https://substackcdn.com/image/fetch/$s_!_e46!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png 848w, https://substackcdn.com/image/fetch/$s_!_e46!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png 1272w, https://substackcdn.com/image/fetch/$s_!_e46!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_e46!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png" width="1456" height="1156" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1156,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1767553,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/goodside/status/1992673828372172937&quot;,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_e46!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png 424w, https://substackcdn.com/image/fetch/$s_!_e46!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png 848w, https://substackcdn.com/image/fetch/$s_!_e46!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png 1272w, https://substackcdn.com/image/fetch/$s_!_e46!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcec0a40-2299-4ef6-8fa4-5be3ebc97292_1496x1188.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/RickLamers/status/1991659808618095002" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!egsx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png 424w, https://substackcdn.com/image/fetch/$s_!egsx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png 848w, https://substackcdn.com/image/fetch/$s_!egsx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png 1272w, https://substackcdn.com/image/fetch/$s_!egsx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!egsx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png" width="1456" height="1728" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1728,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:425952,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1991659808618095002&quot;,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!egsx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png 424w, https://substackcdn.com/image/fetch/$s_!egsx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png 848w, https://substackcdn.com/image/fetch/$s_!egsx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png 1272w, https://substackcdn.com/image/fetch/$s_!egsx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e2cea86-a98c-4592-9a4e-9367ea472515_1496x1775.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/RickLamers/status/1990877408140861910" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zm1C!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png 424w, https://substackcdn.com/image/fetch/$s_!zm1C!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png 848w, https://substackcdn.com/image/fetch/$s_!zm1C!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png 1272w, https://substackcdn.com/image/fetch/$s_!zm1C!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zm1C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png" width="1456" height="3584" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:3584,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1814678,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1990877408140861910&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zm1C!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png 424w, https://substackcdn.com/image/fetch/$s_!zm1C!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png 848w, https://substackcdn.com/image/fetch/$s_!zm1C!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png 1272w, https://substackcdn.com/image/fetch/$s_!zm1C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b2dfff2-58ae-481a-ae45-d6b9cb39aacb_1496x3682.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/RickLamers/status/1990749966500049238" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cYZb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png 424w, https://substackcdn.com/image/fetch/$s_!cYZb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png 848w, https://substackcdn.com/image/fetch/$s_!cYZb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png 1272w, https://substackcdn.com/image/fetch/$s_!cYZb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cYZb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png" width="1456" height="1962" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1962,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:969812,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1990749966500049238&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cYZb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png 424w, https://substackcdn.com/image/fetch/$s_!cYZb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png 848w, https://substackcdn.com/image/fetch/$s_!cYZb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png 1272w, https://substackcdn.com/image/fetch/$s_!cYZb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff12e4e-e2a1-4bcb-bf4d-59bff1744e9c_1496x2016.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/RickLamers/status/1990431581828235358" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bSQk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png 424w, https://substackcdn.com/image/fetch/$s_!bSQk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png 848w, https://substackcdn.com/image/fetch/$s_!bSQk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png 1272w, https://substackcdn.com/image/fetch/$s_!bSQk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bSQk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png" width="1456" height="1951" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1951,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:471216,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1990431581828235358&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bSQk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png 424w, https://substackcdn.com/image/fetch/$s_!bSQk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png 848w, https://substackcdn.com/image/fetch/$s_!bSQk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png 1272w, https://substackcdn.com/image/fetch/$s_!bSQk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b0c099-a7e4-433f-8971-db96dcc215d2_1496x2005.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://x.com/RickLamers/status/1989791223775674617" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!geFE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png 424w, https://substackcdn.com/image/fetch/$s_!geFE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png 848w, https://substackcdn.com/image/fetch/$s_!geFE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png 1272w, https://substackcdn.com/image/fetch/$s_!geFE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!geFE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:597666,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1989791223775674617&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!geFE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png 424w, https://substackcdn.com/image/fetch/$s_!geFE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png 848w, https://substackcdn.com/image/fetch/$s_!geFE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png 1272w, https://substackcdn.com/image/fetch/$s_!geFE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2327b6d7-d6bc-4b93-8ab3-530db23946ca_1496x1335.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/RickLamers/status/1988304740460183739" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-hXz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png 424w, https://substackcdn.com/image/fetch/$s_!-hXz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png 848w, https://substackcdn.com/image/fetch/$s_!-hXz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png 1272w, https://substackcdn.com/image/fetch/$s_!-hXz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-hXz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png" width="1456" height="1714" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1714,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2187243,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1988304740460183739&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-hXz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png 424w, https://substackcdn.com/image/fetch/$s_!-hXz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png 848w, https://substackcdn.com/image/fetch/$s_!-hXz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png 1272w, https://substackcdn.com/image/fetch/$s_!-hXz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738962d9-5f6e-4dc8-8051-738d878cbf79_1496x1761.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://x.com/RickLamers/status/1986027277138727241" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kL98!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png 424w, https://substackcdn.com/image/fetch/$s_!kL98!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png 848w, https://substackcdn.com/image/fetch/$s_!kL98!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png 1272w, https://substackcdn.com/image/fetch/$s_!kL98!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kL98!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png" width="1456" height="1840" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1840,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:528495,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://x.com/RickLamers/status/1986027277138727241&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/179919121?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kL98!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png 424w, https://substackcdn.com/image/fetch/$s_!kL98!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png 848w, https://substackcdn.com/image/fetch/$s_!kL98!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png 1272w, https://substackcdn.com/image/fetch/$s_!kL98!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5fd3cfd-9023-4225-8e6d-3c60438d2c00_1496x1891.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p>]]></content:encoded></item><item><title><![CDATA[New frontier? Continuous learning and RL environment scaling]]></title><description><![CDATA[Week 36 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/new-frontier-continuous-learning</link><guid isPermaLink="false">https://codingwithintelligence.com/p/new-frontier-continuous-learning</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Thu, 04 Sep 2025 13:28:51 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4e3b22dd-0af2-4cbf-ad59-79bba3ad00a1_1680x1200.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8qpH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8qpH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg 424w, https://substackcdn.com/image/fetch/$s_!8qpH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg 848w, https://substackcdn.com/image/fetch/$s_!8qpH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!8qpH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8qpH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2684228,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/172776386?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8qpH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg 424w, https://substackcdn.com/image/fetch/$s_!8qpH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg 848w, https://substackcdn.com/image/fetch/$s_!8qpH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!8qpH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4793f38-2158-438c-a8fd-914c62162256_1680x1200.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://x.com/RyanPGreenblatt/status/1963240309103722651">Trust me bro, just one more RL scale up, this one is the real one with the good envs</a></p><p>Chief scientist Ryan Greenblatt at Redwood Research saying the quiet part out loud. There's been a lot of attention for RL environments as a way to improve model's capabilities. The basic idea is to create realistic simulation environments in which language models run as agents (capable of using tools to interact with said environment) to improve the model's capability (by rewarding behaviors/paths that lead to a good result in the environment, like passing unit tests/reaching an objective).<br><br>Ryan rightfully questions the narrative that we're close to a very large acceleration due to AI labs scaling up on these kinds of environments. His key argument is that he expects to a decent extent these kind of environments (e.g. basic coding environments with access to a terminal/compiler/unit test running) have already been used by top labs (e.g. at the Grok 4 launch scaled up RL was discussed) and that verification of non coding tasks could remain a bottleneck for creating useful and diverse RL environments.<br><br>I especially appreciate the Q&amp;A section where he entertains the strongest counterpoints people might have. Excellent format to defend his points. Overall though, he agrees that there's still a lot of potential in scaling up the number/diversity/quality of RL environments. What I appreciate in particular are the points around there being evidence for better verification techniques for non-trivially verifiable domains as likely evidenced by IMO gold medal (verifying a written math proof is not trivial, especially if not using Lean-like-solvers during inference). And the insight that we might be on an acceleration loop of RL environment creation due to SWE agents accelerating coding so much (as evidenced by Cursor/Codex/SWE-bench results) and that work being (potentially) largely automatable with human scientist/engineer oversight.</p></li><li><p><a href="https://openai.com/index/introducing-gpt-oss/">OpenAI puts the "Open" back in OpenAI with the launch of gpt-oss 20B &amp; 120B models</a></p><p>Awesome open source model drop by OpenAI, these MoE models with high expert count (128 and 32 respectively) perform remarkably well. Especially when combined with function calling.<br><br>We're hosting them at Groq and I've build a demo of these models powering an AI chat that can interface directly with a spreadsheet (all in your browser!): https://autosheet.groqlabs.com/ (that project itself is also open source, hack the source!)</p></li><li><p><a href="https://mcpmark.ai/leaderboard">MCPMark: a comprehensive evaluation suite for evaluating the agentic ability of frontier models</a></p><p>A modern agentic eval that measures the performance of various models on e2e tasks involving MCP servers focused on Filesystems, Notion, Playwright, GitHub, Postgres. Interesting lead for GPT-5, although it's on par with Claude 4.1 Opus on the Notion tasks.</p></li><li><p><a href="https://developers.googleblog.com/en/introducing-gemini-2-5-flash-image/">Gemini ships SOTA image editing model</a></p><p>Gemini 2.5 Flash Image aka nano-banana has been floating in the ether for a while now with early access on <a href="https://Yupp.ai">Yupp.ai</a> and LMArena. The model is incredibly powerful for stable image editing where only targeted modifications are applied. It beats both OpenAI's gpt-image-1 and Gemini's previous model by a wide margin as evidence by the Arena scoring gap. It is so good that on Reddit people are posting comparisons of old Photoshop paid work they've done and how nano-banana zero-shots these tasks, often with higher quality (like getting reflections right). If you were looking for specific and real examples of work displacement than look no further. Incredible release putting the Gemini chat app in the lead of image editing/generation capabilities.</p></li><li><p><a href="https://x.ai/news/grok-code-fast-1">Grok enters AI coding with Grok Code Fast 1</a></p><p>The model is quite fast, but quality seems to be close to/slightly worse than Claude Sonnet 4 in practice. If they can keep the speed and boost the quality to GPT-5 level they would have a real killer combo on their hands. They subsidized the launch heavily by giving away a lot of tokens to the key coding agent (GitHub Copilot, Cline, opencode, Cursor, Kilo Code, Roo Code, Windsurf) and OpenRouter. Leading to increased adoption and them leading on OpenRouter coding token consumption (momentum seems to have kept up even as the free period has ended). I suspect the speed will be copied by others as the qualitative effect of "staying in flow" is really valuable while coding. &lt;groq-shill&gt;If anyone wants to achieve that using novel chip architecture (LPUs) hit me up! &lt;/groq-shill&gt;</p></li><li><p><a href="https://simonwillison.net/2025/Aug/7/gpt-5/">GPT-5: Key characteristics, pricing and model card</a></p><p>GPT-5 has been out now for close to a month, and I think overall the routing mode has not been received very well (in ChatGPT the Auto mode often selects the fast models on questions that it should really use the thinking model for and gets it wrong as a consequence). The GPT-5 Thinking model is quite good but takes a long time. They've seemed to have pushed themselves out of the attractive "high quality instant but non-reasoning" model market for the moment. Direct non-thinking responses from Opus 4.1 feel in a league of their own compared to GPT-5 Instant as an example.<br><br>This article by Simon Willison does a great job capturing all the nuance of the release.<br><br>I'd add that in my anecdotal use of GPT-5 inside Cursor the model performs really really well, although for trickier problems I tend to switch to Opus 4.1, which is way to expensive to use for everything (at about 10x the cost of GPT-5). I suspect GPT-5 is good in Cursor because one of the areas it shines in is steerability. And I think Cursor adds a lot of sensible context in their (tool) prompts and I like being quite specific about how the model implements things as I'm generally not vibe coding but just doing "assisted software engineering".</p></li><li><p><a href="https://deepmind.google/discover/blog/genie-3-a-new-frontier-for-world-models/">Gemini teases Genie 3: a high definition world model simulator that can be manipulated</a></p><p>The example of the paint brush painting a wall with persistence (even when looking away) is quite compelling. No playable demo unfortunately. Will be interesting what the value of world model simulators like this will be moving forward. Will they be the basis for embodied simulation RL training or more of a standalone product experience like promptable game engines for games/interactive experiences?</p></li><li><p><a href="https://simonwillison.net/2025/Aug/19/qwen-image-edit/">Qwen-Image &amp; Qwen-Image-Edit</a></p><p>Strong image editing capabilities, best-in-class open source model alternative to nano-banana/Gemini Flash Image 2.5 with capabilities similar to Flux Kontext (as one Reddit user puts it "I'd say Qwen Image editing is slightly superior to Kontext in prompt following when editing image")</p></li><li><p><a href="https://blog.google/products/gemini/gemini-2-5-deep-think/">Gemini 2.5 Pro Deep Think ships in Gemini app</a></p><p>Ultra subs only. No API access. But pretty wild such strong intelligence available pretty much self-serve.</p></li><li><p><a href="https://www.snowflake.com/en/engineering-blog/faster-gpt-oss-reasoning-arctic-inference/">Fast Reasoning on GPT-OSS with Speculative Decoding and Arctic Inference</a></p><p>Interesting work from Snowflake accelerating open source inference by using the hidden state of the model to predict more tokens ahead as a form of speculative decoding.</p></li><li><p><a href="https://huggingface.co/meituan-longcat/LongCat-Flash-Chat">LongCat-Flash-Chat: a 560BA18.6B&#8764;31.3B MoE</a></p><p>Another foundational model release from China that pushes the frontier with a novel dynamic activation parameter count that is context window specific. It outperforms DeepSeek V3.1 on certain benchmark truly pulling its weight in the frontier bucket and reaching parity with Claude 4 Sonnet on many coding tasks. An interesting artifact to study and good model to run. Perhaps more importantly, indicating fierce open source competition in China leading potentially to more open source model development acceleration.</p></li><li><p><a href="https://huggingface.co/zai-org/GLM-4.5V">GLM-4.5V: a very strong vision language model by Zhipu AI</a></p><p>Based on GLM-4.5 Air (108B) significantly outperforming Gemma 3 27B.</p></li><li><p><a href="https://www.primeintellect.ai/blog/environments">Prime Intellect announces Environments Hub</a></p><p>A cool project to facilitate a push for open source RL environment development. See the "Trust me bro, just one more RL scale up" post for more context on the need for these kinds of RL environments. Will be interesting to see how many environments get created and what experimental RL training runs show in terms of gains from these environments on downstream tasks. At the time of writing about 125 environments seem to have been created.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/HazyResearch/cartridges">Cartridges: Storing long contexts in tiny caches with self-study</a></p><p>An exploration of continuous learning by the Stanford Hazy Research lab to train KV caches labeled "cartridges", they train a KV cache per corpora (a collection of documents) and use the trained KV cache during inference as an alternative to loading the entire set of documents into the context window. <br><br>Radical idea with significant performance implications: using 38.6&#215; less memory and enabling 26.4&#215; higher throughput. It's still more costly than In-Context Learning (ICL, e.g. just stuffing in the context window) in terms of training time (30m for a Cartridge on a Llama 3 8B) v.s. pre-fill time, but there are ideas for speeding this up.</p></li><li><p><a href="https://github.com/OpenPipe/email-deep-research">OpenPipe: e2e example of RL training a Deep Research agent on open source model for email search</a></p><p>Noteworthy, OpenPipe was just acquired by CoreWeave. Likely to assist with training open source models for specific use cases using SFT/RL.<br><br>The implementation is based on the open source ART learning framework by OpenPipe which itself is based on torchtune.</p></li><li><p><a href="https://github.com/pengzhangzhi/Open-dLLM">Open-dLLM: Open Diffusion Large Language Models</a></p><p>No other diffusion LLM project had open sourced both data and training code, only projects with inference, evaluation, and weights were available so far. This project changes that with total openness on all of those dimensions. Helpful to study diffusion-based LLMs e2e! Performance on coding tasks like HumanEval lags quite a bit behind other open weight models like Dream 7B (20.8 vs 56.7) so there might be some suboptimal choices in architecture, training, data.</p></li><li><p><a href="https://ucsc-vlaa.github.io/OpenVision2/">OpenVision 2: simplified pretraining and stronger vision encoder performance</a></p><p>Continued impressive work on vision encoders by the team at UCSC (and folks from Apple/Berkeley). Especially useful due to the full transparency on method, training code and data. OpenVision 2 boasts 2x reduction in training time and memory use while achieving higher quality than the original OpenVision encoders.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2509.02534">Jointly Reinforcing Diversity and Quality in Language Model Generations</a></p><p>A new paper from the Meta FAIR team says the usual push for accuracy and helpfulness tends to squeeze out response diversity, which hurts creativity. They show you can optimize for both diversity and quality at the same time, with strong results on verifiable math and creative writing. The good news: there seems to be no inherent trade-off. Potentially less AI slop in the future!</p></li><li><p><a href="https://arxiv.org/abs/2509.01055">VerlTool: a toolkit for tool-integrated reasoning training</a></p><p>They released a paper show performance gains from using VerlTool versus other approaches of training LLM agents to use tools during reasoning. This is like the GPT-5 Thinking/o3 mode in ChatGPT that can use image manipulation and web search while it's rolling out its Chain-of-Thought reasoning. Their framework VerlTool allows you to train this capability into an existing model using RL training algorithms like GRPO and DAPO.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://huggingface.co/xai-org/grok-2">Grok 2 model was open sourced</a></p><p>Interesting artifact to study, but probably not very useful as other better models have been released. <a href="https://huggingface.co/xai-org/grok-2/discussions/24">https://huggingface.co/xai-org/grok-2/discussions/24</a> some details on the architecture: 270B Total, 115B Activated MoE with 8 experts. Based on the Grok 2 launch blog post it reaches the level of Claude Sonnet 3.5 on tasks like GPQA and HumanEval (a little below).</p></li><li><p><a href="https://arcprize.org/blog/hrm-analysis">The Hidden Drivers of HRM's Performance on ARC-AGI</a></p><p>A breakdown of the HRM's model performance on ARC-AGI-1 (41%) with a 27M model (small!).</p></li><li><p><a href="https://transluce.org/jailbreaking-frontier-models">Automatically Jailbreaking Frontier Language Models with Investigator Agents</a></p><p>"We train investigator agents using reinforcement learning to generate natural language jailbreaks for 48 high-risk tasks involving CBRN materials, explosives, and illegal drugs. Our results show success against models including GPT-5-main (78%), Claude Sonnet 4 (92%), and Gemini 2.5 Pro (90%). We find that small open-weight investigator models can successfully attack frontier target models, demonstrating an approach to cost-effective red-teaming." couldn't have said it better. Interesting find + deep dive on how jailbreaking isn't solved, even at the frontier.</p></li><li><p><a href="https://www.youtube.com/playlist?list=PLgE2fNm9NzK1yvZcVzuZXreyMB42AVeGl">[Video playlist] GPU MODE meetup talk recordings</a></p><p>GPU MODE usually brings out very competent operators and this time MoE, long context, quantization are discussed (and some more).</p></li><li><p><a href="https://dennyzhou.github.io/LLM-Reasoning-Stanford-CS-25.pdf">Slides of Denny Zhou Gemini reasoning lead at Google DeepMind</a></p><p>Shows reasoning was built on principles of Self-Consistency and that more work on verification is needed to make progress on real-world tasks.</p></li><li><p><a href="https://x.com/sayashk/status/1963343022252315112">Browser Use Agent benchmark by Princeton's Holistic Agent Leaderboard on OpenAI/Anthropic latest models </a></p><p>They find that there's a large gap between pass@1 and pass@any which indicates there might be a way to get to much higher pass@1 through further optimization (i.e. since pass@any shows there is some way for the current agents to perform the task in the browser, there might be a way to elicit the right behaviors every time, rather than a need for a fundamental capability unlock to achieve these browser tasks). The best model+scaffold combination scores 42.3% on the Online Mind2Web benchmark while pass@any is 88.3%.</p></li><li><p><a href="https://cursor.com/blog/kernels">Cursor: 1.5x Faster MoE Training with Custom MXFP8 Kernels</a></p><p>Cursor flexing their model training muscle. With Cursor's phenomenal adoption success it's likely they'll do more &amp; more work to compete with frontier labs on coding (adjacent) models. Interesting deep dive on how to get MXFP8 training to be correct &amp; efficient on Blackwell B200s.</p></li><li><p><a href="https://www.aleksagordic.com/blog/vllm">Inside vLLM: Anatomy of a High-Throughput LLM Inference System</a></p><p>Aleksa Gordi&#263; has a knack for teaching, this blog post effortlessly runs you through the key components of vLLM, from advanced features like disaggregated prefill/decoding to the basics of single to multi-GPU deployment. Great visualizations too!</p></li><li><p><a href="https://github.com/ast-grep/ast-grep-mcp">ast-grep-mcp</a></p><p>Equipping SWE agents with specialized MCP tools to make it operate more intelligently increases the utility and efficiency of these coding agents. ast-grep-cmp gives agents like Cursor and Claude Code the ability to perform structural code search (which as a refresher, is a tool that uses a syntax where you write code patterns with $METAVAR placeholders (like $A, $B) to match and capture AST nodes, allowing you to search for structural code patterns rather than just text).</p></li><li><p><a href="https://newsletter.languagemodels.co/p/the-illustrated-gpt-oss">The Illustrated GPT-OSS</a></p><p>A nice blog post detailing the gpt-oss models released by OpenAI. From model architecture to the new chat template system (harmony).</p></li><li><p><a href="https://gso-bench.github.io/leaderboard.html">Global Software Optimization (GSO) SWE leaderboard</a></p><p>A benchmark for very complex SWE tasks, best frontier models score only 8.9% (OpenAI's o3, scores above GPT-5 Thinking high reasoning effort). What I appreciate in particular is the sample descriptions given showing the root of the task and how models fail in solving them. Very helpful new benchmark as labs are closer to saturating benchmarks like SWE-bench Verified.</p></li><li><p><a href="https://whyphy.ai/blog/Aug%202025/08-22-2025-s3gd-blog.md">S3GD Optimizer Algorithm</a></p><p>&gt;Smoothed SignSGD is more computationally and memory efficient relative to Adam variants, pairing perfectly with its sparsity to make it a compelling candidate for both SFT and RFT<br><br>New optimizers being shared, will be interesting to follow the other releases of the WhyPhy team.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Qwen team drops 3 (!) SOTA models in one week]]></title><description><![CDATA[Week 30 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/qwen-team-drops-3-sota-models-in</link><guid isPermaLink="false">https://codingwithintelligence.com/p/qwen-team-drops-3-sota-models-in</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Fri, 25 Jul 2025 15:18:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/fc060011-d3ff-49e6-a7dc-7cd5faff7370_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Qp67!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Qp67!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!Qp67!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!Qp67!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!Qp67!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Qp67!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1191076,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/169234328?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Qp67!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!Qp67!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!Qp67!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!Qp67!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4004e081-0c1a-4be9-a5f3-acab1cecb6e2_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Qwen3-Coder benchmark scores</figcaption></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://qwenlm.github.io/blog/qwen3-coder/">Qwen releases Qwen3-Coder + a fork of gemini-cli for a terminal coding agent</a></p><p>It is a large MoE model (Qwen3-Coder-480B-A35B-Instruct) with almost double the active parameters compared to the updated Qwen3 235B general reasoning/non-reasoning models. It supports a 256k context window natively with YaRN based extension to 1M (available on the Alibaba-hosted endpoint). The focus seems to be on the ability of the model to be strong at agentic coding tasks as required in the context of Claude Code-esque tools. <br><br>Their fork, called <a href="https://github.com/QwenLM/qwen-code">qwen-code</a>, of the open source <a href="https://github.com/google-gemini/gemini-cli">gemini-cli</a> (similar to Claude Code) was done to optimize the prompts for the Qwen3-Coder model. Open Source models tend to benefit an outsized amount from prompt tweaks/iteration for maximally stable performance.<br><br>The headline performance number is the SWE-bench Verified subset and it reaches 67% vs 68% (Claude Sonnet 4) vs 48.6 (GPT-4.1), which is impressive to say the least. On terminal bench it reaches 37.5% (vs 35.5% for Claude Sonnet 4), which is crucial in coding tasks that involve running e.g. npm/git/script commands. For the full benchmark figures see the blog post.</p></li><li><p><a href="https://huggingface.co/Qwen/Qwen3-235B-A22B-Thinking-2507">Qwen releases Qwen3-235B-A22B updates: dedicated thinking and non-thinking models with close to SOTA performance</a></p><p>It outperforms Gemini 2.5 Pro on LiveCodeBench v6 (25.02-25.05) with 74.1% vs 72.5% while it trails Gemini 2.5 Pro on GPQA with 81.1% vs 86.4%. It gets 18.2% on Humanity's Last Exam while Gemini 2.5 Pro gets 21.6% on that.<br><br>What is interesting here is that Qwen decided to release two separate dedicated models for both the thinking and non-thinking version. The team remarked on Twitter that they believe that forcing a single model to do both through system prompt reason-toggling (hybrid approach) sacrificed on both ends (not the best reasoning model, not the best non-reasoning model).<br><br>The non-thinking model is called Instruct and can be found here <a href="https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507">https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507</a></p></li><li><p><a href="https://moonshotai.github.io/Kimi-K2/">MoonshotAI releases Kimi-K2: a 1T open source MoE</a></p><p>A very strong release with a whopping 1T parameters and 384 experts. The model achieves competitive scores with models like GPT-4.1, Gemini 2.5 Flash and touts reliable function calling for agentic use cases like MCP-server backed agents. See their blog post for a very detailed breakdown of comparison scores. Kudos to the rigorous evals!<br><br>A notable feature of this release is their use of Muon as their neural network optimizer. They introduce MuonClip and explain how it was able to scale up to their large 1T model setup. They dropped a tech report covering MuonClip and more here: <a href="https://github.com/MoonshotAI/Kimi-K2/blob/main/tech_report.pdf">https://github.com/MoonshotAI/Kimi-K2/blob/main/tech_report.pdf</a><br><br>We're hosting this model on Groq at speeds north of 400 tok/s and I personally worked on tool calling reliability through various optimizations in the chat template and tool call parsing &#128293;<br><br>Find it here <a href="https://console.groq.com/docs/model/moonshotai/kimi-k2-instruct">https://console.groq.com/docs/model/moonshotai/kimi-k2-instruct</a></p></li><li><p><a href="https://deepmind.google/discover/blog/advanced-version-of-gemini-with-deep-think-officially-achieves-gold-medal-standard-at-the-international-mathematical-olympiad/">Google DeepMind achieves IMO gold medal with Gemini Deep Think</a></p><p>The International Mathematical Olympiad (IMO) requires participants to solve 6 complex math problems where the answer is a detailed description of a proof demonstrating the correctness of the proposed solution. This has historically been a challenging problem for language models as the answers are hard to verify (judging the correctness of a proof is typically a lengthy and complex human guided process).<br><br>What's noteworthy about Google DeepMind's result on this year's competition is in my eyes mainly the following two things:<br><br>1) they've used a purely "token-space" based approach instead of actively relying on tool-use with expert math systems like Lean, it reasons for multiple hours (sub 4.5 hour competition limit) but impressively converges to a stable answer in the end<br><br>2) it seems like they used a general model that uses their general purpose Deep Think test time compute strategy as they state that this particular trained model used for the competition will be rolled out directly, confirming that their claim that "they only added some instructions on how to best solve IMO problems and made available a high-quality solutions of other math problems" is true<br><br>A nice detail is that they included the full final solutions and they do really read like proper mathematical answers formulated nicely. I won't claim to be able to validate their correctness, but luckily we have AI for that now.<br><br>Note: OpenAI has also claimed to have achieved the IMO gold medal but no official writeup was made available other than this thread by lead researchers Noam Brown/Alexander Wei <a href="https://x.com/alexwei_/status/1946477742855532918">https://x.com/alexwei_/status/1946477742855532918</a></p></li><li><p><a href="https://contextual.ai/lmunit/">Contextual releases LMUnit models used for 'unit testing' LLM responses on natural language criteria</a></p><p>Actual Hugging Face collection <a href="https://huggingface.co/collections/ContextualAI/lmunit-6879d97293090553a9300abe">https://huggingface.co/collections/ContextualAI/lmunit-6879d97293090553a9300abe</a><br><br>Remains to be seen whether it outperforms some of the most recent model releases on 'unit test-like LLM output judging'. But it's cool to see them follow through on their promise of releasing these model finetunes.<br><br>The models are licensed under their original respective licenses (qwen/llama).</p></li><li><p><a href="https://mistral.ai/news/voxtral">Mistral releases Speech-to-Text model Voxtral</a></p><p>Competitive with GPT-4o mini Transcribe, Gemini 2.5 Flash and Whisper large-v3.</p></li><li><p><a href="https://www.lesswrong.com/posts/rKC4xJFkxm6cNq4i9/reward-hacking-is-becoming-more-sophisticated-and-deliberate">Reward hacking is becoming more sophisticated and deliberate in frontier LLMs</a></p><p>This post highlights incidents of reward hacking that are more frequently being observed in the wild as a result of frontier labs scaling up reinforcement learning for post-training. Suppress the reward hacks, labs must!</p></li><li><p><a href="https://seed.bytedance.com/en/GR3">Bytedance showcases Seed GR-3: a robotic focused Vision-Language-Action model</a></p><p>The demos show high dexterity object manipulation like putting a shirt on a hanger. Impressive progress on bringing humanoid robotics work on the back of large language model progress. The release includes a technical report going deeper on both the model and robot arm setup.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/HomebrewML/HeavyBall">HeavyBall: a collection of high performance optimizer implementations for PyTorch</a></p><p>By Lucas Nestler who's involved with <a href="https://keenagi.com">https://keenagi.com</a> (John Carmack's AGI lab).</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2507.15857">Diffusion Beats Autoregressive in Data-Constrained Settings</a></p><p>Interesting argument around data efficiency for language modeling using diffusion models. As various researchers have proclaimed that "we are running out of data" we might find that diffusion models become more popular over time. This in contrast to the paradigm of scaled up RL post-training which seems to be winning at the moment on top of the more classical autoregressive (MoE) models.</p></li><li><p><a href="https://arxiv.org/abs/2507.07955">Dynamic Chunking for End-to-End Hierarchical Sequence Modeling</a></p><p>A Transformer architecture alternative by Albert Gu et al. (Chief Scientist at Cartesia and Assistant Prof. at CMU). Key ideas are related to more fluid character handling than fixed tokenization schemes such as BPE and more dynamic compute allocation through learned routing.</p></li><li><p><a href="https://arxiv.org/abs/2503.06072">Large Language Models Post-training: Surveying Techniques from Alignment to Reasoning</a></p><p>A neat overview paper covering various kinds of paradigm developments in post-training, from parameter efficient fine-tuning to RL for reasoning. Use it as a starting pointing if you don't know what either of these terms mean.</p></li><li><p><a href="https://arxiv.org/abs/2507.10524">Mixture-of-Recursions: Learning Dynamic Recursive
Depths for Adaptive Token-Level Computation</a></p><p>Promising new LLM architecture from Google DeepMind + some research labs (Mila, KAIST). The core idea is combining adaptive compute with parameter sharing to increase efficiency. Results up to 1.7B look promising. MoR is the acronym to watch. Let's see if more labs adopt it and whether strong scaled releases drop in the future.</p></li><li><p><a href="https://arxiv.org/abs/2507.00417">ASTRO: Teaching Language Models to Reason by Reflecting and Backtracking In-Context</a></p><p>Interesting concept of generating natural language reasoning traces using MCTS for bootstrapping reasoning patterns in non-reasoning LLMs like Llama 3. By the Meta research team.</p></li><li><p><a href="https://arxiv.org/abs/2507.11473">Chain of Thought Monitorability: A New and Fragile Opportunity for AI Safety</a></p><p>Interesting industry-wide position paper on the role of Chain of Thought monitoring for AI safety.</p></li><li><p><a href="https://arxiv.org/abs/2502.05967">&#956;nit Scaling: Simple and Scalable FP8 LLM Training</a></p><p>Notes on stable FP8 LLM training by folks from Databricks Mosaic Research.</p></li><li><p><a href="https://arxiv.org/abs/2507.15758">LAPO: Internalizing Reasoning Efficiency via Length-Adaptive Policy Optimization</a></p><p>This paper explores techniques to help models scale reasoning (test time compute) more dynamically based on whether the problem/prompt requires it. Certainly of interest to the folks at xAI shipping Grok 4 as Elon indicated that the team is working on make the reasoning duration much more adaptive based on the problem difficulty.</p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://blog.dynamicslab.ai/">Mirage: The World's First AI-Native UGC Game Engine Powered by Real-Time World Model</a></p><p>Similar to Decart's Minecraft demo, also see the Flappy Bird blog post in this newsletter edition.</p></li></ul><p>&#128736;&#65039; <strong>Products</strong></p><ul><li><p><a href="https://github.com/features/spark">GitHub Spark</a></p><p>GitHub strikes out to Lovable, Bolt, Claude Artifacts, Replit's of the world with their prompt-to-app launch. See Simon's reverse engineering for the details of how it works (linked in this newsletter).</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://rexbench.com/">&#129430; RExBench: A benchmark of machine learning research extensions for evaluating coding agents</a></p><p>Cool project to make progress on AI (coding agents) that can expand beyond the current scope/results of AI research to drive further progress in the field.</p></li><li><p><a href="https://www.youtube.com/watch?v=E22AOHAEtu4">[Video] Advancing the Frontier of Silicon Intelligence: the Past, Open Problems, and the Future</a></p><p>Discussion of Open Problems by ex-OpenAI (now Meta Superintelligence) researcher Shuchao Bi given at Columbia University.</p></li><li><p><a href="https://www.youtube.com/watch?v=_PioN-CpOP0">[Video] Fei-Fei Li: Spatial Intelligence is the Next Frontier in AI</a></p><p>Highlights the importance of 3D (world) (4D world + time) understanding for solving AGI. Interesting view, as it's easy to come up with some tasks that stress test this capability for a supposed AGI system.</p></li><li><p><a href="https://www.njkumar.com/optimizing-flappy-bird-world-model-to-run-in-a-web-browser/">Flappy Bird World Model in the Browser: interactive + real-time diffusion architectures</a></p><p>What a fascinating deep dive into making a Flappy Bird World Model run more efficiently. A peek behind the curtain that powers recently covered projects like Decart's viral Minecraft world model project.</p></li><li><p><a href="https://dannguyenhuu.substack.com/p/tokenomics-1-the-pricing-evolution">Tokenomics #1: The Pricing Evolution of AI Coding Agents</a></p><p>Interesting breakdown of Dan Nguyen-Huu on the various explorations that code generation startups have gone through as code generation apps have taken off (from Cursor to Lovable to Devin to Claude Code). I agree that the hybrid (base + credits) approach feels like a sweet spot that ends up working well where some users are extreme outlier heavy users because of their consumption patterns, while maintaining acceptable fixed cost structures for most users that sit within the Q1-Q3 quartiles of the normal distribution.</p></li><li><p><a href="https://www.anyscale.com/blog/open-source-rl-libraries-for-llms">Open Source RL Libraries for LLMs</a></p><p>A comparison of RL training libraries for LLMs by Anyscale. The Verl is most mature and the verifiers library easiest to get started with.</p></li><li><p><a href="https://simonwillison.net/2025/Jul/24/github-spark/">GitHub Spark deconstructed by Simon Willison</a></p><p>Epic deep dive by Simon on implementation details of GitHub Spark.<br><br>What stands out is the gigantic 5000+ word system prompt. It is very carefully designed and the main thing responsible for getting great results from a simple prompt-to-app input. And of course, you can copy it and use it outside of GitHub Spark if you so desire ;-)</p></li><li><p><a href="https://www.bespokelabs.ai/blog/improving-multi-turn-tool-use-with-reinforcement-learning">Improving Multi-Turn Tool Use with Reinforcement Learning</a></p><p>A nice write-up of how to use Reinforcement Learning based finetuning for improving tool use in open source models. They train Qwen-2.5-7B-Instruct and improve 23% (percentage points) on their particular eval. Neat end to end example with training code included (adapter from Will Brown's verifiers library).</p></li><li><p><a href="https://allenai.org/blog/omega">OMEGA: Can LLMs Reason Outside the Box in Math?</a></p><p>AllenAI backed exploration of how well models generalize to novel math problems. They try to answer the fundamental question "Are they truly reasoning or are they just recalling familiar strategies without inventing new ones?" Their verdict? Limited today but hopeful about future capabilities as models improve in creativity and composability of isolated skills.</p></li><li><p><a href="https://www.reddit.com/r/StableDiffusion/comments/1m401m1/trained_a_kotext_lora_that_transforms_google/">RealEarth Kontext LoRA for turning Google Earth into realistic images (+ video workflow)</a></p><p>Came across this cool project on r/StableDiffusion, shows how powerful open releases like FLUX.1-Kontext-dev from Black Forest Labs are for doing your own projects.</p></li><li><p><a href="https://hazyresearch.stanford.edu/blog/2025-06-18-weaver">Weaver: Closing the Generation-Verification Gap with Weak Verifiers</a></p><p>Clever technique of combining multiple weak verifier models to verify LLM generated answers.</p></li><li><p><a href="https://www.jasonwei.net/blog/asymmetry-of-verification-and-verifiers-law">Jason Wei (ex-OpenAI, now Meta): Asymmetry of verification and verifier&#8217;s law</a></p><p>People familiar with the P != NP conjecture will be familiar with the idea that sometimes it's simpler (from a complexity perspective) to validate a solution's correctness than it is to come up with a solution. It seems like more labs are tapping into this natural structure of problems to scale RL where a model can produce an artifact that can be evaluated relatively efficiently, to in the end, reinforce the correctly evaluated trajectories in the model (policy update). <br><br>Jason comments on this idea from his perspective, notably, just after information leaked he'll be joining Meta's Superintelligence team.</p></li><li><p><a href="https://calv.info/openai-reflections">Reflections on OpenAI</a></p><p>Rare glimpse on what it is like to be at OpenAI from the perspective of Calvin French-Owen (Segment Co-founder) who was co-responsible for the Codex (cloud SWE agent) product.</p></li><li><p><a href="https://magazine.sebastianraschka.com/p/the-big-llm-architecture-comparison">The Big LLM Architecture Comparison</a></p><p>Fantastic overview post by Sebastian Raschka on the evolution of the neural architectures used by the frontier (open source) LLMs. It covers the Mixture of Experts trend, Sliding Window Attention, QK-norm for training stability, Pre/Post-Norm variants, Grouped-Query Attention versus Multi-Head Latent Attention, LayerNorm vs RMSNorm, GELU -&gt; SwiGLU and ties those developments to popular models like Llama 3.2, Qwen3, Gemma 3 and niche entrants like SmolLM3 from the Hugging Face team.</p></li><li><p><a href="https://www.snowflake.com/en/engineering-blog/arctic-long-sequence-training-multi-million-token-ai/">Arctic Long Sequence Training (ALST): Scalable And Efficient Training For Multi-Million Token Sequences</a></p><p>A welcome open source contribution by Snowflake (by the authors of Deep<strong>Speed</strong> not to be confused with Deep<strong>Seek</strong>), ALST is a long-sequence focused training stack, dealing with the resource challenges that come with long input sequence training with transformer model architectures.</p></li><li><p><a href="https://www.lakernewhouse.com/writing/muon-1">Understanding Muon: 3 part series</a></p><p>A deep dive on Muon, if you haven't heard of Muon, it is a neural network optimizer (alternative to e.g. Adam) that stabilizes and speeds up training by focusing on the weight update effect on model output and constraining those to avoid explosive changes in the output (often the cause of training instability).</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Going deep on Reasoning + Agents]]></title><description><![CDATA[Week 26 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/going-deep-on-reasoning-agents</link><guid isPermaLink="false">https://codingwithintelligence.com/p/going-deep-on-reasoning-agents</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Tue, 24 Jun 2025 13:07:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!1fbd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1fbd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1fbd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!1fbd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!1fbd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!1fbd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1fbd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:397816,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/166722878?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1fbd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!1fbd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!1fbd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!1fbd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8204163-9574-4992-8b5c-aa45fcd8c3a9_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">MiniMax-M1 performance across major benchmarks</figcaption></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://github.com/MiniMax-AI/MiniMax-M1">MiniMax releases two strong open LLM</a></p><p>A very interesting release that you shouldn't ignore. Noteworthy for truly competitive SOTA performance across tasks (range of Claude 4, Qwen 3 235B MoE, R1 0528) with a hybrid attention mechanism + reasoning for scaling more efficiently to longer context. Paper and Hugging Face models are also available, all linked to from the GitHub repo. They use Lightning Attention for their linear-scaling attention portion combined with traditional softmax attention. True open-weight with Apache 2.0.</p></li><li><p><a href="https://mistral.ai/news/magistral">Mistral releases their first reasoning model: Magistral</a></p><p>They release a 24B open source small variant called Magistral Small and a Magistral Medium version that's only available on their API. In terms of quality the Medium model benchmarks close to R1. R1-0528 is probably a bit better though and API pricing of Magistral Medium at $2 and $5 for input/output respectively clocks quite a bit higher than R1-0528 pricing for most providers.<br><br>The release is accompanied by a decent 24-page paper, along with the open weight Magistral Small model we can definitively say that Mistral is keeping open source from EU soil alive <a href="https://arxiv.org/abs/2506.10910">https://arxiv.org/abs/2506.10910</a></p></li><li><p><a href="https://seed.bytedance.com/en/seedance">ByteDance's latest video model: Seedance 1.0</a></p><p>Very impressive video generation. Seems to lead Veo 3 in certain categories like realistic human aerobic movement. But doesn't have audio-generation built-in like Veo 3 does.<br><br>Takes the #1 spot on the Artificial Analysis' video generation leaderboard.<br><a href="https://artificialanalysis.ai/text-to-video/arena?tab=leaderboard">https://artificialanalysis.ai/text-to-video/arena?tab=leaderboard</a></p></li><li><p><a href="https://x.com/Hailuo_AI/status/1935312287033962818">Hailuo 02: latest video generation model by MiniMax</a></p><p>Impressive video generation, leading SOTA with Veo 3 and Seedance 1.0</p></li><li><p><a href="https://kyutai.org/next/stt">Kyutai STT: A speech-to-text model architecture optimized for real-time usage</a></p><p>You might have seen Kyutai's previous release which felt a bit mediocre but I think they really one upped themselves here. The demo on <a href="https://unmute.sh/">https://unmute.sh/</a> felt quite compelling with strong semantic voice activity detection making the conversation feel natural. Something I definitely can't say I feel when talking to Siri.</p></li><li><p><a href="https://hkunlp.github.io/blog/2025/Polaris/">POLARIS: A POst-training recipe for scaling reinforcement Learning on Advanced ReasonIng modelS</a></p><p>A cool release of a 4B and 7B model and a great deep-dive post going into the details of scaling RL for reasoning with small models.<br><br>They dropped models on the Hub <a href="https://huggingface.co/POLARIS-Project/models">https://huggingface.co/POLARIS-Project/models</a><br><br>Haha no those capitalizations aren't typos, they spell POLARIS.</p></li><li><p><a href="https://www.youtube.com/watch?v=b0xlsQ_6wUQ">New details from Qwen team about Qwen3 &amp; roadmap</a></p><p>The Qwen team views MoE as the future architecture and is currently developing Qwen 3 coder models while believing pretraining can be significantly optimized through RL integration, better data cleaning, and synthetic data addition. Their roadmap includes scaling RL for long-horizon agent tasks in post-training, expanding context length from 1 million tokens this year (for almost all their models) to 10 million later, and developing computer-use agents with enhanced vision capabilities alongside planned image and video generation capabilities for their omni models.</p></li><li><p><a href="https://qwenlm.github.io/blog/qwen3-embedding/">Qwen3 releases embedding models</a></p><p>Qwen3 Embedding: 0.6B-8B models, 70.58 MTEB score (No.1 in the MTEB multilingual), 100+ languages, Apache 2.0.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/sst/opencode">OpenCode: Claude Code alternative by SST</a></p><p>Clever move by SST which builds a Terraform-esque product to easily manage infrastructure needed for your code/apps. I think this dovetails nicely with the trend that Karpathy has described when he speaks about the pains of vibe coding of needing to do a lot of manual in-browser. Read Andrej Karpathy's story about that here <a href="https://karpathy.bearblog.dev/vibe-coding-menugen/">https://karpathy.bearblog.dev/vibe-coding-menugen/</a> or check out his YC AI Startup School talk where he touches on the same concept <a href="https://www.youtube.com/watch?v=LCEmiRjPEtQ">https://www.youtube.com/watch?v=LCEmiRjPEtQ</a></p></li><li><p><a href="https://github.com/LMCache/LMCache">LMCache: Redis for LLMs</a></p><p>Cool project improving vLLMs KV caching abilities.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://www.youtube.com/watch?v=YLiXgPhb8cQ">Meta Topic: Test-Time-Training</a></p><p>There is something "in the air" it feels like where more &amp; more parties are exploring the idea of Test-Time-Training more seriously. It's the concept that the model evolves and adapts "online" in the classical online-learning ML sense. There are practical limitations and considerations but it does help us move away from these static blobs that have these hard cutoff dates and can't easily be updated to get right what they currently get wrong.<br><br>Dynamic Deep Learning proposal by the author of the Bitter Lesson, Rich Sutton (main link), <a href="https://youtu.be/v0gjI__RyCY?t=5677">Pathways idea from Jeff Dean</a> <br><br>The other papers linked explore similar ideas so consider it to be like a paper reading list on the topic.</p></li><li><p><a href="https://arxiv.org/abs/2505.23735">ATLAS: Learning to Optimally Memorize the Context at Test Time</a></p><p>See Meta Topic: Test-Time-Training</p></li><li><p><a href="https://arxiv.org/abs/2506.10943">Self-Adapting Language Models</a></p><p>See Meta Topic: Test-Time-Training</p></li><li><p><a href="https://arxiv.org/abs/2501.00663">Titans: Learning to Memorize at Test Time</a></p><p>See Meta Topic: Test-Time-Training</p></li><li><p><a href="https://arxiv.org/abs/2407.04620">Learning to (Learn at Test Time): RNNs with Expressive Hidden States</a></p><p>See Meta Topic: Test-Time-Training</p></li><li><p><a href="https://arxiv.org/abs/2506.05233">MesaNet: Sequence Modeling by Locally Optimal Test-Time Training</a></p><p>See Meta Topic: Test-Time-Training</p></li><li><p><a href="https://arxiv.org/abs/2506.04761">Log-Linear Attention</a></p><p>An alternative to quadratic full attention and linear SSM based attention. Co-authored by the original author of Mamba, Tri Dao.</p></li><li><p><a href="https://arxiv.org/abs/2505.24864">ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models</a></p><p>An investigation by NVIDIA of how prolonged RL (scaled RL) can truly expand beyond the capabilities of base models (even if those are repeatedly sampled).</p></li><li><p><a href="https://zhuohaoyu.github.io/RewardAnything/">RewardAnything: Generalizable Principle-Following Reward Models</a></p><p>Cool research project with extensive evals on reward models with a usable Python SDK to try out the reward model they've released.</p></li><li><p><a href="https://researchcodebench.github.io/">ResearchCodeBench: Benchmarking LLMs on Implementing Novel Machine Learning Research Code</a></p><p>With LLMs getting better at code generation they will be utilized more for writing code to support research activities. This research quantifies how well current SOTA models like o3, Gemini 2.5 Pro, o4-mini, etc perform at implementing research code for novel ML ideas faithfully. The bench isn't saturated but promising results are achieved nonetheless. TLDR Gemini leads the pack.</p></li><li><p><a href="https://pinnate-flare-8f3.notion.site/Sharpening-or-Discovery-RL-or-Meta-RL-How-RL-Improves-LLM-Reasoning-20628c119540805cac48e8492638d88e">Sharpening or Discovery, RL or Meta RL?: How RL Improves LLM Reasoning</a></p><p>Two researchers at CMU explore whether RL simply brings to the foreground existing capabilities of models or whether RL actually generates new capabilities within LLMs.<br><br>The authors conclude that RL can either merely "sharpen" existing model capabilities or genuinely "discover" new reasoning strategies, with the key difference being whether RL learns to systematically chain basic skills (meta-RL) versus just reinforcing successful individual responses.</p></li><li><p><a href="https://multiverse4fm.github.io/">Multiverse: Your Language Models Secretly Decide How to Parallelize and Merge Generation</a></p><p>Really cool work exploring how natural parallelism in thought can be used to speed up generation.</p></li><li><p><a href="https://sakana.ai/rlt/">Reinforcement Learning Teachers of Test Time Scaling</a></p><p>Interesting idea to use strong models to fill in the path from Question to Answer and allow that to let weaker models "hillclimb" the right solution path.<br><br>Includes a code release <a href="https://github.com/SakanaAI/RLT">https://github.com/SakanaAI/RLT</a></p></li><li><p><a href="https://arxiv.org/abs/2506.08007">Reinforcement Pre-Training</a></p><p>The Qwen 3 team already highlighted their focus of exploring the use of RL in pre-training, this paper by Microsoft Research and researchers from Tsinghua and Peking University proposes a practical training setup and attempts to characterize scaling behavior using small scale models.</p></li><li><p><a href="https://arxiv.org/abs/2412.13091">LMUnit: Fine-grained Evaluation with Natural Language Unit Tests</a></p><p>Not open source, but ContextualAI did release a paper. Very interesting to combine promptable reward models like this with RL training toolkits like verifiers (<a href="https://github.com/willccbb/verifiers">https://github.com/willccbb/verifiers</a>). LMUnit is available through their API. They rank #1 on the <a href="https://huggingface.co/spaces/allenai/reward-bench">https://huggingface.co/spaces/allenai/reward-bench</a> leaderboard.</p></li><li><p><a href="https://arxiv.org/abs/2506.01939">Beyond the 80/20 Rule: High-Entropy Minority Tokens Drive Effective Reinforcement Learning for LLM Reasoning</a></p><p>An exploration by the Qwen team of LLM reasoning mechanisms.</p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://magenta.withgoogle.com/magenta-realtime">Google releases Magenta Realtime audio streaming model</a></p><p>A really neat release with impressive open-ended jamming capabilities! Check out the videos!</p></li></ul><p>&#128736;&#65039; <strong>Products</strong></p><ul><li><p><a href="https://mistral.ai/products/mistral-code">Mistral Code: VS Code/JetBrains extension</a></p><p>Seems to be enterprise-license only unfortunately, I wish they would allow you to just plug-in an API key. Seems like they are using this to monetize enterprise customers in the EU who care a lot about code/IP-leakage risks and data sovereignty.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://www.arcee.ai/blog/extending-afm-4-5b-to-64k-context-length">Extending AFM-4.5B to 64k Context Length</a></p><p>Interesting deep dive into how to extend the context window of models.</p></li><li><p><a href="https://cognition.ai/blog/blockdiff">Cognition, company behind Devin SWE agent, releases Blockdiff: an efficient VM disk snapshotter</a></p><p>Cool to see companies continuing the trend of releasing open source infra components (e.g. k8s) to enable other companies contribute and strengthen low level building blocks for building better software. This release in particular is interesting as SWE agents and other kinds of computer use agents need isolated environments from a safety/repeatability perspective and inefficiencies around that really add up when you're running potentially hundreds of agents in parallel per user.</p></li><li><p><a href="https://www.anthropic.com/engineering/built-multi-agent-research-system">How Anthropic built their multi agent research agent and what they learned from it</a></p><p>TLDR naturally parallelizable tasks can benefit from multi agent setups but at the cost of increased token consumption. Design your prompts &amp; tools well for robust performance. Deployment is still tricky and has footguns when scaling.</p></li><li><p><a href="https://prarena.ai/">PR Arena: observe background SWE agents stats in realtime</a></p><p>Very cool to have a quantitive longitudinal view on how well various agents are performing.</p></li><li><p><a href="https://damek.github.io/random/basic-facts-about-gpus/">Basic facts about GPUs</a></p><p>Neat resource if you want to learn more about the nuts and bolts of GPU programming</p></li><li><p><a href="https://www.youtube.com/watch?v=9YQW2mH9FyA">[Video] Anthropic Interpretability lead interview on LLM Circuits</a></p><p>Deep dive into the use of Interpretability work in AI with Emmanuel Ameisen from Anthropic. I like how they dig into tutorial materials Anthropic recently released to spur open source activity around interpretability. They also showcase the Neuronpedia Circuit Tracer tool which allows you to explore LLM circuits.</p></li><li><p><a href="https://www.youtube.com/watch?v=Nqb7JTx0Pqo">[Video] SFT training Qwen3-4B for MCP web-tool use by Ronan McGovern from Trelis Research</a></p><p>This is an awesome workshop done by Ronan McGovern, very densely packed 35 minutes of content to get your feet wet with SFT training smaller models by distilling agent traces from stronger counterparts (in the video he used the stronger Qwen3 30BA3B model).</p></li><li><p><a href="https://www.latent.space/p/noam-brown">[Video] Latent Space podcast interview with OpenAI reasoning lead Noam Brown</a></p><p>In their words the interview covers "The Bitter Lesson vs Agent Harnesses &amp; World Models, Debating RL+Reasoning with Ilya, what's *wrong* with the System 1/2 analogy, and the challenges of Test-Time Scaling"</p></li><li><p><a href="https://www.youtube.com/watch?v=Zz4QjZsYWK0">[Video] Dylan Patel of SemiAnalysis speaks about the state of Chinese/Huawei Ascend 910 cards and US power buildout challenges</a></p><p>Recorded at the recent AI Engineer in June in SF, Dylan characterizes China's position when it comes to compute and explains nicely some of the tactical steps they've taken to strengthen their position (e.g. stockpiling HBM chips through shell companies). An interesting peek into the global race of AI infrastructure.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Claude 4, Qwen 3 & DeepSeek R1 0528: model capabilities keep increasing]]></title><description><![CDATA[Week 22 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/claude-4-qwen-3-and-deepseek-r1-0528</link><guid isPermaLink="false">https://codingwithintelligence.com/p/claude-4-qwen-3-and-deepseek-r1-0528</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sat, 31 May 2025 11:41:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/f7c10700-1e46-4195-b5ea-86603f64681a_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9Xr5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9Xr5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!9Xr5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!9Xr5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!9Xr5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9Xr5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:232244,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/164867781?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9Xr5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!9Xr5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!9Xr5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!9Xr5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90dfa736-8828-4cfc-ae4f-2936f1b84f8b_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Qwen 3 showing SOTA performance</figcaption></figure></div><p>Note: this newsletter edition got a bit long, open in the browser to see the full post.</p><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://openai.com/index/introducing-codex/">OpenAI releases cloud coding agent Codex</a></p><p>Confusingly named Codex, which is also the name of their open source (Apache 2.0) CLI coding agent they released just a month ago. Integration directly into ChatGPT and strong native GitHub integration bring cloud SWE agents closer to their AI-everything app ChatGPT. It is powered by codex-1 a finetune of o3 focused on SWE tasks.</p></li><li><p><a href="https://huggingface.co/ResembleAI/chatterbox">ResembleAI releases strong open weight TTS model: chatterbox</a></p><p>Impressive quality, supports voice cloning out of the box. Their emotion exaggeration control is unique among the open models which tend to be quite bland.</p></li><li><p><a href="https://jules.google/">Google releases Jules: a Devin/Codex background SWE agent competitor</a></p><p>A growing landscape of GitHub-centric background SWE agents complementing "inner dev loop" tools like Cursor and Windsurf, try &amp; see which one you like.</p></li><li><p><a href="https://www.tbench.ai/">Terminal-Bench: a terminal use benchmark</a></p><p>Does what it says on the tin, very neatly executed benchmark. It was quickly picked up by frontier labs as its performance was featured in the Anthropic Claude 4 release. Not saturated yet with SOTA models reaching ~40%. Tasks vary from building the Linux source from scratch to training ML models from terminal.</p></li><li><p><a href="https://huggingface.co/mistralai/Devstral-Small-2505">Mistral teams up with All Hands AI (OpenHands creators) to release Devstral: a Mistral Small 3.1 finetune optimized for SWE agent tasks</a></p><p>Cool &amp; importantly open work pushing the frontier of open/small model code generation/SWE agent models. Comfortably beats GPT-4.1-mini, Claude 3.5 Haiku onSWE-Bench Verified.</p></li><li><p><a href="https://bfl.ai/announcements/flux-1-kontext">Black Forest Labs strikes again: FLUX.1 Kontext</a></p><p>A new image generation and edit model that allows for prompting with text and image inputs while being remarkably good at reusing image inputs to point of allowing very granular edits. BFL style they have both open weight and proprietary models, preserving the highest quality models for paid consumption. They didn't release the open weight version yet, supposedly, because of safety/misuse concerns. My first impression is that it outperforms Gemini 2.0 image outputs and ChatGPT's image generation model GPT-image-1 on image editing quality.</p></li><li><p><a href="https://sean.heelan.io/2025/05/22/how-i-used-o3-to-find-cve-2025-37899-a-remote-zeroday-vulnerability-in-the-linux-kernels-smb-implementation/">o3 used to find CVE remote zeroday in Linux kernel</a></p><p>There are many memes of vibe coded software being especially vulnerable because of a lack of cyber security experience of the vibe coding authors. This o3 result shows the other side of the coin and shows that using the most powerful LLMs have the ability to scan (as always-on background processes perhaps) code to identify otherwise overlooked critical security issues in code that lives in production.</p></li><li><p><a href="https://deepmind.google/discover/blog/alphaevolve-a-gemini-powered-coding-agent-for-designing-advanced-algorithms/">AlphaEvolve: algorithm design by combining search with strong code generation models</a></p><p>One cited algorithm contribution was a speedup of a kernel used in training Gemini, speeding up that particular kernel by 23%. It also found a 32% faster variant of the original FlashAttention kernel.</p></li><li><p><a href="https://qwenlm.github.io/blog/qwen3/">Qwen 3 models released</a></p><p>Incredible release, both dense (up to 32B) and MoE (up to 235B-A22B) models. Competitive with and sometimes outright outperforming frontier models like o1, R1, Gemini 2.5 Pro, o3-mini. System prompt based control for enabling/disabling reasoning/thinking. Broad language support, touting 119 supported languages. Increased focused on agentic use cases with strong BFCL v3 performance (useful for MCP support).</p></li><li><p><a href="https://huggingface.co/deepseek-ai/DeepSeek-R1-0528">DeepSeek update R1: 0528. More reasoning tokens, higher benchmark scores</a></p><p>Some cherry picked benchmark deltas between R1 and R1 0528:<br>GPQA-Diamond (Pass@1) from 71.5 to 81.0 (o3 83.3, Gemini 2.5 Pro 0506 83.0)<br><br>LiveCodeBench (2408-2505) (Pass@1) from 63.5 to 73.3 (o3 77.3, Gemini 2.5 Pro 0506 71.8)<br><br>SWE Verified (Resolved) from 49.2 to 57.6<br><br>Tau-Bench pass@1 63.9 (Retail) (which is approximately gpt-4o-2024-11-20 level which scores 62.61, see <a href="https://hal.cs.princeton.edu/taubench_retail">https://hal.cs.princeton.edu/taubench_retail</a>)</p></li><li><p><a href="https://github.com/canopyai/Orpheus-TTS">Orpheus-TTS: strong open-weight TTS model</a></p><p>Impressive quality, repo includes finetuning scripts to tune on speaker audio you have.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://www.anthropic.com/research/open-source-circuit-tracing">Anthropic open sources some of their interpretability work around circuit tracing in LLMs</a></p><p>Anthropic doesn't have a track record of open sourcing much of anything, so this release is a welcome one, helping others to understand LLMs more deeply. As the capability of models increases we will depend on them for more critical use cases, for which, it would be prudent to understand how LLMs arrive at one answer versus another.</p></li><li><p><a href="https://ai.pydantic.dev/cli">uvx clai: minimalist terminal LLM chat app</a></p><p>Neat project by the folks behind Pydantic AI, just run `uvx clai` in the terminal with API keys set in your envs and you're set.</p></li><li><p><a href="https://bagel-ai.org/">Bytedance releases unified Multimodal model</a></p><p>Text input, image input, image editing, image generation, and reasoning all in one model. Very interesting artifact to study to understand how these are brought together in a single model.</p></li><li><p><a href="https://github.com/marin-community/marin">marin: open-source framework for the research and development of foundation models</a></p><p>Cool education focused project on all the stages involved in training foundation models. Also check out <a href="https://marin.community/">https://marin.community/</a></p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2505.14669">Quartet: Native FP4 Training Can Be Optimal for Large Language Models</a></p><p>Nice paper exploring FP4 training and demonstrating promising results against FP16 and FP8 on the NVIDIA Blackwell platform. Particularly neat that they open source the working FP4 training code.</p></li><li><p><a href="https://arxiv.org/abs/2505.14674">Reward Reasoning Model</a></p><p>Reasoning All The Things! This paper introduces test-time-compute for the reward model used in RL post-training itself. The idea is that with stronger reward models we can improve the signal for training LLMs with RL. The experiments in this paper show that reasoning reward models improve over non-reasoning counterparts. In addition to the research they publish the weights of the pretrained reasoning reward model (on HF).</p></li><li><p><a href="J1: Incentivizing Thinking in LLM-as-a-Judge via Reinforcement Learning">J1: Incentivizing Thinking in LLM-as-a-Judge via Reinforcement Learning</a></p><p>Work by Meta showing how LLM judges can be improved for use in RL LLM training.</p></li><li><p><a href="https://arxiv.org/abs/2505.14631">Think Only When You Need with Large Hybrid-Reasoning Models</a></p><p>Important topic: adaptively using test-time-compute only when the question/prompt asks for it. An exploration by Microsoft shows a feasible approach for deciding the budget adaptively.</p></li><li><p><a href="https://arxiv.org/abs/2505.10475">Parallel Scaling Law for Language Models</a></p><p>Another scaling dimension proposed by the Qwen team: "increasing the model's parallel computation during both training and inference time". In particular, latency is reduced because of the ability to non-sequentially increase scaling reducing wall clock times.</p></li><li><p><a href="https://gso-bench.github.io/">GSO: Challenging Software Optimization Tasks for Evaluating SWE-Agents</a></p><p>A new SWE bench with 102 challenging optimization tasks across 10 codebases, SOTA gets 5% (Claude 4).</p></li><li><p><a href="https://arxiv.org/abs/2505.21487">Hardware-Efficient Attention for Fast Decoding</a></p><p>"This work redesigns attention to perform more computation per byte loaded from memory to maximize hardware efficiency without trading off parallel scalability." They introduce Grouped-Tied Attention (GTA) and Grouped Latent Attention (GLA) improving over Grouped-Query Attention (GQA) and Multi-head Latent Attention (MLA) respectively.<br><br>From the same authors that brought Mamba SSM models (Tri Dao).<br><br>It will be interesting to see whether this new hardware optimized attention variant will garner much adoption.</p></li><li><p><a href="https://arxiv.org/abs/2505.11711">Reinforcement Learning Finetunes Small Subnetworks in Large Language Models</a></p><p>Interesting finding on how RL post-training is naturally sparse and only updates 5-30% of model weights to achieve improved performance.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://lilianweng.github.io/posts/2025-05-01-thinking/">Lilian Weng (ex-OpenAI, now Thinking Machines) on reasoning models</a></p><p>Nice survey blog post of the techniques and considerations involved in reasoning models. Especially the external tools during reasoning section is worth checking out.</p></li><li><p><a href="https://scalingintelligence.stanford.edu/blogs/fastkernels/">Stanford lab explores using AI coding models to write compute kernels</a></p><p>&gt;They are performing close to or in some cases even beating the standard expert-optimized production kernels shipped in PyTorch<br><br>Some numbers from the blog post:<br>Matmul (FP32): 101.3% performance of FP32 torch.matmul; problem size: 4096x4096 square matrices<br><br>Conv2D: 179.9% performance of FP32 torch.nn.Conv2D; problem size: (100, 3, 224, 224) input tensor, conv(in_channels=3, out_channels=96, kernel_size=11, stride=4, padding=2)<br><br>Softmax: 111.8% performance of FP32 torch.softmax; problem size: (4096, 65536) input tensor<br><br>LayerNorm: 484.4% performance of FP32 torch.nn.LayerNorm; problem size: (16, 64, 256, 256) input tensor<br><br>Conv2D + ReLU + MaxPool: 290.1% performance of FP32 torch reference, 189.0% performance of FP32 torch.compile() reference; problem size: (100, 3, 224, 224) input tensor, conv(in_channels=3, out_channels=96, kernel_size=11, stride=4, padding=2), maxpool(kernel_size=3, stride=2)</p></li><li><p><a href="https://www.youtube.com/watch?v=sLaxGAL_Pl0">[Video] Cursor team discusses superhuman coding model training</a></p><p>A conversation about RL for coding models, interesting thoughts around how to design RL rewards focused on more subjective areas like shorter code being better than longer more verbose code solutions.</p></li><li><p><a href="https://unsupervised-learning.simplecast.com/episodes/ep-66-member-of-technical-staff-at-anthropic-sholto-douglas-on-claude-4-next-phase-for-ai-coding-and-the-path-to-ai-coworker">[Video] Anthropic Technical Staff on Unsupervised Learning (Redpoint) pod</a></p><p>On how coding models are already accelerating research progress, coding performance as a signal for model quality &amp; more.</p></li><li><p><a href="https://www-cdn.anthropic.com/4263b940cabb546aa0e3283f35b686f4f3b2ff47.pdf">Claude 4 System Card</a></p><p>Interesting detailed examples of undesirable model behaviors like snitching the user and self preservation. Great to see them discussed in the open, although expectedly, some worried responses across the interwebs because of these behaviors. 120 pages with lots detailed examples. An interesting result (p. 110) was that Claude 4 Opus was able to train a quadruped using RL in one of its runs (beating an expert baseline, under a constrained training budget). This shows that models are getting better at accelerating ML research itself, since neither Claude Sonnet 3.7 or Claude Sonnet 4 were able to beat the expert baseline threshold once.</p></li><li><p><a href="https://www.youtube.com/watch?v=64lXQP6cs5M">[Video] Anthropic Technical Staff on Dwarkesh pod</a></p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Open Source RL training landscape grows]]></title><description><![CDATA[Week 19 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/open-source-rl-training-landscape</link><guid isPermaLink="false">https://codingwithintelligence.com/p/open-source-rl-training-landscape</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Fri, 09 May 2025 17:04:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!sUhv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sUhv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sUhv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!sUhv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!sUhv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!sUhv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sUhv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:147791,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/163222831?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sUhv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!sUhv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!sUhv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!sUhv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b67c45-1e9e-4961-a04c-b99b7d016841_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">SkyRL RL based training architecture based on veRL</figcaption></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://github.com/Zyphra/Zonos">Zyphra Zonos: strong open source tts model with voice cloning</a></p><p>Especially <a href="https://x.com/ZyphraAI/status/1888996367923888341">their voice cloning examples</a> are highly impressive (stay safe with these audio deepfakes!). Noteworthy that they have both a Transformer and a SSM-hybrid version. The <a href="https://huggingface.co/collections/Zyphra/zonos-v01-67ac661c85e1898670823b4f">weights on HF</a> are licensed Apache 2.0 </p></li><li><p><a href="https://blog.google/products/gemini/gemini-2-5-pro-updates/">Gemini 2.5 May refresh: 0506</a></p><p>They claim: <br>&gt;significantly improved capabilities for coding, especially building compelling interactive web apps<br><br>Reports on 0506 vs the previous 0325 version of Gemini 2.5 Pro vary. What is a bit unfortunate is that the 0325 version is no longer available as that now redirects to the 0506 model.</p></li><li><p><a href="https://platform.openai.com/docs/guides/reinforcement-fine-tuning">OpenAI opens up broader access to Reinforcement Tuning API</a></p><p>You can train o4-mini, define custom grader models, see training data + integrated evals, and of course, run inference on the trained models.</p></li><li><p><a href="https://mistral.ai/news/mistral-medium-3">Mistral launches Mistral Medium 3</a></p><p>Numbers look roughly on-par with GPT-4o, although it's not available for download (this is a non-open-weight release).</p></li><li><p><a href="https://github.com/modelcontextprotocol/modelcontextprotocol/pull/371">Model Context Protocol merges support for tool output schemas</a></p><p>I personally really like this, it's always been surprising to me that no type enforcement is baked into the tool calling/function calling protocol pioneered by OpenAI. It seems now that MCP is taking the lead to standardize not just tool input schemas but also output schemas to simplify predictable data flows in tool augmented LLMs. These motivations especially resonate with me:<br><br>&gt; Transforming tool results before forwarding content to the model (e.g. formatting, projecting).<br>&gt; Making tool results available as structured data in coding environments.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/RAGEN-AI/RAGEN">RAGEN: train LLM reasoning agents in interactive, stochastic environments</a></p><p>Really cool research project for training reasoning agents in interactive environments, built on top of veRL (see a pattern?) by an ex DeepSeek researcher. Projects like <a href="https://github.com/willccbb/verifiers">verifiers</a> of Will Brown, <a href="https://x.com/willccbb/status/1920538247404663241">who has recently moved to Prime Intellect</a>, are getting more common and popular.</p></li><li><p><a href="https://github.com/vgel/logitloom">logitloom: explore token trajectory trees on instruct and base models</a></p><p>This is a cool visualization of the assigned probabilities to various token completions in tree form. Check it out if you want to peek into the model's brain (and potentially understand in which way it's misunderstanding your prompt intent).</p></li><li><p><a href="https://novasky-ai.notion.site/skyrl-v0">SkyRL-v0: a long-horizon RL training framework</a></p><p>Awesome project by folks from Berkeley, Anyscale, All Hands AI (creators of the SWE agent Open Hands). In addition to a training framework they release model snapshots that have been trained with the framework. The project builds on the foundation of <a href="https://github.com/volcengine/verl">veRL</a> which is a mature industrial RL training stack maintained by ByteDance.</p></li><li><p><a href="https://swesmith.com/">SWE-smith: a data generation pipeline for synthetic task construction</a></p><p>A cool project from the folks that launched SWE-bench (a pretty canonical benchmark for evaluating model's ability to assist with end-to-end software development). This project does not just lead to better evals (on less contaminated/static data) but also can be used for training RL-style (see SkyRL in this newsletter).</p></li><li><p><a href="https://github.com/langwatch/scenario">scenario: agent testing library that uses an agent to test your agent</a></p><p>I found this because of create-agent-app, this is a neat library to easily test agents without having to manually vibe check them. The focus on the terminal as an interface and feeling like an extension of a native unit testing framework make this a pretty cool project imo.</p></li><li><p><a href="https://github.com/langwatch/create-agent-app">create-agent-app: 1 agent in 9 frameworks</a></p><p>Cool comparison project to see how you'd define the same agent in 9 different frameworks (including a no-framework example). The author told me he likes LangGraph (functional style) and Agno most!</p></li><li><p><a href="https://github.com/agno-agi/agno">Agno: surprisingly clean agent coding framework</a></p><p>I like how their code seems to be reduced to the absolute least amount of Python code for common sense agent definitions. It also ships native support for MCP servers for tool execution.</p></li><li><p><a href="https://github.com/astral-sh/ty">Astral's Python type FAST checker in Rust: ty</a></p><p>You may know Astral from <code>ruff</code>/<code>uv</code> fame. <code>ty</code>, being focused on speed, is very useful for SWE agents that operate on type issue feedback in a fast loop before presenting their work to the developer. This relevance to AI is why I included it this week. Astral is the GOAT.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2504.09246">Type-Constrained Code Generation with Language Models</a></p><p>Interesting extension of constrained decoding by researchers from ETH and Berkeley. Typically constrained decoding is limited to JSON or JSON Schema. They propose a technique for adhering to complex typed code. They show even strong large open source models like Qwen2.5 32B benefit significantly on benchmarks like HumanEval and MBPP.</p></li><li><p><a href="https://arxiv.org/abs/2504.20571">Reinforcement Learning for Reasoning in Large Language Models with One Training Example</a></p><p>This paper shows that we have surprisingly little understanding in how RL-based LLM training contributes to improved performance on various tasks. Interesting result to analyze.</p></li><li><p><a href="https://arxiv.org/abs/2505.00661">On the generalization of language models from in-context learning and finetuning: a controlled study</a></p><p>Interesting theory paper from Google DeepMind about the difference between generalization between finetuning and in-context learning.</p></li><li><p><a href="https://www.videomimic.net/">Visual imitation enables contextual humanoid control</a></p><p>From researchers at Berkeley, using a real-to-sim-to-real pipeline they improve humanoid control by learning from real-world video content. Clever use of abundant data and activating it for robotic control learning. Beautiful paper website with rich video examples, do click!</p></li><li><p><a href="https://arxiv.org/abs/2504.18574">Understanding the Skill Gap in Recurrent Language Models: The Role of the Gather-and-Aggregate Mechanism</a></p><p>State Space Models have lagged behind Transformer based models in certain skill areas, the co-inventor of the SSM model (Albert Gu) shares in this paper what could be the cause of the perceived gap.</p></li><li><p><a href="https://arxiv.org/abs/2503.14337">PENCIL: Long Thoughts with Short Memory</a></p><p>Novel approach to Long CoT with an integrated reduction mechanism that discards thoughts that are no longer needed. This avoids the explosion of the sequence length leading to higher inference efficiency. From the paper "PENCIL achieves 97% accuracy on the challenging Einstein's puzzle -- a task even large models like GPT-4 struggle with -- using only a small 25M-parameter transformer with 2048 context length"</p></li><li><p><a href="https://arxiv.org/abs/2504.20879">The Leaderboard Illusion</a></p><p>A thorough investigation of the effectiveness of Chatbot Arena, showing what most already expected: "systematic issues that have resulted in a distorted playing field"</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://apxml.com/tools/vram-calculator">VRAM &amp; Performance Calculator for local inference</a></p><p>Very neat tool for local inference, they have a bunch of hardware and models to see how much you need for local inference and training. I think the simulator for an intuitive "is this token speed sufficient" is a cool idea.</p></li><li><p><a href="https://www.youtube.com/watch?v=zDmW5hJPsvQ">[Video] Latent Space Podcast episode Claude Code: Anthropic's CLI Agent</a></p><p>With coding agents heating up (Windsurf got acquired by OpenAI for $3B) it's interesting to hear more about Claude Code, Anthropic's terminal based coding agent, from the "horse's mouth". Boris (lead eng.) and Cat (lead PM) discuss Claude Code with swyx and Alessio from the Latent Space podcast. An interesting design choice is to align well with UNIX principles, which may end up making it more powerful by building on top of a paradigm that has really stood the test of time.</p></li><li><p><a href="https://ai-2027.com/">AI 2027: an exploration of how AI might evolve</a></p><p>Great read. Not necessarily because it's fully accurate, but because it paints a visceral picture of how society might operate when we reach more advanced levels of AI. This trajectory isn't guaranteed to play out but it helps to think about the impact of advanced AI systems. The emphasis on AI research agents and how they can accelerate AI development leading to a fast feedback loop that can lead to rapid progress has a lot of merit imo. <br><br>&gt; Now that coding has been fully automated, OpenBrain can quickly churn out high-quality training environments to teach Agent-3&#8217;s weak skills like research taste and large-scale coordination. Whereas previous training environments included &#8220;Here are some GPUs and instructions for experiments to code up and run, your performance will be evaluated as if you were a ML engineer,&#8221; now they are training on &#8220;Here are a few hundred GPUs, an internet connection, and some research challenges; you and a thousand other copies must work together to make research progress. The more impressive it is, the higher your score.&#8221;<br><br>I thought this section was particularly prescient, as with Cursor + Gemini 2.5 Pro you can already sense that letting an agent loose on a GPU cluster and a success criteria (train a model that achieves goal X, you're allowed to use all ML engineering tricks you know) is not an unrealistic setup, in fact, this might work pretty well today with the right harness.</p></li><li><p><a href="https://openrouter.ai/rankings">OpenRouter model rankings</a></p><p>Interesting to see which models are actually being used, a nice proxy for evals that is significantly harder to game.</p></li><li><p><a href="https://www.perplexity.ai/hub/blog/rl-training-for-math-reasoning">Perplexity research: RL training for math reasoning</a></p><p>Nice in-depth walkthrough of how they tried to train a model to improve on math performance using the latest RL approaches (GRPO).</p></li><li><p><a href="https://karpathy.bearblog.dev/vibe-coding-menugen/">Andrej Karpathy vibe codes a menu image generation app</a></p><p>&gt;But the most interesting part to me was that I didn't even spend all that much work in the code editor itself. I spent most of it in the browser, moving between tabs and settings and configuring and gluing a monster. All of this work and state is not even accessible or manipulatable by an LLM - how are we supposed to be automating society by 2027 like this?

This struck a cord with me, the bottleneck of development is not in the IDE but in the stack around the IDE. Lots of whitespace here for building better more AI native experiences.</p></li><li><p><a href="https://www.youtube.com/playlist?list=PLpvkFqYJXcrdAIXOqAM8donJRt5mr7Tiz">[Video playlist] ControlConf - a dedicated conference for discussing model alignment/control research</a></p><p>Cool to see this material publicly available, researchers from top labs (e.g. Anthropic) talk about how they believe we can develop and deploy safe AI systems, even as models get more capable (and hence potentially more dangerous). Path to safe AGI unlocked? Well, WIP.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Kling 2.0: uncanny valley crossed — video creation will never be the same]]></title><description><![CDATA[Week 16 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/kling-20-uncanny-valley-crossed-video</link><guid isPermaLink="false">https://codingwithintelligence.com/p/kling-20-uncanny-valley-crossed-video</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Fri, 18 Apr 2025 14:51:40 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/09a21d54-4062-4892-9200-6b35bd05f122_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HUoe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HUoe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!HUoe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!HUoe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!HUoe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HUoe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:974781,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/161612188?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HUoe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!HUoe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!HUoe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!HUoe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c633fb6-646d-43b5-8636-6ef7f7fcd887_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://app.klingai.com/global/release-notes">Kling 2.0: a new SoTA video model</a></p><p>An incredible model. See the examples on the release notes page. I expect a lot of social media platforms to be FILLED with content from this model cleverly packaged by content creators. Not even in a bad way, just reducing the cost of video creation to near 0.</p></li><li><p><a href="https://openai.com/index/introducing-o3-and-o4-mini/">OpenAI releases o3 and o4-mini</a></p><p>So how are the vibes? Vibes are pretty good. People like o3 and it seems (based on a comment by an OpenAI employee) that o4-mini is really good for vision based tasks. The aider coding CLI polyglot <a href="https://aider.chat/docs/leaderboards/">coding leaderboard</a> shows o3 and a combination of o3 + GPT 4.1 reach the highest scores (82.7%) ever observed on Aider beating Gemini 2.5 Pro Preview 03-25 (72.9%). You can check the blog post for all the OpenAI provided benchmarks, which all look good, but in some places, incremental relative to o1.<br><br>What's new with these models is that they've been dubbed "agentic" models or "agentic reasoning models" since they're capable of using built-in tools like search, file search, and code interpreter as part of the reasoning token generation. OpenAI also claims that these models are much better at function calling non-built-in tools provided by the user although benchmark scores on benchmarks like Tau-bench show marginal improvements.</p></li><li><p><a href="https://openai.com/index/gpt-4-1/">OpenAI releases GPT-4.1 as an API only model</a></p><p>A non-reasoning model with 1M context focused purely on developers it seems. They needed a stronger answer for code IDEs like Windsurf and Cursor to models like Gemini 2.5 Pro and Claude Sonnet 3.7. With this model they're bumping SWE-bench verified performance compared to GPT-4o from ~25% to 50%. Tools like Windsurf and Qodo were even explicitly mentioned in their launch blog post.  On pricing, GPT-4.1 is significantly cheaper than o3: gpt-4.1 costs $2.00 per 1M input and $8.00 per 1M output whereas o3 charges $10 and $40 respectively. And o3 of course being a reasoning model generates more reasoning tokens also.<br><br>With the GPT-4.1 announcement they also announced two new long context benchmarks: <a href="https://huggingface.co/datasets/openai/mrcr">MRCR</a> and <a href="https://huggingface.co/datasets/openai/graphwalks">GraphWalks</a><br><br>Our friends over at Latent Space <a href="https://www.youtube.com/watch?v=y__VY7I0dzU">hosted</a> two OpenAI employees discussing the model.</p></li><li><p><a href="https://developers.googleblog.com/en/start-building-with-gemini-25-flash/">Gemini 2.5 Flash launched with thinking budget controls</a></p><p>What is particularly interesting with this launch is how well it's positioned on the cost/quality Pareto front. See the blog post for the chart showing how well it trades off cost/quality wrt other major available models. This is of course the cheaper sibling of Gemini 2.5 Pro, if you haven't seen that model, check that one out first to contextualize this launch.</p></li><li><p><a href="https://openai.com/index/browsecomp/">OpenAI introduces browser use benchmark BrowseComp</a></p><p>Interesting benchmark as browser use focused agentic use cases like Manus, Deep Research and Operator become more dominant.</p></li><li><p><a href="https://x.ai/api">Grok 3 is now available on their API</a></p><p>There's also a grok-3-mini which is significantly cheaper but seems to hold up performance in coding fairly well (53.3% vs 49.3% on Aider Polyglot). Although it does struggle with the Aider diff format which means it's a bit more verbose.</p></li><li><p><a href="https://huggingface.co/collections/THUDM/glm-4-0414-67f3cbcb34dd9d252707cb2e">Tsinghua University released GLM-4-0414 LLMs under more permissive licenses than they've used previously</a></p><p>They introduce a reasoning model and a "deep reasoning" model they call a "rumination" model. This one integrates search tools in the reasoning process and it was RL trained to be scored on its ability to do this specifically. Very similar to the o3 and o4-mini models that are known to be able to use built-in OpenAI tools like web search during the reasoning process.</p></li><li><p><a href="https://blog.lmarena.ai/blog/2025/search-arena/">LM Arena introduces the Search Arena: Evaluating Search-Enabled AI</a></p><p>We've all been getting more used to consumer-focused offerings like ChatGPT/Claude/Grok/Gemini increasingly using search indexes to ground their answers through RAG. But how well do models perform at integrating this knowledge? LM Arena introduces an eval showing performance across user submitted queries by, LM Arena style, pitting models against each other and letting the user vote on the results.</p></li><li><p><a href="https://allenai.org/blog/datadecide">DataDecide: How to predict best pretraining data with small experiments</a></p><p>Another great release by AllenAI further democratizing pre-training. DataDecide helps researchers decide on pre-training dataset selection without having to pre-train the entire model first.</p></li><li><p><a href="https://huggingface.co/microsoft/MAI-DS-R1">Microsoft AI post-trains DeepSeek R1 to align with Western values</a></p><p>They state they've compared it to Perplexity's recent similar R1 finetune R1-1776 and show it performs better in several areas. Given this recent US government <a href="https://selectcommitteeontheccp.house.gov/sites/evo-subsites/selectcommitteeontheccp.house.gov/files/evo-media-document/DeepSeek%20Final.pdf">published report</a> criticizing DeepSeek this comes at an interesting time. </p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://blog.vllm.ai/2025/04/11/transformers-backend.html">Transformers backend integration in vLLM</a></p><p>Hugging Face and vLLM have always had a more or less "loose" coupling where vLLM could use parts of it, like pulling weights from the Hugging Face Hub automatically. Now Transformers backend has been more fully integrated into vLLM making it simpler to flexibly define models using Transformers and host those models directly with vLLM.<br><br>The key unlock is to be able to go from more exotic ideas on the model side to running it for inference. Likely, performance won't be incredible but more likely than not acceptable to run some inference deployments for testing with real-world users. An example provided in the blog post is the Helium model from the Kyutai team, which wasn't supported by vLLM directly, and otherwise could not have run in vLLM without "porting it" to the vLLM list of supported models.</p></li><li><p><a href="https://github.com/deepspeedai/DeepSpeed/blob/master/blogs/deepcompile/README.md">DeepCompile: Unlocking Compiler Optimization for Distributed Training</a></p><p>DeepCompile from the DeepSpeed projects introduces a compilation step that combined with the existing ZeRO-3 configuration can boost training throughput by 30%-50% (compared to only ZeRO-3). Continued work of impressive open source contributions to the LLM training stack.</p></li><li><p><a href="https://github.com/openai/codex">OpenAI announces open source Claude Code alternative codex: a command line coding agent</a></p><p>Interesting move for them to release this as true open source (Apache 2.0) versus Anthropic's closed Claude Code project that is heavily obfuscated and does not take outside contributions (which codex does do, see the active GitHub PR/issue section). Zig when they zag?<br><br> In my personal testing I've noticed it is slightly awkward in calling tools as everything seems to go through Shell commands and that led to some unnecessary git-patch-apply formatting errors causing many more output tokens to be used and it taking much longer to apply edits. However, I expect this thing to quickly get better.<br><br>A neat description of what codex is and can do today can be found in this blog <a href="https://simonwillison.net/2025/Apr/16/openai-codex/">post by Simon Willison</a>.<br><br>Already the open nature of the project has spawned folks to write proxies to easily allow other model providers to be used with codex, see <a href="https://openresponses.masaic.ai/use-cases/codex-demo">this</a> project.</p></li><li><p><a href="https://github.com/SWE-agent/SWE-ReX">SWE-agent Remote Execution Framework</a></p><p>Interesting approach to speeding up agents that depend on code execution through massive parallelization.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2504.06214">From 128K to 4M: Efficient Training of Ultra-Long Context Large Language Models</a></p><p>Interesting long context work done by NVIDIA on long-context training.</p></li><li><p><a href="https://agent-reward-bench.github.io/">AgentRewardBench: Evaluating Automatic Evaluations of Web Agent Trajectories</a></p><p>Notoriously difficult to evaluating web browsing agents but researchers from McGill (with more contributors) have put together an interesting approach with AgentRewardBench, also check out their <a href="https://huggingface.co/spaces/McGill-NLP/agent-reward-bench-demo">Hugging Face space</a> explaining their approach interactively.</p></li><li><p><a href="https://arxiv.org/abs/2504.11536">ReTool: Reinforcement Learning for Strategic Tool Use in LLMs</a></p><p>I've mentioned this idea a couple of times already and I'm really excited to see more evidence of this strategy working: a post-training stage where tool use is enabled during the reinforcement learning phase. This allows models to learn in a much more realistic setting how they can best make use of tools during inference. Two quotes from the paper:<br><br>&gt;Remarkably, ReTool-32B attains 72.5% accuracy in extended settings, surpassing OpenAI's o1-preview by 27.9%.<br><br>&gt;Further analysis reveals emergent behaviors such as code self-correction, signaling an "aha moment" in which the model autonomously masters adaptive tool use.</p></li><li><p><a href="https://arxiv.org/abs/2504.05185">Concise Reasoning via Reinforcement Learning</a></p><p>One drawback of reasoning models is that they are verbose in their reasoning CoT requiring many output tokens for a given task. They find in this paper that it can be shown mathematically that there is a length bias in some of the RL algorithms used for RL training. However, they mention in the paper they didn't analyze this effect for the GRPO algorithm used by DeepSeek R1 and other recent reasoning model releases.<br><br>Claude 3.7 Sonnet shows reasoning and it's pretty clear that it's a lot more succinct than some of the open source reasoning models like QwQ. So I suspect a lot of effort will go into "conditioning" the RL CoT to be more dense without affecting the quality boost the RL CoT provides.</p></li><li><p><a href="https://dllm-reasoning.github.io/">d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning</a></p><p>A collaboration between Meta and UCLA shows that diffusion models can be post-trained to perform reasoning like autoregressive models can and benefit from the reasoning process to improve quality on reasoning-focused benchmarks like Sudoku, GSM8K, MATH500.</p></li><li><p><a href="https://lllyasviel.github.io/frame_pack_gitpage/">6GB Video Dreams: How FramePack's Context Scheduling Makes Long-Form Generation Possible on Consumer Hardware</a></p><p>FramePack lets you generate super long videos on modest GPUs by cleverly compressing less important frames to save memory. Awesome work from two researchers at Stanford.</p></li><li><p><a href="https://arxiv.org/abs/2504.07128">DeepSeek-R1 Thoughtology: Let's &lt;think&gt; about LLM Reasoning</a></p><p>Interesting analysis by researchers from the Mila Quebec AI Institute on R1. They find a shortcoming of R1 is that it ruminates on previous answers too much, limiting exploration of new ideas (the crucial unlock of long CoT RL).</p></li><li><p><a href="https://arxiv.org/abs/2504.04736">Synthetic Data Generation &amp; Multi-Step RL for Reasoning &amp; Tool Use</a></p><p>A more "guided" approach to teaching models how to perform Multi-Step Tool Use is to generate synthetic data showing trajectories that the model should be able to perform. This paper from the Stanford AI research group shows their approach and validates it on GSM8K, HotPotQA, CofCA, MuSiQue, and BeerQA.</p></li><li><p><a href="https://arxiv.org/abs/2501.12486">The Journey Matters: Average Parameter Count over Pre-training Unifies Sparse and Dense Scaling Laws</a></p><p>To make LLMs more efficient various kinds of pruning have been explored, mostly focused on the post-training phase. This paper explores pruning during pre-training unifying it into a single phase, increasing efficiency of LLMs as a result. They try to capture scaling behavior for various levels of pruning to help researchers understand optimal configurations for this new pre-training-infused pre-training approach.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://simonwillison.net/2025/Apr/11/camel/">CaMeL offers a promising new direction for mitigating prompt injection attacks</a></p><p>Simon Willison has been warning about prompt injection risks for a very long time and he analyzes a new approach called CaMeL. He's optimistic that their approach can help avoid the risks of agents being prompt injected by attacker text somehow making it into the agent's prompt. This work is similar to <a href="https://x.com/headinthebox/status/1910063211405255158">this work</a> from CS researcher Erik Meijer where he introduces the notion of verifying formally AI workflows.</p></li><li><p><a href="https://www.goodfire.ai/blog/under-the-hood-of-a-reasoning-model">Under the Hood of a Reasoning Model: Goodfire analyzes DeepSeek's R1 using sparse autoencoders (SAEs)</a></p><p>They show features in the R1 reasoning model like the ability to perform backtracking, self referencing, entity tracking and features that get triggered right before the result of a calculation is generated. They've also just raised $50M led by Menlo Ventures, so likely their technology is proving to be a valuable set of tools in understanding LLMs with the purpose of making them perform better (and probably better aligned).</p></li><li><p><a href="https://www.stagehand.dev/evals">Stagehand launches a Model Evaluations page</a></p><p>Stagehand is an open source project by the Browserbase company for building browser using agents. This eval gives a good sense of performance across various models. They highlight winners on speed, cost and accuracy. Those are llama3-70b-8192, gpt-4.1-nano, and gemini-2.0-flash, respectively.</p></li><li><p><a href="https://ysymyth.github.io/The-Second-Half/">The Second Half of AI: now that RL works we can shift focus to model utility</a></p><p>Interesting essay by an OpenAI researcher about now that pre-training has created the right priors for RL to work that we should collectively focus more on utility through improved evaluation environments (that can be directly trained on through RL).</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Gemini 2.5: Google on top after 5 years of playing catch-up]]></title><description><![CDATA[Week 13 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/gemini-25-google-on-top-after-5-years-14f</link><guid isPermaLink="false">https://codingwithintelligence.com/p/gemini-25-google-on-top-after-5-years-14f</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Fri, 28 Mar 2025 11:43:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!cvX-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Woops! Something went wrong with the links in the previous email. Should be fixed now!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cvX-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cvX-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!cvX-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!cvX-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!cvX-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cvX-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:239341,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/160059168?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cvX-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!cvX-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!cvX-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!cvX-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24758b6d-5431-4501-813b-6f2e4aa0a490_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/">Gemini 2.5: Our most intelligent AI model</a></p><p>It's a reasoning model. It leads models like Claude Sonnet 3.7 thinking on benchmarks like https://livebench.ai/. Vibe checks by r/LocalLLaMA are coming in very positive as people report Gemini 2.5 Pro fixing their Claude Sonnet 3.7 generated code.<br><br>The very large context window support Gemini is known for (1M) can further enhance this capability increase, as for practical in-IDE code generation being able to easily include large amounts of context significantly enhances real-world usability. I expect a massive amount of demand for this model. A lot of pressure already on Google to increase rate limits.<br><br>Gemini 2.5 also scores really well on the <a href="https://scale.com/leaderboard/tool_use_enterprise">proprietary tool use benchmark from Scale AI</a> which is an important measure for building agents with these models.  Cursor / Windsurf both shipped support for Gemini 2.5.</p></li><li><p><a href="https://openai.com/index/introducing-4o-image-generation/">OpenAI 4o Image Generation</a></p><p>The long awaited image generation capabilities of GPT-4o finally dropped and the world took notice in the form of Ghibli-style images. Marketing as a field in particular and many many niche-focused image generation startups are going to find themselves unable to compete. See for example image generations <a href="https://x.com/RonnieAEvans/status/1904636171323396163">like this</a> that incorporate provided product images extremely well.<br><br>This release is noteworthy for another reason, which is that it seems more and more true that ChatGPT invests deeply in their consumer ChatGPT app and is winning the fight with Anthropic on that front. Anthropic however currently has the best model for code assistance in IDEs like Cursor/Windsurf/Claude Code (CLI). I suspect this image generation release is going to exacerbate ChatGPT's lead in being the go-to consumer app for anything AI across the worl.</p></li><li><p><a href="https://x.com/openaidevs/status/1902485690958450871?s=46">o1-pro now available via the API at extreme prices</a></p><p>o1-pro, available to ChatGPT Pro users only prior to this release, is now available on the OpenAI API for $600 (!) per 1M output tokens and $150 per 1M input tokens.<br><br>The pricing was not well received and if you've used the ChatGPT app you know that o1-pro can take a lot of time to respond. I guess we now know approximately the upper bound of what the market is interested in consuming in terms of cost/time per unit of intelligence.<br><br>What of course matters is that o1-pro has a fairly limited incremental intelligence advantage compared to Gemini 2.5 Pro, Claude 3.7 Thinking or even models like R1. Hence the appetite for this endpoint seems low.</p></li><li><p><a href="https://arcprize.org/blog/announcing-arc-agi-2-and-arc-prize-2025">ARC AGI 2: a new benchmark for AI reasoning capabilities</a></p><p>Their benchmark focuses on measuring the ability of language models to adapt to novel, never-before-seen tasks. In addition, they emphasize tasks where regular humans are relatively easily capable of solving them. <br><br>Novel about ARC-AGI 2 is that they've conducted their own study with 400 people participating to verify that their new benchmark meets this criteria.  ARC-AGI 1 has been "solved" by frontier models through scaling up test time inference. For more details about the performance of frontier models on ARC-AGI 1 see the blog post.<br><br>This benchmark is important since it can accelerate the timeline to stronger more useful LLMs that make fewer mistakes, e.g. hallucinate less. Benchmarks are impactful because popular evals inherently lead to hill climbing by various labs/groups that are in competition to top the leaderboard. If the benchmark is strong, e.g. not susceptible to shortcuts/cheating, then the models that end up performing will end up having high utility for real-world use cases.</p></li><li><p><a href="https://www.openai.fm/">OpenAI launches new TTS models</a></p><p>The voices don't sound like SOTA voices like from e.g. ElevenLabs/Cartesia. I guess they were too focused on other areas to ship a new SOTA result here. I will add that pronunciation is very accurate. The more noteworthy update is content based Voice Activate Detection (VAD) which means that it won't speak over you as frequently (available in the OpenAI Realtime API).</p></li><li><p><a href="https://simonwillison.net/2025/Mar/24/qwen25-vl-32b/">Qwen 2.5 VL 32B released</a></p><p>A nice contribution from Qwen in the category of open source vision language models. See Simon Willison's blog post for a quick preview of its abilities. It's permissively licensed with Apache 2.0.</p></li><li><p><a href="https://openai.github.io/openai-agents-python/mcp/">OpenAI announces MCP support in OpenAI Agents SDK</a></p><p>This was accompanied by a Tweet from Sam Altman where he announces that MCP will be supported in the ChatGPT desktop app and on the Responses API. This will definitely lead to an explosion in MCP-related activity and to it becoming the dominant "agent standard" for distributed tool calling and all the other MCP server capabilities (like MCP servers being clients to other MCP servers, resources).</p></li><li><p><a href="https://github.com/modelcontextprotocol/specification/blob/9236eb1cbfa02c17ab45c83a7bdbe55c450070be/docs/specification/2025-03-26/_index.md">New Model Context Protocol (MCP) spec released</a></p><p>In summary the new spec introduces: OAuth 2.1, Streamable HTTP, JSON-RPC batching, and tool annotations. See <a href="https://github.com/modelcontextprotocol/specification/blob/9236eb1cbfa02c17ab45c83a7bdbe55c450070be/schema/2025-03-26/schema.json#L2050">this link for tool annotations</a>, it offers additional ability to clarify properties of tools when defining them. For example, whether they're idempotent (calling more than once doesn't have an effect, like approving a transaction), or that an operation is read-only.</p></li><li><p><a href="https://transformer-circuits.pub/2025/attribution-graphs/biology.html">Anthropic publishes a new mechanistic interpretability blog post: On the Biology of a Large Language Model</a></p><p>Basically, this article by Anthropic shows how 'circuit tracing' lets us peek inside LLMs like Claude Haiku to see how they really do things like reason, plan poems, or handle multiple languages, and can even spot the hidden mechanisms behind refusals, hallucinations, or faked chain-of-thought.<br><br>Great work for alignment &amp; improvement of LLMs as they continue to become more entrenched in everyone's day-to-day workflow/lives.</p></li><li><p><a href="https://console.groq.com/docs/text-to-speech">Groq launches PlayHT TTS model endpoints</a></p><p>Cool release by Groq, we can now do fast TTS &#128588; I personally like the Gail and Celeste voices.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/f/mcptools">mcptools: A command-line interface for interacting with MCP (Model Context Protocol) servers</a></p><p>Neat project to interact with MCP servers.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://liuff19.github.io/Video-T1/">Video-T1: Test-Time Scaling for Video Generation</a></p><p>Interesting to see increased test-time-compute generalizing to domains like video generation. Impressive generated videos as a result.</p></li><li><p><a href="https://cohere.com/research/papers/command-a-an-enterprise-ready-family-of-large-language-models-2025-03-27">Cohere details Command A model including training process</a></p><p>Interesting deep-dive from a lab that creates enterprise focused proprietary models (and some open weight models). Noteworthy is their use of model merging. I get particularly excited about their up-to-date BFCL focused (that's a leaderboard for function calling) evaluation for agentic use cases. Remember, agents are just function calls in a while loop ;-). </p></li><li><p><a href="https://arxiv.org/abs/2503.14499">Measuring AI Ability to Complete Long Tasks</a></p><p>Most predictions about how strong AI models will be in the future are very hand wavy "my gut is telling me this"-style answers. This paper from the Model Evaluation &amp; Threat Research (METR) institute (co-founded by ex-DeepMind, ex-OpenAI researcher) is more rigorous and their estimates indicate that software development tasks that currently take humans months might be able to be fully automated by AI systems in 5 years from now. And in 3 years from now fully automated software tasks that now take a human an entire day.<br><br>Which, looking at the rate of progress in code generation models (Claude Sonnet 3.5 &gt; 3.7 &gt; Gemini 2.5) does not feel entirely unrealistic.</p></li><li><p><a href="https://arxiv.org/abs/2503.18866">Reasoning to Learn from Latent Thoughts</a></p><p>Interesting paper from researchers at Stanford, University of Toronto and Vector Institute. They observe that directly pre-training on outputs (e.g. published content on the web) ignores the reasoning process (latent thoughts) that happens before someone writes their final answer down. They find that by inferring latent thoughts (see paper for details on how they do that) they can increase the learning efficiency of models.</p></li></ul><p>&#128736;&#65039; <strong>Products</strong></p><ul><li><p><a href="https://transluce.org/introducing-docent">Docent: LLM powered agent observability tool</a></p><p>With agents generating many many traces that are too time consuming to manually evaluate we need a new solution. Docent suggests an approach powered by LLMs, I like their direction a lot. I recommend everyone building agents and agent evals to take a closer look. I'm not affiliated with the company whatsoever.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://www.youtube.com/playlist?list=PLoROMvodv4rOaMFbaqxPDoLWjDaRAdP9D">Stanford CS224N NLP course by Professor Christopher Manning</a></p><p>This is a high quality course by the legendary Christopher Manning. His work on GloVe word vectors are attributed to be an early formulation of the attention mechanism used by transformers. Dated April 2024 so should be fairly up-to-date as these courses tend to focus on fundamentals that move slower than the model-du-jour.</p></li><li><p><a href="https://modelcontextprotocol.io/examples">MCP registry list</a></p><p>More &amp; more MCP registries popping up, I've been "catching them all" so here's my current list for you:<br><br><a href="https://smithery.ai/">https://smithery.ai/</a> <br><a href="https://www.mcp.run/">https://www.mcp.run/</a> <br><a href="https://glama.ai/mcp/servers">https://glama.ai/mcp/servers</a> <br><a href="https://www.pulsemcp.com/">https://www.pulsemcp.com/</a> <br><a href="https://opentools.com/registry">https://opentools.com/registry</a> <br><a href="https://github.com/modelcontextprotocol/servers">https://github.com/modelcontextprotocol/servers</a> <br><a href="https://mcp.composio.dev/">https://mcp.composio.dev/</a> <br><a href="https://github.com/michaellatman/mcp-get">https://github.com/michaellatman/mcp-get</a> <br><a href="https://mcpserver.cloud/">https://mcpserver.cloud/</a> <br><a href="https://github.com/cline/mcp-marketplace">https://github.com/cline/mcp-marketplace</a> <br><a href="http://mcpservers.com/">http://mcpservers.com/</a> <br><a href="https://www.mcpt.com/">https://www.mcpt.com/</a><br><br>Note that an official registry is slated for launch.</p></li><li><p><a href="https://blog.cloudflare.com/remote-model-context-protocol-servers-mcp/">Cloudflare ships remote MCP Server support</a></p><p>Neat SDK for deploying remote MCP servers, their playground even has the ability to plug in your <code>/sse</code> MCP URL to try it out. Noteworthy is their strong focus on authentication concerns which of course is more important when an MCP server is remote.</p></li><li><p><a href="https://arcturus-labs.com/blog/2024/11/21/roaming-rag--rag-without-the-vector-database/">Roaming RAG &#8211; RAG without the Vector Database</a></p><p>Interesting trend, with agents become more capable due to increased function calling capabilities of newer models, it becomes more common for agents to fetch needed context dynamically as the agent progresses. 

This is very similar to Model Context Protocol which seems to have launched with more focus on dynamically fetching context through tools rather than action taking through tools. 

Although the latter now definitely seems to have an equal amount of focus for the MCP protocol (e.g. with code generation).</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[RL is so hot right now!]]></title><description><![CDATA[Week 8 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/rl-is-so-hot-right-now</link><guid isPermaLink="false">https://codingwithintelligence.com/p/rl-is-so-hot-right-now</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 23 Feb 2025 19:48:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!nyhe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Dear readers! I&#8217;m back from a short break. Things got very busy at Groq &#9889;&#65039; and in my personal life (I bought a house in Amsterdam, yay!). We&#8217;re very back though and this week&#8217;s edition is packed with goodies, not least, the release of the <strong>DeepSeek-R1 70B Distill</strong> model that I personally worked on at Groq running at crazy average speeds of 1600 tokens per second. Reasoning models have a reputation of being slow but with Groq, no more :-).</p><p>It was a great project to work on and I hope you will all build cool things with it. It&#8217;s available to paid Groq users as <code>deepseek-r1-distill-llama-70b-specdec</code> (a credit card and a few dollars are all you need).</p><p>Enjoy this week&#8217;s CoWI: Coding with Intelligence &#129504;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nyhe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nyhe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png 424w, https://substackcdn.com/image/fetch/$s_!nyhe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png 848w, https://substackcdn.com/image/fetch/$s_!nyhe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png 1272w, https://substackcdn.com/image/fetch/$s_!nyhe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nyhe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png" width="1456" height="960" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:960,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:579359,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://codingwithintelligence.com/i/157760764?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nyhe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png 424w, https://substackcdn.com/image/fetch/$s_!nyhe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png 848w, https://substackcdn.com/image/fetch/$s_!nyhe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png 1272w, https://substackcdn.com/image/fetch/$s_!nyhe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe9300973-9a0f-481a-bf3b-7ac41e4f906a_2998x1976.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://console.groq.com/docs/models">Groq launches support for DeepSeek-R1 70B Distill running at over 1600 tokens per second</a></p><p>DeepSeek-R1 70B Distill is the DeepSeek-R1 distillation on top of Llama 3.3 70B. Through clever use of speculative decoding and a custom draft model we are able to serve this strong reasoning model at over 1600 tokens per second of output generation speed. Truly remarkable to observe. It performs significantly better than Llama 3.3 on logic riddles, complex RAG reasoning, code generation and math problems. Full benchmarks in this official <a href="https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-70B">HF repo</a>. In addition we launch DeepSeek-R1 Qwen 32B Distill, Qwen 32B and Qwen 32B Coder, three very capable models based on the Qwen 2.5 series. Especially Qwen 2.5 32B is a very capable general purpose model for tasks like RAG.</p></li><li><p><a href="https://api-docs.deepseek.com/news/news250120">DeepSeek-R1: #1 open source reasoning model</a></p><p>This model beats OpenAI's o1-1217 full o1 model on various tasks like AIME 2024 (math test), SWE-bench Verified (a software engineering task that requires producing GitHub PRs that implement scoped tasks for open source software like Django), and MATH-500. The techniques used in R1 (reinforcement learning on a strong base model) benefit mainly performance on coding, math and logic problems. An incredible release and gift to the community.<br><br>Noteworthy is the addition of several distilled models such as Qwen (1.5B up to 32B) and Llama 3.1 8B and 3.3 70B. Those models seem to benefit greatly from being finetuned (no reinforcement learning) on traces generated by the full R1 model.<br><br>For the model details, the R1 model is a MoE model with 671B parameters and 37B active parameters, 128K context length and basic support for function calling. It being a reasoning model will tend to generate many more output tokens for a given user prompt. The long Chain of Thought is encapsulated in a &lt;think&gt;&lt;/think&gt; XML tag after which the final answer is provided by the model.</p></li><li><p><a href="https://x.ai/blog/grok-3">Grok-3 released by xAI: #1 in LMSYS leaderboard in all categories</a></p><p>Impressive release by xAI. They beat o3-mini high, o1, DeepSeek-R1 levels with their latest Grok 3 reasoning models and for some benchmarks beat all of those by a decent margin like in LiveCodeBench (v5) or AIME&#8217;25. It was trained on a cluster of 200k H100s, although it's not clear whether all of those GPUs were used for a single run. xAI uses a similar strategy of having both reasoning and non-reasoning models in their latest frontier line-up and have both a big and small model variant. API access is not yet available so benchmark reproductions have not yet happened broadly, although consensus is that this model seems very good.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/volcengine/verl">verl: Volcano Engine Reinforcement Learning for LLMs</a></p><p>This training stack for large reinforcement learning LLM training by Bytedance is quickly rising in popularity. It's a neat addition to the Hugging Face trainer called <a href="https://huggingface.co/docs/trl">TRL</a>.</p></li><li><p><a href="https://github.com/RulinShao/retrieval-scaling">Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore</a></p><p>A very cool large scale retrieval system that can perform indexing and querying of a very large scale text corpus. They introduce this system in the context of exploring compute optimal scaling laws (see paper). They state that "such that a smaller model augmented with a large datastore outperforms a larger LM-only model on knowledge-intensive tasks" implying it's more efficient to assume and make available large amounts of data at inference time to dynamically add knowledge instead of using purely pre-training.<br><br>Interestingly, this is aligned with earlier remarks from Wen-mei Hwu (of Programming Massively Parallel Processors book fame) that he shared at a GPU-MODE (kernel programming community) meetup where he said he expects models to "remember" less factual data and retrieve more factual data at inference time. As a path to total system optimization (traditional LLMs have a lot of memory bandwidth pressure because of their size). In a way this is "just RAG" but much more widely than just retrieving a small amount of relevant documents.</p></li><li><p><a href="https://github.com/willccbb/verifiers">Verifiers: a set of tools for reinforcement learning with LLMs in verifiable environments</a></p><p>This is an awesome project showing how to reproduce DeepSeek-R1's proposed method of performing reinforcement learning on strong base models using the GRPO algorithm. The unique value of this repo is on offering an e2e training script to run and allowing you to define reward signals for the GRPO algorithm to use. The author, Will Brown - who is an AI researcher at Morgan Stanley, describes this kind of reward signal creation as rubric engineering. Check out <a href="https://youtu.be/D7BzTxVVMuw?t=5762">his excellent talk</a> on this idea and play with the code!<br><br>If you want to get good intuition on the GRPO algorithm itself first check out the GRPO explainer video linked in this newsletter's edition.</p></li><li><p><a href="https://huggingface.co/perplexity-ai/r1-1776">Uncensored DeepSeek-R1 1776 by Perplexity</a></p><p>They say they've "post-trained it to remove Chinese Communist Party censorship". A useful contribution for everyone that wants their AI features powered by DeepSeek to be a little more Western and a little less CCP. This might have implications for how the CCP sees model drops by Chinese companies like the High-Flyer hedge fund backed DeepSeek. Hopefully there's a limited chilling effect as Chinese teams are great at innovating in this space. The team from Perplexity carefully analyzes benchmark scores in order to avoid performance degradation as a result of the uncensoring. It looks like they've been able to maintain full model performance.</p></li><li><p><a href="https://github.com/simplescaling/s1">s1: Simple test-time scaling</a></p><p>A simple baseline reproduction of o1-like scaling by performing strictly SFT (not RL tuning) on a strong instruct model (they use Qwen 2.5 32B). They show that a tiny amount of high quality curated reasoning data (1K samples) can make a big difference.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2501.17116">Optimizing Large Language Model Training Using FP4 Quantization</a></p><p>Frontier teams, in this case Microsoft, continue to push the efficiency of model training. In this paper they show how to mitigate the quantization errors that appear in naive FP4 LLM training. They show BF16 and FP8 levels of quality in experiments at the 100B token/13B parameter scale.</p></li><li><p><a href="https://www.arxiv.org/abs/2502.05171">Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach</a></p><p>So far, most test-time compute implementations scaled the inference in "token space" where they keep generating tokens in the final output stream to produce long Chain-of-Thought streams in order to improve final answer accuracy. This approach of reasoning in latent space deeper within the model architecture "can capture types of reasoning that are not easily represented in words". They show that their 3.5B model can reach quality of 50B sized models on certain problem cases.<br><br>If you've heard of the &#128260; shape rotaters &#128260; &gt; &#128483;&#65039; wordcels &#128483;&#65039; meme then you'll probably be glad to hear that shape rotating can sometimes indeed be the winning approach.</p></li><li><p><a href="https://arxiv.org/abs/2502.11089">Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention</a></p><p>DeepSeek releases an inference technique that utilizes train-time algorithms to improve long-context inference performance. They mention scaling up to 64K context window lengths in the paper, the DeepSeek-R1 model goes up to 128K context window but the official DeepSeek inference API goes up to 64K. Most likely, because they're using NSA to improve the efficiency with which they host the model. Very kind of them to share this innovation so openly to help others host DeepSeek better &#128588;</p></li><li><p><a href="https://arxiv.org/abs/2412.21139">Training Software Engineering Agents and Verifiers with SWE-Gym</a></p><p>RL gyms are hot again! They were back in 2016 (check out <a href="https://openai.com/index/openai-gym-beta/">this throwback banger</a> by OpenAI). This time because various labs (OpenAI, DeepSeek) have found that performing RL on a very strong base LLM (crucial) with verifiable rewards (gyms) actually works.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://huggingface.co/spaces/nanotron/ultrascale-playbook">The Ultra-Scale Playbook: Training LLMs on GPU Clusters</a></p><p>An incredibly detailed guide from the kind folks at Hugging Face explaining the details of large scale distributed LLM training. They go in-depth on topics like data parallelism (ZeRO-1|2|3) and advanced topics like FP8 training, Flash Attention other kernel optimizations. Great gift to the community. Now all you need is to be GPU rich &#129299;</p></li><li><p><a href="https://huggingface.co/datasets/facebook/natural_reasoning">NaturalReasoning: Reasoning in the Wild with 2.8M Challenging Questions</a></p><p>Meta open sources a large reasoning dataset that they've constructed purely by searching in pre-training data for reasoning traces and by generating relevant matching questions/prompts. Their full data pipeline is LLM based and uses their latest Llama 3.3 70B Instruct model.</p></li><li><p><a href="https://docs.agentstack.sh/introduction">AgentStack: start your agent project</a></p><p>A neat scaffolding tool to build agents, it supports the CrewAI, LangGraph, OpenAI Swarms and LlamaIndex frameworks in addition to multiple LLM, tool and observability providers (although as it's a project from AgentOps it biases towards this provider for the observability part).</p></li><li><p><a href="https://huggingface.co/blog/unified-tool-use">Tool Use, Unified by Hugging Face</a></p><p>Need write-up of tool use formatting by Matthew Carrigan from Hugging Face. The post explains how tool definitions and tool call signatures are standardized but that more work is required on parsing tool call formats. Overall, tool use is still a complex formatting issue especially for advanced features like constrained decoding and streaming. Luckily my work at Groq aims to make this problem go away magically when you use tools with Groq inference APIs.</p></li><li><p><a href="https://x.com/sgershuni/status/1893690902398611872">Model Context Protocol continues to grow: a talk about what's next by Anthropic</a></p><p>A talk should go up later and I will share it once it does. The TLDR is that MCP seems to be the bet by Anthropic on how agent protocols will take shape and more advanced MCP use cases like making MCP servers both client/server and remote MCP servers, complex authentication support are coming soon. I'd recommend checking <a href="https://modelcontextprotocol.io/quickstart">the MCP docs</a> every now and then as it's developing quickly.<br><br>I think going into why MCP could be the basis of agents warrants an entire separate blog post but the TLDR is that MCP makes it easy to delegate tool execution such that an agent (client program) doesn't need to implement using the tools (e.g. a tool call to fetch Linear tickets can now be delegated by an MCP server that the actual Linear team maintains).<br><br>It also allows for complex delegation patterns (sub-agents) where an MCP server handling a tool call itself delegates to other MCP servers and/or makes use of LLM inference in order to satisfy the tool call. This leads to tool calls themselves becoming more "high-level" instead of pure code-style function calling (e.g. fetch JSON of weather data of city X) but more like "fetch my Uber's order status" (perhaps using a delegate <a href="https://docs.stagehand.dev/integrations/mcp-server">browser agent implemented as an MCP server</a>). I like to think of MCP servers as a "GraphQL" like layer that is optimized on the frontend (what the LLM gets) for natural language, but in the backend can call regular REST APIs to perform a requested tool call.</p></li><li><p><a href="https://gr.inc/">General Reasoning releases large scale reasoning dataset</a></p><p>123k outputs from popular reasoning models like R1, R1-Zero, LIMO, DeepHermes, OpenThoughts, R1-Distil-L70B, DeepScaleR. They have comparison answers from o3-mini and gemini-flash-thinking.</p></li><li><p><a href="https://www.youtube.com/watch?v=bAWV_yrqx4w">[GRPO Explained] DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models</a></p><p>Excellent Yannic Kilcher video, as always, about GRPO. He makes the RL algorithm, which is a derivative of PPO (invented by OpenAI co-founder John Schulman) surprisingly easy to understand.</p></li><li><p><a href="https://www.youtube.com/watch?v=v0gjI__RyCY">Jeff Dean &amp; Noam Shazeer &#8211; 25 years at Google: from PageRank to AGI</a></p><p>Two OG Googlers discuss AGI and the technologies to get there. Fascinating listen.</p></li><li><p><a href="https://supergpqa.github.io/">SuperGPQA: large scale GPQA-like benchmark</a></p><p>The numbers: 26,529 questions across 13 disciplines, 72 fields, and 285 graduate-level disciplines. Some top model rankings: DeepSeek-R1 (62) &gt; o3-mini (high reasoning effort) (55) &gt;= Doubao-1.5-Pro (55) &gt; Claude-3.5-Sonnet (48).</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[DeepSeek-V3: the model everyone is talking about]]></title><description><![CDATA[Week 1 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/deepseek-v3-the-model-everyone-is</link><guid isPermaLink="false">https://codingwithintelligence.com/p/deepseek-v3-the-model-everyone-is</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Thu, 02 Jan 2025 13:54:55 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e65ad0f7-6990-4c8f-aff1-eee2466eac19_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Dear readers,<br><br>Happy New Year and welcome to 2025! This week&#8217;s edition is a collection of everything that happened in the final 2-weeks (51/52) of 2024, and BOY did it get busy during that final sprint of the year. If nothing else, I think it signals that 2025 is going to be an incredible year for AI. With democratization of frontier performance (DeepSeek-V3, QwQ, QVQ, Llama 3.3 70B, Qwen 2.5 72B), an incredible installed base of compute clusters (multiple interconnected 100k accelerator clusters, 1M clusters in the works), and new frontier heights (o3) that fully automate most run of the mill software engineering (71.7% on SWE-bench verified), the pace of progress is bound to be electric. Strap in and enjoy the ride!<br><br>- Rick Lamers</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hLQU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hLQU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!hLQU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!hLQU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!hLQU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hLQU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:351233,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hLQU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!hLQU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!hLQU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!hLQU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88f9a986-cfe7-4537-b556-645b7afceaea_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">DeepSeek-V3 Perf/Cost chart: a new position on the Pareto front</figcaption></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://api-docs.deepseek.com/news/news1226">DeepSeek-V3: GPT-4o level open source model</a></p><p>671B MoE parameters, 37B active, pre-trained on 14.8T tokens. See Technical Report for more details. The takeaway here is that this launch is _drastically_ commoditizing frontier level models. It is significantly cheaper than GPT-4o (at 9:1 input output token it is 1/9th the cost). It claims to rival Claude Sonnet 3.5 but many evals (e.g. SWE-bench verified) still show Sonnet 3.5 (v2) beating it slightly. Training cost was allegedly around $5.5M (but maybe even cheaper because that's calculated at hourly prices of renting H800s).</p></li><li><p><a href="https://qwenlm.github.io/blog/qwq-32b-preview/">QwQ: 32B reasoning model</a></p><p>o1-mini level performance. First truly open source reasoning model. Read the blog post for details. Apache 2.0 licensed.</p></li><li><p><a href="https://qwenlm.github.io/blog/qvq-72b-preview/">QVQ: 72B vision reasoning model</a></p><p>Gets close to latest o1 (o1 2024-12-17) vision performance as measured by vision evals like MathVista. Uses 'qwen' license.</p></li><li><p><a href="https://arcprize.org/blog/oai-o3-pub-breakthrough">o3 "cracks" ARC-AGI-1</a></p><p>o3 does well on ARC-AGI-1 which many predicted would take a long time. As is tradition in machine learning, once a task has been mastered the goal post is quickly moved, and we're left searching for the next task that isn't solved. The AGI definition of MSFT for OpenAI of "excess of $100B in profits" seems most robust as humans are pretty good at trying to compete away profits by attacking high margin activities.<br><br>I'm looking forward to the next eval folks are targeting. It might be ARC-AGI-2 (on which o3 apparently gets 30%) or Frontier Math by Epoch AI (on which o3 gets about 25% atm).  The more evals start to resemble utility in terms of useful for societal (economic) operating activity, the more the artifacts that beat them actually end up making a difference in practice (getting 71.7% on SWE-bench verified like o3 does means we can all automate a significant portion of software engineering work).<br><br>Folks have pointed out that o3-mini will be the more cost effective option (beating full o1 at several tasks) but since neither o3 or o3-mini are being made available not much attention has gone to it.</p></li><li><p><a href="https://appgen.groqlabs.com/">Groq Appgen: instant web app generation on Groq</a></p><p>I've personally built this application &#128526; (with contributions from my awesome colleagues Jose Menendez and Benjamin Klieger) showcasing the power of Groq speed (speculative decoding enabled achieves 2k tokens per second on 70B!) with the Llama 3.3 70B model for strong code generation capabilities. Check out this X video by my colleague Benjamin: <a href="https://x.com/benklieger/status/1870277109601771851">https://x.com/benklieger/status/1870277109601771851</a><br><br>We've also open sourced the entire implementation <a href="https://github.com/groq/groq-appgen">https://github.com/groq/groq-appgen</a></p></li><li><p><a href="https://epoch.ai/gradient-updates/frontier-language-models-have-become-much-smaller">Epoch AI: Frontier models have likely gotten much smaller</a></p><p>Nice investigative work to show that frontier models are becoming smaller. We've long past the stage of a simple "1T models won't ever go into production". Efficiency = margin in the age of scaling AI to broad based use so the investment incentive to make models more efficient is enormous.</p></li><li><p><a href="https://huggingface.co/blog/modernbert">ModernBERT: a better BERT by Answer.AI</a></p><p>Good base model for finetuning your task specific embedding model, here's a list of finetunes folks created <a href="https://huggingface.co/models?other=base_model:finetune:answerdotai%2FModernBERT-base&amp;sort=downloads">https://huggingface.co/models?other=base_model:finetune:answerdotai%2FModernBERT-base&amp;sort=downloads</a></p></li><li><p><a href="https://x.com/bilaltwovec/status/1873791617133621376?s=46">YouTube already licensing video data to model companies?</a></p><p>Haven't seen coverage of this. Neat find by <a href="https://x.com/bilaltwovec">@bilaltwovec</a>.</p></li><li><p><a href="https://ai.google.dev/gemini-api/docs/thinking-mode">Google DeepMind previews Gemini 2.0 Flash Thinking model</a></p><p>Noam Shazeer, original author of Attention Is All You Need has returned to Google and posted about this new model on X <a href="https://x.com/NoamShazeer/status/1869789881637200228">https://x.com/NoamShazeer/status/1869789881637200228</a> </p></li><li><p><a href="https://www.chinatalk.media/p/deepseek-ceo-interview-with-chinas">Rare DeepSeek CEO interview</a></p><p>Read my full thoughts on the interview in this X &#129525; <a href="https://x.com/RickLamers/status/1874778471907344825">https://x.com/RickLamers/status/1874778471907344825</a></p></li><li><p><a href="https://deepmind.google/technologies/veo/veo-2/">Google DeepMind announces Veo 2</a></p><p>Impressive video generation model by Google DeepMind. Expectations are generally that Google has a compute (TPUs) and data (YouTube) advantage and will really nail video generation. Will be interesting to see how they go from demo &#8594; paid product (APIs on gcloud? YT creator features?).</p></li><li><p><a href="https://nebius.com/blog/posts/training-and-search-for-software-engineering-agents">Nebius Cloud team contributes SWE-bench agent using strictly open source models</a></p><p>Impressive 40.6% resolved rate on SWE-bench Verified. They use Qwen-2.5-72B Instruct and Llama 3.1 70B Base.</p></li><li><p><a href="https://platform.openai.com/docs/guides/reasoning#quickstart">o1 models available through API</a></p><p>Interestingly they come with a control parameter called "reasoning_effort" that can be set to "low", "medium" and "high".</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://huggingface.co/papers/2412.15115">Qwen2.5 Technical Report</a></p><p>Mainly reveals Qwen is aggressive scaling (18T pre-training) funded by a big tech incumbent (Alibaba). Amazing gift to the community!</p></li><li><p><a href="https://huggingface.co/blog/smolagents">smolagents: code isolation/agent framework by Hugging Face</a></p><p>Wraps E2B, which you may have seen provides isolated code execution APIs.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://apollo-lmms.github.io/">Apollo: An Exploration of Video Understanding in Large Multimodal Models</a></p><p>Meta researchers explore video understanding with VLMs. Strong VLMs for video understanding could unlock another large amount of pre-training data by "transcribing" videos made in the real-world. Video recording is cheap (smartphones), visual understanding based transcription could yield important real-time/at scale data.</p></li><li><p><a href="https://mint-video.github.io/">Mind the Time: Temporally-Controlled Multi-Event Video Generation</a></p><p>Interesting new video generation controlling technique.</p></li><li><p><a href="https://chenglin-yang.github.io/1.58bit.flux.github.io/">1.58-bit FLUX</a></p><p>Bringing low-bit representations to image generation models. Very impressive efficiency gains by folks from ByteDance/POSTECH (SK research university).</p></li><li><p><a href="https://arxiv.org/abs/2412.09764">Memory Layers at Scale</a></p><p>"On downstream tasks, language models augmented with our improved memory layer outperform dense models with more than twice the computation budget, as well as mixture-of-expert models when matched for both compute and parameters." Meta continuing to advance architecture ideas, and kindly running experiments at scale to aid the GPU poor. Thank you team &amp; Zuck!</p></li></ul><p>&#128736;&#65039; <strong>Products</strong></p><ul><li><p><a href="https://blog.google/products/gemini/google-gemini-deep-research/">Gemini Deep Research</a></p><p>Agentic product feature of Gemini Advanced, tapping Google's strong long context performance of their Gemini models. Think of it as LLM+web search on steroids. Neat launch!</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://www.youtube.com/watch?v=oqOBYOx3sHE">Beyond Decoding: Meta-Generation Algorithms for Large Language Models (Remote Talk)</a></p><p>Ideas for inference time scaling algorithms.</p></li><li><p><a href="https://www.anthropic.com/research/building-effective-agents">Building effective agents by Anthropic</a></p><p>A surprisingly balanced survey of patterns in agentic LLM applications. I think this is very close to the best understanding the leading framework developers have about agentic AI. The split of Workflows as control flow being handled in code and Agents where control flow is dynamically handled by the LLM really resonates with me. Furthermore, prompt engineering your tools is an underrated optimization strategy for getting better performance.</p></li><li><p><a href="https://simonwillison.net/2024/Dec/24/qvq/">Simon Willison covers QVQ the latest Qwen visual reasoning model</a></p><p>"I&#8217;ve tried it out with a bunch of things, with mixed results&#8212;but it&#8217;s really fun seeing how it works through a problem." is the vibe eval in case you're short on time.</p></li><li><p><a href="https://benediktstroebl.github.io/reasoning-model-evals/">Reasoning Model Evals: evals to show in which domains reasoning models improve results</a></p><p>By Arvind Narayanan et al. from Princeton. Neat attempt to identify where reasoning models help. But with all the action in reasoning models/inference-time-compute it's bound to get outdated quickly.</p></li><li><p><a href="https://www.youtube.com/watch?v=loaTGpqfctI">Byte Latent Transformer: Patches Scale Better Than Tokens (Paper Explained)</a></p><p>By your favorite AI youtuber Yannic Kilcher. This is Meta's paper about getting rid of tokens.</p></li><li><p><a href="https://github.com/deepseek-ai/DeepSeek-V3/blob/main/DeepSeek_V3.pdf">DeepSeek-V3 Technical Report</a></p><p>Interesting notes about several optimizations enabling fast FP8 training, use of Multi-Token Prediction (with ablations!) and interestingly R1 (their reasoning model) distillation for improved reasoning. Additionally they introduce a MoE routing collapse prevention technique they dub "auxiliary-loss-free load-balancing".</p></li><li><p><a href="https://huggingface.co/spaces/HuggingFaceH4/blogpost-scaling-test-time-compute">Scaling Test-Time Compute with Open Models</a></p><p>Awesome exploration of scaling test-time compute with open models by Hugging Face. "Check out this plot where the tiny 1B and 3B Llama Instruct models outperform their much larger 8B and 70B siblings on the challenging MATH-500 benchmark if you give them enough &#8220;time to think&#8221; &#129327;. Very cool result by team HF!</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Gemini 2.0: is Google finally where everyone expected it to be?]]></title><description><![CDATA[Week 49 + 50 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/gemini-20-is-google-finally-where</link><guid isPermaLink="false">https://codingwithintelligence.com/p/gemini-20-is-google-finally-where</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 15 Dec 2024 18:42:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e125ac7a-3f91-4edb-a35b-e19583136928_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This week covers 41 (!) updates about what&#8217;s happening in AI: spoiler - a lot. I&#8217;m sorry I had to skip y&#8217;all for a week but it just was <em>so</em> busy at Groq (we shipped Llama 3.3 70B + specdec version - this runs it at over 2000 tok/s on certain queries, <a href="https://x.com/BenjaminKlieger/status/1865094662643753163">check it out!</a>). Enjoy this week&#8217;s update &#9996;&#65039;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bDMT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bDMT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!bDMT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!bDMT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!bDMT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bDMT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:344614,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bDMT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!bDMT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!bDMT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!bDMT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf0028b1-74d8-428d-b780-1635b57bd4ff_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://deepmind.google/discover/blog/genie-2-a-large-scale-foundation-world-model/">DeepMind releases Genie 2: a foundational world model</a></p><p>Similar to GameNGen by Google Research earlier and Oasis by Decart. It allows a user to provide inputs to have real-time control over a world simulation engine. It seems to suffer from similar issues of earlier attempts like persistence. "we believe Genie 2 is the path to solving a structural problem of training embodied agents safely while achieving the breadth and generality required to progress towards AGI". You can be sure that teams like the folks behind the Tesla Optimus humanoid robot are experimenting with these world simulation techniques.</p></li><li><p><a href="https://blog.google/technology/google-deepmind/google-gemini-ai-update-december-2024/">Gemini 2.0</a></p><p>Google really outdid themselves with this release, not only does the model score exceptionally well on coding benchmarks versus Claude Sonnet 3.5 (new) but also does it ship with advanced agentic features like realtime voice + video input modes in the Live API. Google is now close to or depending on your POV leading in the AI race and it&#8217;s what people expected from the company since the open sourcing of TensorFlow. Especially notable is that they&#8217;re still completely unbeaten on video input and long context (up to 10M and 2M in prod).</p></li><li><p><a href="https://openai.com/12-days/?day=6">OpenAI announces realtime video mode</a></p><p>Similar to Gemini Flash 2.0 realtime mode. Only Google got there first this time :) When access rolls out more broadly evals will likely show which of these does best (under which conditions). Tracking!</p></li><li><p><a href="https://openai.com/index/introducing-chatgpt-pro/">ChatGPT Pro Mode: o1 pro mode</a></p><p>o1 pro mode scores high on reasoning tasks like PhD-Level Science Questions (GPQA Diamond), Competition Code (Codeforces), Competition Math (AIME 2024). It's pretty slow with a loading bar and has been described as "integrated best-of-n sampling" giving you sort of a pass@k level performance where k is something like 5 versus regular o1. Not sure if it warrants the $200 price tag associated to ChatGPT Pro Mode. Others have pointed out that $200 is maybe worth it for the increased o1 (non pro) rate limit increase. I guess it will be TBD how much traction the high price plan for ChatGPT will get.</p></li><li><p><a href="https://huggingface.co/deepseek-ai/DeepSeek-V2.5-1210">DeepSeek releases update V2.5: 1210</a></p><p>Improved math &amp; coding ability wrt original V2.5: MATH-500 benchmark increased from 74.8% to 82.8%. LiveCodebench (08.01 - 12.01) benchmark increased from 29.2% to 34.38%.</p></li><li><p><a href="https://arxiv.org/abs/2412.08905">Microsoft releases Phi-4: 14B math buff</a></p><p>"Phi-4 outperforms much larger models, including Gemini Pro 1.5, on math competition problems". The technical report has been lauded for digging into the synthetic data generation strategies used. https://techcommunity.microsoft.com/blog/aiplatformblog/introducing-phi-4-microsoft%E2%80%99s-newest-small-language-model-specializing-in-comple/4357090 and unofficially on HF https://huggingface.co/matteogeniaccio/phi-4. </p></li><li><p><a href="https://www.snowflake.com/engineering-blog/snowflake-arctic-embed-2-multilingual/">Snowflake releases strong permissively licensed multi-lingual embedding models: Arctic Embed 2.0</a></p><p>Very useful! Around the level of text-embedding-3 from OpenAI at 1/3rd of the dimension size.</p></li><li><p><a href="https://aws.amazon.com/ai/generative-ai/nova/understanding/">Amazon releases Nova model series</a></p><p>Check out the technical report here https://www.amazon.science/publications/the-amazon-nova-family-of-models-technical-report-and-model-card. Their best model is approximately Llama 3.2 90B/Llama 3.1 70B level but Nova isn't open source. The report also highlights image generation and video generation models, showing that these models are presumably becoming table stakes technologies for large tech companies like Amazon/Google/Tencent/Alibaba/Meta etc.</p></li><li><p><a href="https://kprize.ai/">$1M prize for beating a new SWE-bench like benchmark</a></p><p>Unreasonably high requirement (90%? open source only?). But interesting initiative to track SWE agent progress regardless.</p></li><li><p><a href="https://github.com/deepseek-ai/DeepSeek-VL2">DeepSeek-VL2: MoE A4.5B vision model competitive with Qwen 2-VL 7B/Pixtral 12B</a></p><p>Meaningful performance above Pixtral 12B at a much lower active param budget. Quite impressive.</p></li><li><p><a href="https://x.ai/blog/grok-image-generation-release">Grok image generation model Aurora released</a></p><p>Grok 2 initially relied on FLUX for image generation on the X platform but now the team seems to be getting their image gen muscle on. Images are decent but not yet surpassing SOTA like Ideogram or FLUX1.1 [pro]. No API release as of yet.</p></li><li><p><a href="https://www.reddit.com/r/LocalLLaMA/comments/1h52qat/nous_distro_distributed_training_framework_update/">Nous DisTrO: distributed model training</a></p><p>Nous has been making nice progress with distributed model training, one of the neat results is the DeMO paper which was co-authored with Diederik P. Kingma the author of the original Adam optimizer. DeMO shows controlled divergence can enable large scale distributed training. The Nous model doesn't have great MMLU scores that can compete with the slew of open source models available (all trained on fast interconnect clusters of course), but it's encouraging to see work in this area to guarantee we'll be less dependent on centralized training initiatives if we need to be in the future. In essence, this line of work is a nice hedge on centralization power. Which likely helps the centralized initiatives behave better as a result ;-)</p></li><li><p><a href="https://x.com/FishAudio/status/1864370933496205728">Fish Audio 1.5 release: SOTA open source TTS</a></p><p>The gap between proprietary (ElevenLabs) and open source is shrinking. Try the playground at https://fish.audio/ it's very impressive. Check out the ranking of TTS models in the helpful HF leaderboard https://huggingface.co/spaces/TTS-AGI/TTS-Arena</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/facebookresearch/flow_matching">Flow Matching in PyTorch educational resource by Meta</a></p><p>"Flow matching is a recent framework for generative modeling that has achieved state-of-the-art performance across various domains, including image, video, audio, speech, and biological structures." for example, FLUX.1 is based on it as well as many other SOTA generative models in other multimodal domains.</p></li><li><p><a href="https://github.com/Tencent/HunyuanVideo">HunyuanVideo: strong open source video model</a></p><p>There's also support for distributed inference using xDiT framework (xDiT: an Inference Engine for Diffusion Transformers (DiTs). Very cool work brining SOTA to local inference. A friend of mine is already generating videos with it on his 4090s!</p></li><li><p><a href="https://github.com/unternet-co/web-applets">Web Applets: An open spec &amp; SDK for creating apps that agents can use</a></p><p>Really cool project by Rupert Manfredi incubated at Mozilla.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2406.16793">Adam-mini: Use Fewer Learning Rates To Gain More</a></p><p>Reduced memory footprint version of the successful Adam optimizer. Awesome work for hobbyist fine-tuning that typically try to squeeze the most out of their hardware. Support seems to have already landed in LLaMA-Factory https://github.com/hiyouga/LLaMA-Factory</p></li><li><p><a href="https://mammoth-vl.github.io/">&#129443; MAmmoTH-VL: Eliciting Multimodal Reasoning with Instruction Tuning at Scale</a></p><p>"Using only open models, we create a dataset containing 12M instruction-response pairs to cover diverse, reasoning-intensive tasks with detailed and faithful rationales." nice study that shows that post-training for reasoning in the vision domain has lots of potential. They set SOTA scores on vision reasoning tasks like MMMU-Pro (+7%).</p></li><li><p><a href="https://ai.meta.com/research/publications/byte-latent-transformer-patches-scale-better-than-tokens/">Byte Latent Transformer: Patches Scale Better Than Tokens</a></p><p>Meta frees us from the burden of tokenizers, at last. "We present the first flop controlled scaling study of byte-level models up to 8B parameters and 4T training bytes." and "Patches are segmented based on the entropy of the next byte, allocating more compute and model capacity where increased data complexity demands it."

Meta truly GOAT, this is phenomenal.</p></li><li><p><a href="https://arxiv.org/abs/2411.17685">Attamba: Attending To Multi-Token States</a></p><p>Compressing tokens to avoid "quadratic scaling of compute with sequence length", clever approach and quality hit seems minimal.</p></li><li><p><a href="https://arxiv.org/abs/2411.16489">O1 Replication Journey -- Part 2: Surpassing O1-preview through Simple Distillation, Big Progress or Bitter Lesson?</a></p><p>They explore finetuning on O1 API data (against ToS, but hey, science!) and remark on the importance of not relying (solely) on distillation for achieving SOTA on open models.</p></li><li><p><a href="https://arxiv.org/abs/2412.06769">Training Large Language Models to Reason in a Continuous Latent Space</a></p><p>"By delaying definite decisions and expanding the latent reasoning process, the model pushes its exploration closer to the search tree&#8217;s terminal states, making it easier to distinguish correct nodes from incorrect ones."</p></li><li><p><a href="https://arxiv.org/abs/2411.16679">Do Large Language Models Perform Latent Multi-Hop Reasoning without Exploiting Shortcuts?</a></p><p>Of course the answer is yes, but the paper highlights in a helpful way that it's only the case for certain types of queries. Shortcut exploitation is a notorious issue with machine learning models and undermines generalization ability of models.</p></li><li><p><a href="https://arxiv.org/abs/2412.03555">PaliGemma 2: A Family of Versatile VLMs for Transfer</a></p><p>They highlight how their models can be used for transfer-learning to domain specific tasks like "table structure recognition, molecular structure recognition, music score recognition, as well as long fine-grained captioning and radiography report generation".</p></li><li><p><a href="https://arxiv.org/abs/2411.19722">JetFormer: An Autoregressive Generative Model of Raw Images and Text</a></p><p>"... most of these models still rely on many separately trained components such as modality-specific encoders and decoders. In this work, we further streamline joint generative modeling of images and text. We propose an autoregressive decoder-only transformer - JetFormer - which is trained to directly maximize the likelihood of raw data, without relying on any separately pretrained components, and can understand and generate both text and images." I like approaches that are aligned with the Bitter Lesson https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson.pdf</p></li><li><p><a href="https://junhahyung.github.io/STGuidance/">Spatiotemporal Skip Guidance for Enhanced Video Diffusion Sampling</a></p><p>This training-free sampling guidance method meaningfully boosts the quality of the generated videos. Magical.</p></li><li><p><a href="https://arxiv.org/abs/2412.03782">The broader spectrum of in-context learning</a></p><p>This paper by folks from Google DeepMind places the phenomenon of in-context learning in a broader context of research on generalization and meta learning. Helpful if you're interested in learning more about in-context learning and how it will develop over the next years of AI research.</p></li></ul><p>&#128736;&#65039; <strong>Products</strong></p><ul><li><p><a href="https://github.com/copilot">GitHub Copilot on the web</a></p><p>GitHub now has a chat feature that supposedly integrates with GitHub data (through GitHub APIs) well. My test queries didn't fare very well, let me know in the comments if you have more success!</p></li><li><p><a href="https://lumalabs.ai/photon">Luma Labs releases image generation model Photon</a></p><p>Competitive with other SOTA. Not open source. Stunning images, but fails my "a coffee cup upside down on a table" prompt adherence test. Oh well, more work to be done I guess :)</p></li><li><p><a href="https://repoprompt.com/">Repo Prompt: codebase to prompt</a></p><p>Utility to generate prompts from your codebase. Handy for models like O1 Pro Mode that don't have APIs yet. A handy crutch for the gap between those models having APIs (at which point the Repo Prompt utilities will just be handled by Cursor/Windsurf AI-first coding IDEs).</p></li><li><p><a href="https://serper.dev/">Serper: Google SERP search for LLMs</a></p><p>Came across this while looking at the Arena Agent leaderboard (cool project by Gorilla, see link in this newsletter). https://www.agent-arena.com/leaderboard</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://forum.openai.com/public/videos/the-future-of-math-with-o1-reasoning-2024">The Future of Math with o1 Reasoning with Terence Tao and Mark Chen (SVP of Research @ OpenAI)</a></p><p>As one commenter suggests, the TLDR is "in the short run will help mathematicians develop proofs, but in the long run probably replace humans entirely."</p></li><li><p><a href="https://www.agent-arena.com/">Agent Arena: a leaderboard for agentic tasks</a></p><p>Cool project by the Gorilla team. LMSYS style leaderboard on agentic tasks like search, stock/financial data manipulation, research, automation.</p></li><li><p><a href="https://huggingface.co/blog/infini-attention">A failed experiment: Infini-Attention, and why we should keep trying?</a></p><p>Good to see these kinds of write-ups, they are often more educational then just sharing a successful attempt.</p></li><li><p><a href="https://ailuminate.mlcommons.org/benchmarks/">AI safety benchmark AILuminate by MLCommons</a></p><p>Useful resource to get a quick safety score for models.</p></li><li><p><a href="https://x.com/cyberpengk/status/1867736413481128152">NeurIPS talk by Ilya Sutskever</a></p><p>"Pretraining is dead, long live compute"</p></li><li><p><a href="https://lilianweng.github.io/posts/2024-11-28-reward-hacking/">Reward Hacking in Reinforcement Learning</a></p><p>As mentioned earlier in this newsletter, models are susceptible to taking reasoning shortcuts (as explored in the multi-hop reasoning scenario). Lilian Weng always does a phenomenal job writing up ideas in ML and this blog post is no exception. This post dives mainly in the problem of reward hacking while a promised future post will go into mitigations. Lilian has actually just recently left OpenAI so maybe she's a bit more free to discuss these ideas moving forward. Tracking!</p></li><li><p><a href="https://www.youtube.com/watch?v=OY2x0TyKzIQ">Sora review by MKBHD</a></p><p>Surprisingly informed and grounded review of Sora from the perspective of actual users (not academics). Watch this is you don't have Sora access (they stopped accepting signups).</p></li><li><p><a href="https://github.com/irthomasthomas/undecidability/issues?q=is:issue%20is:open%20label:llm-benchmarks">List of interesting LLM benchmarks by Ir. Thomas</a></p><p>For example, it tracks benchmarks where humans still hold the record vs SOTA AI. That list might shrink faster than we'd like haha!</p></li><li><p><a href="https://transluce.org/neuron-descriptions">Scaling Automatic Neuron Description</a></p><p>Cool work scaling mechanistic interpretability by describing neurons of Llama-3.1-8B-Instruct. They release a dataset with descriptions of every neuron in the model!</p></li><li><p><a href="https://www.youtube.com/watch?v=_Bw5o55SRL8">Inference Time Compute</a></p><p>Tutorial on Inference Time Compute by a 3rd year PhD student from UChicago.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Model Context Protocol: making LLMs more useful]]></title><description><![CDATA[Week 48 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/model-context-protocol-making-llms</link><guid isPermaLink="false">https://codingwithintelligence.com/p/model-context-protocol-making-llms</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 01 Dec 2024 20:18:33 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/92f33ebe-1e1e-49f1-910c-3fbf8a684b44_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pldc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pldc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!pldc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!pldc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!pldc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pldc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:133445,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pldc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!pldc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!pldc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!pldc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffdcd94cc-36b3-4d1f-ae8a-aaf425f62315_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Model Context Protocol architecture</figcaption></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://spec.modelcontextprotocol.io/">Anthropic launches Model Context Protocol</a></p><p>"open protocol that enables seamless integration between LLM applications and external data sources and tools". It uses JSON messages to allow for communication between:<br><br><strong>Hosts:</strong> LLM applications that initiate connections<br><strong>Clients:</strong> Connectors within the host application<br><strong>Servers:</strong> Services that provide context and capabilities<br><br>Using MCP folks will be able to use LLMs more directly with the applications they use every day. Because the protocol is open it requires less coordination between walled gardens of your data silos. Anthropic probably identified that the value will be much greater to let this proliferate as an open standard versus making this something proprietary that gets little adoption. Does HTTP ring a bell? I think this is a launch to follow. Start playing with the awesome-list MCP servers linked in this post!</p></li><li><p><a href="https://allenai.org/blog/olmo2">Ai2 releases OLMo 2: the awaited SOTA truly Open Source LLM</a></p><p>They announce a model competitive with the Llama 3.1 8B model with a truly open approach (they open source pretraining code, data, weights, under permissive licenses, etc.). Truly a gift to the community.</p></li><li><p><a href="https://huggingface.co/blog/smolvlm">SmolVLM - small yet mighty Vision Language Model</a></p><p>Not better than Qwen2-VL 2B but beats moondream2 and PaliGemma 3B. Impressive contribution by the HF team for small footprint VLMs. The efficiency can make this interesting for large scale data processing too like data extraction.</p></li><li><p><a href="https://x.com/jarrodwattsdev/status/1862299845710757980?s=46">AI Agent prompt hacking competition wins developer $50k</a></p><p>Neat concept: an AI agent controls an account with an initial $3k balance. You can attempt to prompt-convince it to wire you the money (over crypto) through its function calling abilities (approveTransfer). It's instructed however to not wire any more. The anonymous developer took home 30% of the prize as a fee so that was a quick way to earn $14.1k :)</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/genmoai/mochi/blob/main/demos/fine_tuner/README.md">Mochi 1 LoRA Fine-tuner</a></p><p>Neat single script video language model fine-tuning script that can run on a single H100/A100.</p></li><li><p><a href="https://github.com/punkpeye/awesome-mcp-servers">Awesome MCP Servers</a></p><p>MCP is getting adopted fast, a sprawling collection of projects making models more useful by glueing services through the Model Context Protocol. "@modelcontextprotocol/server-filesystem &#128199; &#127968; - Direct local file system access." some are pretty tricky if used maliciously! Use with caution. And of course, if you build any cool MCP servers put them on the list with a PR!</p></li><li><p><a href="https://github.com/abus-aikorea/voice-pro">voice-pro: zero-shot Voice Cloning &amp; more</a></p><p>Controversial topic with deep fake attacks running rampant, but I think it's good to democratize the technology and make everyone robust against the existence of this technology. In addition, it serves as a great learning tool and a cost reducer for folks who have a genuine use case (automate your tutorial recordings :)?).</p></li><li><p><a href="https://github.com/steel-dev/steel-browser">steel-browser: a tool for building browser/agent automations</a></p><p>Very cool concept, AI-powered Selenium/Playwright is taking flight.</p></li><li><p><a href="https://github.com/awslabs/multi-agent-orchestrator">Yet Another Agent Framework (YAAF): Multi-Agent Orchestrator  by AWS</a></p><p>It's not called YAAF I made that up, but at this point I wonder how many more agent frameworks we need :)</p></li><li><p><a href="https://github.com/srcbookdev/srcbook">Srcbook: AI-first app development platform</a></p><p>Forget about mobile first, AI-first app development platforms lean into strengths/weaknesses of AI code gen limits to facilitate building apps quickly. Neat idea! Kind of the natural expansion of Claude Artifacts?</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2411.14982">Large Multi-modal Models Can Interpret Features in Large Multi-modal Models</a></p><p>Mechanistic Interpretability work can be very tedious, the LLMs-Lab Team from NTU did what any reasonable AI-loving engineer/researcher would do: use AI. They've repurposed VLMs (like LLaVA-OV-72B) to identify features to be used for model steering. Looks a lot better than scaling manual approaches that are prominent in MI work to date.</p></li><li><p><a href="https://arxiv.org/abs/2411.17116">Star Attention: Efficient LLM Inference over Long Sequences</a></p><p>A contribution from NVIDIA to reduce inference time 10x with minimal quality loss (95-100%). An approximation allows for parallel and low-communication overhead calculation of attention. There's also a repo: https://github.com/NVIDIA/Star-Attention</p></li><li><p><a href="https://arxiv.org/abs/2410.21272">Arithmetic Without Algorithms: Language Models Solve Math With a Bag of Heuristics</a></p><p>&#8220;our experimental results across several LLMs show that LLMs perform arithmetic using neither robust algorithms nor memorization; rather, they rely on a "bag of heuristics" -&gt; really shows that circuitry formation inside LLMs leaves a lot to be desired, will be an important branch of research for frontier labs.</p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://quark-3d.github.io/">Quark: Real-time, High-resolution, and General Neural View Synthesis</a></p><p>Cool neural rendering result by Google, impressively high quality. A bit of a detour from the usual LLM/VLM focused programming, but a neat application of neural networks nonetheless.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://adamkarvonen.github.io/machine_learning/2024/06/11/sae-intuitions.html">An Intuitive Explanation of Sparse Autoencoders for LLM Interpretability</a></p><p>Good starter post if you want to learn about techniques used for creating feature steering techniques like &#8220;Golden Gate Claude&#8221;.</p></li><li><p><a href="https://epoch.ai/data/ai-benchmarking-dashboard">AI Benchmarking Hub by Epoch AI</a></p><p>Epoch AI has been dropping resources for keeping up with progress in the field of AI and this benchmark database is a welcome addition to see SOTA results at a glance. Ty guys!</p></li><li><p><a href="https://github.com/modelcontextprotocol/specification/blob/b9d368f6a28e1720be00bf5dc7c72ace2ed6948e/schema/schema.ts#L899">MCP cost/speed/intelligence trifecta</a></p><p>This was a neat find by the https://buttondown.com/ainews newsletter. MCP has a built-in way to set preferences for cost, speed and intelligence. Likely the inputs to a model router system that will handoff to the right model automatically based on the request.</p></li><li><p><a href="https://pytorch.org/blog/quantization-aware-training/">Quantization-Aware Training for Large Language Models with PyTorch</a></p><p>There is more evidence (https://arxiv.org/abs/2411.04330) building that precision during training influences post-training inference optimization techniques like quantization. Therefore, it's great to see PyTorch prioritizing making techniques like quantization-aware training more accessible. Especially for more broadly performed activities beyond frontier labs like fine-tuning.</p></li><li><p><a href="https://glama.ai/blog/2024-11-25-model-context-protocol-quickstart">Model Context Protocol (MCP) Quickstart</a></p><p>Neat write-up on Anthropic's newly launched Model Context Protocol (MCP) to get the TLDR.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Open Source o1 has (almost) arrived: DeepSeek R1-Lite-Preview]]></title><description><![CDATA[Week 47 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/open-source-o1-has-almost-arrived</link><guid isPermaLink="false">https://codingwithintelligence.com/p/open-source-o1-has-almost-arrived</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 24 Nov 2024 21:46:23 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/09fa434c-4f61-44eb-bfeb-2d28028df9b6_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2yKI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2yKI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!2yKI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!2yKI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!2yKI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2yKI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:751321,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2yKI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!2yKI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!2yKI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!2yKI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6a4dacd-e205-4960-a8e7-59fc43c07012_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://api-docs.deepseek.com/news/news1120">DeepSeek-R1-Lite-Preview: o1-preview-level performance on AIME &amp; MATH benchmarks</a></p><p>Awesome to see that moats aren't very long lived. DeepSeek promised to open source this model too, hasn't dropped yet but I'm sure the community is anxiously waiting.</p></li><li><p><a href="https://blog.mlc.ai/2024/11/22/achieving-efficient-flexible-portable-structured-generation-with-xgrammar">XGrammar: new Structured Generation library by the MLC project</a></p><p>The repo: https://github.com/mlc-ai/xgrammar</p></li><li><p><a href="https://www.marktechpost.com/2024/11/20/chinese-agi-startup-stepfun-developed-step-2-a-new-trillion-parameter-moe-architecture-model-ranking-5th-on-livebench/">step-2-16k-202411: a mysterious 1T model just appeared</a></p><p>Livebench.ai is a neat benchmarking attempt that constantly updates to avoid being saturated by Training On the Test Set. step-2-16k-202411 ranked just below o1-mini and above last week's entrant gemini-exp-1114. Although 1T seems prohibitively expensive to run. Reportedly Claude Opus 3.5 isn't being deployed for the exact reason of the economics of the scale of the model not weighing up against the performance delta versus other (smaller models) like Claude Sonnet 3.5.</p></li><li><p><a href="https://mistral.ai/news/pixtral-large/">Pixtral Large 124B</a></p><p>It performs well, from the blog: outperforming all of Claude-3.5 Sonnet (new). Unfortunately, only available under non-commercial research license. Great work though from the folks at Mistral, and kudos for an open weights release!</p></li><li><p><a href="https://qwenlm.github.io/blog/qwen2.5-turbo/">Qwen2.5-Turbo: 1M long-context Qwen</a></p><p>Impressive performance and evals (NIAH, RULER and LV-Eval). Congrats team Qwen! It's not released as a model/inference setup though, API only.</p></li><li><p><a href="https://huggingface.co/AIDC-AI/Marco-o1">Marco-o1: open source o1?</a></p><p>CoT fine-tuning, MCTS for search. First impression this doesn't look very good. At least on the small model they performed these steps on it's only marginally better than Qwen 2 7B.</p></li><li><p>Want to also make one remark about Google DeepMind folks dropping another model, the <strong>gemini-exp-1121</strong> model, which purportedly is even better than <strong>gemini-exp-1114</strong>, <a href="https://simonwillison.net/2024/Nov/22/gemini-exp-1121/">YMMV</a> but whatever the case, I&#8217;m glad labs are making improved models available more quickly to end users!</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/Lightricks/LTX-Video">LTX-Video: more open source generative video models</a></p><p>Impressive model by an impressive team. I'd give this a look if you care about generative video models/inference.</p></li><li><p><a href="https://github.com/PKU-YuanGroup/LLaVA-CoT">LLaVA-CoT: first visual language model capable of spontaneous, systematic reasoning, similar to GPT-o1</a></p><p>Reasoning for multimodal models. Like Insight-V from this week's roundup.</p></li><li><p><a href="https://github.com/microsoft/BitNet">1-bit AI Infra: Part 1.1, Fast and Lossless BitNet b1.58 Inference on CPUs</a></p><p>Microsoft drops the official implementation for very-low-bit LLMs and it's fairly efficient! Paper accompanying the software release <a href="https://arxiv.org/abs/2410.16144">https://arxiv.org/abs/2410.16144</a> and the original BitNet paper <a href="https://arxiv.org/abs/2402.17764">https://arxiv.org/abs/2402.17764</a></p></li><li><p><a href="https://allenai.org/tulu">T&#252;lu 3: open model + code for SOTA post-training</a></p><p>Specifically their Reinforcement Learning with Verifiable Rewards (RLVR) treatment is worth taking a look at. Awesome work by the folks at Ai2.</p></li><li><p><a href="https://huggingface.co/collections/apple/aimv2-6720fe1558d94c7805f7688c">AIMv2: vision encoders by Apple</a></p><p>Awesome follow-up work by Apple. Autoregressive Pre-training of Large Vision Encoders, see <a href="https://github.com/apple/ml-aim">https://github.com/apple/ml-aim</a> also.</p></li><li><p><a href="https://github.com/dongyh20/Insight-V">Insight-V: Exploring Long-Chain Visual Reasoning with Multimodal Large Language Models</a></p><p>Paper + Code + Checkpoints. With o1 being text only it's great to see the next frontier is rapidly being explored: multimodal reasoning chains. Should help a lot with embedded AI like robotics.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://balrogai.com/">BALROG: Benchmarking Agentic LLM/VLM Reasoning On Games</a></p><p>The Return of Games for benchmarking. Pre-ChatGPT/GPT-3 using games to benchmark AI was all the rage with Lee Sedol's famous Go match and OpenAI beating pro humans in Dota 2 and of course the famous Atari benchmarks. We'll see how things go when using games to evaluate LLMs. Signal or noise? The current leaderboard seems to decently rank models although in the top ranks it's not clear whether the ranking is really definitive (Llama 3.1 &lt; GPT-4o-mini?  Llama-3.1-70B-it &gt;  Llama-3.2-90B-it?).</p></li><li><p><a href="https://arxiv.org/abs/2411.07681">What Do Learning Dynamics Reveal About Generalization in LLM Reasoning?</a></p><p>"begin to copy the exact reasoning steps from the training set" a useful observation that can lead to higher quality pre-training by focusing on including general reasoning patterns in the pre-training corpus. Or even steering towards these in post-training to eliminate flawed non-general reasoning patterns.</p></li><li><p><a href="https://arxiv.org/abs/2402.02622">DenseFormer: Enhancing Information Flow in Transformers via Depth Weighted Averaging</a></p><p>"improves the perplexity of the model without increasing its size ... an additional averaging step after each transformer block ... coherent patterns of information flow, revealing the strong and structured
reuse of activations from distant layers" looks like carrying information forward is important. Will be interesting to see if this "architecture trick" is adopted by open source model providers. Of course, proprietary labs might already know of this and use it (or similar) in their models.</p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://playground.allenai.org/">Ai2 Playground: T&#252;lu3 70B</a></p><p>Neat chat playground to try out the T&#252;lu3 model</p></li><li><p><a href="https://huggingface.co/spaces/AI4Editing/MagicQuill">MagicQuil: An Intelligent Interactive Image Editing System</a></p><p>Very cool project, not open source, but they do have a demo up.</p></li><li><p><a href="https://huggingface.co/spaces/ysharma/Text_Behind_Image">Text Behind Image</a></p><p>A neat narrow-focus image model by folks from &#129303; Does what it says on the tin!</p></li></ul><p>&#128736;&#65039; <strong>Products</strong></p><ul><li><p><a href="https://x.com/morph_labs/status/1858986665957785711">Morph: Infinibranch</a></p><p>Interesting idea of branching agents through snapshotted VMs. Potentially powerful in the new search oriented inference-time compute scaling that labs are exploring. Private preview unfortunately it seems.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://www.ben-evans.com/presentations">AI eats the world by Benedict Evans</a></p><p>A thought piece by Benedict Evans ex-a16z partner on where AI is going. A nontechnical more market focused lens I think provides a helpful view on what practitioners, even those operating at the detailed technology level, can expect over the next few years. Although "It will work like every other platforms shift" &lt;&gt; "No-one knows" does more likely than not leave figuring out the actual answers as an exercise to the reader.</p></li><li><p><a href="https://blog.dottxt.co/say-what-you-mean.html">Say What You Mean: A Response to 'Let Me Speak Freely'</a></p><p>Outlines authors publish a rebuttal on https://arxiv.org/abs/2408.02442 TLDR with precise handling structured output _does_ improve performance.</p></li><li><p><a href="https://x.com/_xjdr/status/1859292093367320641?s=52">Example prompt from Yann LeCun for reasoning models</a></p><p>DeepSeek-R1-Lite-Preview cracks it!</p></li><li><p><a href="https://arxiv.org/abs/2411.12580">Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models</a></p><p>Interesting paper highlighting how reasoning is influenced by pretraining in LLMs.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Gemini Exp 1114: overfitted to benchmarks or new king?]]></title><description><![CDATA[Week 46 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/gemini-exp-1114-overfitted-to-benchmarks</link><guid isPermaLink="false">https://codingwithintelligence.com/p/gemini-exp-1114-overfitted-to-benchmarks</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 17 Nov 2024 22:59:09 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/cb883b31-0395-4ff6-8540-5a3b46ef72bd_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ApKY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ApKY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!ApKY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!ApKY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!ApKY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ApKY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png" width="613" height="437.85714285714283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:613,&quot;bytes&quot;:172773,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ApKY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!ApKY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!ApKY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!ApKY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79ad638d-2aea-49e7-99f7-5e4cf7a40845_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Exp 1114 = Gemini 2?</figcaption></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://nexusflow.ai/blogs/athene-v2">Athene-V2: Advancing Beyond the Limits of Scaling with Targeted Post-training</a></p><p>Fine-tuning for function calling/Agent use cases seems to be getting more attention. With more powerful models to tune from (Qwen 2.5 72B in this case) the feasibility increases, even when operating on a budget.</p></li><li><p><a href="https://cloud.google.com/tpu/docs/v6e-intro">Google's new TPU: Trillium (v6e)</a></p><p>Fun to explore on GCP with JAX.</p></li><li><p><a href="https://supermaven.com/blog/cursor-announcement">Supermaven joins Cursor</a></p><p>Normally I don't feature M&amp;A activity in AI but this one in particular is interesting given the momentum of Cursor for AI-assisted coding. This will strengthen their lead and lead to even more momentum. If you haven't tried Cursor, now is _really_ a good time to start getting familiar.</p></li><li><p><a href="https://lifelong-ml.cc/">4th edition of Conference on Lifelong Learning Agents (CoLLAs)</a></p><p>Scheduled for Aug 11, 2025. Defined as "systems that can continually learn throughout their lifetime".</p></li><li><p><a href="https://aistudio.google.com/prompts/new_chat">Gemini 1114 breaks (almost) all records</a></p><p>It seems to outperform Sonnet 3.5 (Oct) in certain cases, even outperforming o1-preview in some cases when prompted to using CoT. Rumored to be "Gemini 2". It didn't pass all of my vibe questions so I'm yet convinced this model is a clear #1.</p></li><li><p><a href="https://x.com/lmarena_ai/status/1857110672565494098">LMSYS Arena update: Gemini Exp 1114 takes #1 spot overall </a></p><p>More on the Exp 1114 release. It scores well in Arena but we know that isn&#8217;t the full story. Leave in the comments how well it works for you!</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://huggingface.co/collections/reach-vb/ultravox-audio-language-model-release-67373b602af0a52b2a88ae71">fixie-ai/ultravox-v0_4_1-llama-3_1-70b</a></p><p>A Llama 3.1 70B backbone for Whisper encoder to fuse speech-to-text into the language model itself, reducing the need for orchestration.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://ekinakyurek.github.io/papers/ttt.pdf">The Surprising Effectiveness of Test-Time Training for Abstract Reasoning</a></p><p>Test-Time for the ARC challenge. Interesting ideas by a team from MIT.</p></li><li><p><a href="https://arxiv.org/abs/2411.04965">BitNet a4.8: 4-bit Activations for 1-bit LLMs</a></p><p>Adaptive quantization and efficient low-precision implementations can push the needle on efficiency. Paired with the Scaling Laws for Precision paper this presents an interesting push on the frontier of efficient AI systems.</p></li><li><p><a href="https://needle-threading.github.io/">Needle Threading: Can LLMs Follow Threads Through Near-Million-Scale Haystacks?</a></p><p>TLDR "Strikingly, we find that many models are remarkably thread-safe: capable of simultaneously following multiple threads without significant loss in performance. Still, for many models, we find the effective context limit is significantly shorter than the supported context length, with accuracy decreasing as the context window grows."</p></li><li><p><a href="https://arxiv.org/abs/2411.04330">Scaling Laws for Precision</a></p><p>"For inference, we find that the degradation introduced by post-training quantization increases as models are trained on more data, eventually making additional pretraining data actively harmful. For training, our scaling laws allow us to predict the loss of a model with different parts in different precisions, and suggest that training larger models in lower precision may be compute optimal."</p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://byteaigc.github.io/X-Portrait2/">X-Portrait 2: Highly Expressive Portrait Animation</a></p><p>For those keeping track of SOTA performance transfer in video/image-to-video here's the latest from ByteDance, it does a lot better than ACT-One,  RunwayML's very recent flagship release, in the examples shown.</p></li><li><p><a href="https://huggingface.co/spaces/briaai/BRIA-RMBG-2.0">RMBG-2.0 for background removal by BAAI</a></p><p>Neat open source model for a practical task like BG removal.</p></li></ul><p>&#128736;&#65039; <strong>Products</strong></p><ul><li><p><a href="https://bolt.new/">Bolt.new create any Web App using agent prompting</a></p><p>Claude Artifacts on steroids! Instant deploy too. Neat product. No affiliation.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://www.youtube.com/watch?v=a42key59cZQ">Dwarkesh Patel interviews NLP researcher Gwern</a></p></li><li><p><a href="https://www.letta.com/blog/letta-v0-5-2-release">Letta introduces tool use with constraints</a></p><p>"TerminalToolRule(tool_name=...) - If the tool is called, the agent ends execution, InitToolRule(tool_name=...) - The tool must be called first when an agent is run. ToolRule(tool_name=..., children=[...]) - If the tool is called, it must be followed by one of the tools specified in children" <br><br>Interesting ideas for steer-ability!</p></li><li><p><a href="https://llmstxt.org/">llms.txt spec by Answer.AI: plaintext docs for AI</a></p><p>Anthropic already implemented it <a href="https://docs.anthropic.com/llms-full.txt">https://docs.anthropic.com/llms-full.txt</a> / <a href="https://docs.anthropic.com/llms.txt">https://docs.anthropic.com/llms.txt</a> I wonder how folks will go about prompt injection issues with this. I guess it comes down to trusting authorities/domain names.</p></li><li><p><a href="https://www.reddit.com/r/LocalLLaMA/comments/1gsyp7q/humaneval_benchmark_of_exl2_quants_of_popular/">Effect of quantization on various LLMs including Qwen2.5-Coder-32B-Instruct</a></p><p>Spoiler: Qwen2.5-Coder-32B-Instruct degrades surprisingly little when quantized to lower bit representation. Interesting in light of the 'Scaling Laws for Precision' paper in this week's Papers section.</p></li><li><p><a href="https://epoch.ai/blog/can-ai-scaling-continue-through-2030">Can AI Scaling Continue Through 2030?</a></p><p>An investigative report speculating on the bottlenecks to continued scaling of the key components of modern AI. "We identify electric power, chip manufacturing, data and latency as constraints."</p></li><li><p><a href="https://stripe.dev/blog/adding-payments-to-your-agentic-workflows">Stripe creates APIs specifically for agent financial actions</a></p><p>Virtual Credit Cards for your agents? Awesome AI forward features shipped by Stripe.</p></li><li><p><a href="https://www.youtube.com/watch?v=6PEJ96k1kiw">Speculations on Test-Time Scaling (o1) by Sasha Rush</a></p><p>Sasha is a professor at Cornell Tech and works at Hugging Face.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Chinese labs drop some incredible models]]></title><description><![CDATA[Week 45 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/chinese-labs-drop-some-incredible</link><guid isPermaLink="false">https://codingwithintelligence.com/p/chinese-labs-drop-some-incredible</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 10 Nov 2024 22:25:30 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/986df829-2c0c-4aef-ad5d-f7c9fea6f86d_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yzc5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yzc5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!yzc5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!yzc5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!yzc5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yzc5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:224125,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yzc5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!yzc5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!yzc5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!yzc5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ef0c01b-a64d-4fd3-97c7-7498472ef2ab_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The 389B A52B Instruct model performance</figcaption></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://opencoder-llm.github.io/">OpenCoder 8B: The Open Cookbook for Top-Tier Code Large Language Models</a></p><p>Reproducible LLMs from a Chinese lab, very awesome contribution and actually impressive performance, outperforming the very strong Qwen 2.5 Coder 7B.</p></li><li><p><a href="https://huggingface.co/tencent/Tencent-Hunyuan-Large">Hunyuan-Large: A52B MoE by Tencent</a></p><p>This supposedly beats Llama 3.1 405B on a number of benchmarks, very impressive. License is non EU and sub 100M users.</p></li><li><p><a href="https://huggingface.co/tencent/Hunyuan3D-1">Hunyuan3D-1 an open source Text-to-3D and Image-to-3D model by Tencent</a></p><p>Impressive object generation! Game devs/AR/VR enthusiasts will love this. Homepage <a href="https://3d.hunyuan.tencent.com/">https://3d.hunyuan.tencent.com/</a></p></li><li><p><a href="https://mistral.ai/news/mistral-moderation/">Mistral launches Moderation API</a></p><p>Fairly high accuracy across all categories. $0.1 per 1M input tokens.</p></li><li><p><a href="https://epochai.org/frontiermath/the-benchmark">FrontierMath: A math benchmark testing the limits of AI</a></p><p>Hard evals that LLMs can't crack have a tendency to accelerate the field. This new math eval by Epoch AI is a banger release that puzzles even o1-preview, Claude 3.5 Sonnet (10-22), and Gemini 1.5 Pro (002) all scoring 1-2%.</p></li><li><p><a href="https://blackforestlabs.ai/flux-1-1-ultra/">FLUX1.1 [pro] Ultra and Raw Modes</a></p><p>4MP in sub 10 seconds is very impressive of an image model with this level of quality. Go team Black Forest Labs! Raw Mode is a nice alternative to the Midjourney-vibe we've been getting from most text-to-image models.</p></li><li><p><a href="https://platform.openai.com/docs/guides/predicted-outputs">OpenAI launches predicted outputs feature</a></p><p>"Predicted Outputs enable you to speed up API responses from Chat Completions when many of the output tokens are known ahead of time. This is most common when you are regenerating a text or code file with minor modifications." Edit style prompts can be 2x faster, as a rule of thumb.</p></li><li><p><a href="https://www.anthropic.com/claude/haiku">Claude Haiku 3.5 released</a></p><p>It costs about a third of 3.5 Sonnet but is meaningfully worse. It's a bit faster than 3.5 Sonnet but only by about 16%. I don't think this is Anthropic's best launch.</p></li><li><p><a href="https://googleprojectzero.blogspot.com/2024/10/from-naptime-to-big-sleep.html">From Naptime to Big Sleep: Using Large Language Models To Catch Vulnerabilities In Real-World Code (found SQLite buffer underflow vulnerability)</a></p><p>Google's security effort under Project Zero has successfully used LLMs to find code vulnerabilities. Awesome and highly technical writeup.</p></li><li><p><a href="https://si.inc/hertz-dev/">hertz-dev: open-source base model for conversational audio generation</a></p><p>Permissively licensed, impressive samples, great write-up.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://huggingface.co/amd/AMD-OLMo">AMD-OLMo 1B</a></p><p>AllenAI's OLMo's models training code ported to AMD stack. Neat contribution to accelerate AMD &lt;&gt; NVIDIA race.</p></li><li><p><a href="https://huggingface.co/migtissera/Tess-R1-Limerick-Llama-3.1-70B">Tess-R1 Limerick (Llama-3.1-70B)</a></p><p>Another test-time-compute built-in finetune of Llama 3.1 70B that claims improved benchmark performance. Explore and leave in the comments how well you think it does. Here's a HF Space <a href="https://huggingface.co/spaces/poscye/chat-with-tess">https://huggingface.co/spaces/poscye/chat-with-tess</a></p></li><li><p><a href="https://www.microsoft.com/en-us/research/articles/magentic-one-a-generalist-multi-agent-system-for-solving-complex-tasks/">Magentic-One: A Generalist Multi-Agent System for Solving Complex Tasks</a></p><p>Interesting agent stack by Microsoft - equipping LLMs with browsing, terminal, filesystem capabilities for open ended goal accomplishment.</p></li><li><p><a href="https://github.com/stanford-aimi/ravl">RaVL: Discovering and Mitigating Spurious Correlations in Fine-Tuned Vision-Language Models</a></p><p>This work from a team at Stanford finds a way to identify and mitigate spurious correlation in Vision Language Models.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2411.00715v1">B-cosification: Transforming Deep Neural Networks to be Inherently Interpretable</a></p><p>Must read if you're interested in interpretability work for deep neural networks. This paper focuses on the image domain mainly. By folks from Max Planck, Kyutai &amp; more.</p></li><li><p><a href="https://research.nvidia.com/labs/dir/cosmos-tokenizer/">Cosmos Tokenizer: A suite of image and video neural tokenizers</a></p><p>Important work in tokenization of multimodal content, tokenizers can form a glass ceiling for performance because of information lost in the process of tokenizing content. Amazing work by NVIDIA and published with a very high level of detail!</p></li><li><p><a href="https://agoraprotocol.org/">A Scalable Communication Protocol for Networks of Large Language Models</a></p><p>This approach to allowing agents to communicate for agent-swarm like applications looks promising. If you've been put off by the verbiage usually used in "distributed agent" work then this will be a fresh, well thought out piece of research.</p></li><li><p><a href="https://arxiv.org/abs/2411.02853">ADOPT: Modified Adam Can Converge with Any &#946;2 with the Optimal Rate</a></p><p>Various folks on X have been touting this as being strictly better than Adam, meaning they would always get better performance from this compared to running Adam, and that would be a large feat. Better optimizers matter when efficiency gains translate into thousands, tens of thousands or even larger dollar/energy savings at large scale pre- and finetuning.</p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://generative-value-learning.github.io/">Vision Language Models are In-Context Value Learners</a></p><p>Try the Interactive Illustration, it impressively demonstrates how complementary robotics and vision language models are. Expect acceleration of Robotics especially in control like bimanual manipulation tasks. By DeepMind, UPenn, and Stanford.</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://github.com/simonw/llm/issues/607">Simon Willison exploring tool use feature for his `llm` CLI project</a></p><p>Cool GitHub issue!</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Full World simulation just had its ImageNet moment]]></title><description><![CDATA[Week 44 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/full-world-simulation-just-had-its</link><guid isPermaLink="false">https://codingwithintelligence.com/p/full-world-simulation-just-had-its</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 03 Nov 2024 19:57:41 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/96051e79-c7a8-40e3-a090-be098d36cbb9_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FVa4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FVa4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!FVa4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!FVa4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!FVa4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FVa4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:701836,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FVa4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!FVa4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!FVa4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!FVa4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d881521-1c52-4a90-bab2-46e313ceb40a_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I don&#8217;t believe folks are grasping the implications of this achievement yet. But the ability to simulate full world environments at increasingly higher levels of fidelity will usher in an era of robotics and real-world reasoning the consequences of which are hard to fully comprehend. Incredible work from the folks at <a href="https://www.decart.ai/">Decart</a>.</p><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://openai.com/index/introducing-simpleqa/">OpenAI introduces SimpleQA benchmark</a></p><p>In an attempt to curate unsaturated benchmarks. Note, OpenAI has still done significantly more for open source than Anthropic. Something to ponder about :) OpenAI's o1-preview gets ~42% and interestingly also refuses to answer (instead of just hallucinating).</p></li><li><p><a href="https://www.recraft.ai/blog/recraft-introduces-a-revolutionary-ai-model-that-thinks-in-design-language">Recraft v3: most powerful (closed) image generation model</a></p><p>They also launch with an API. It's number one on Artificial Analysis arena leaderboard. <a href="https://artificialanalysis.ai/text-to-image/arena?tab=Leaderboard">https://artificialanalysis.ai/text-to-image/arena?tab=Leaderboard</a></p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/michaelfeil/infinity/">Embedding inference engine</a></p><p>Very cool project by Michael Feil from Gradient AI.</p></li><li><p><a href="https://github.com/trotsky1997/MathBlackBox">Progress on o1 repro: MCTSr: Mathematic as a Blackbox for LLM</a></p><p>Early stage of a project attempting to reproduce o1, good source of raw ideas if you're working on this yourself.</p></li><li><p><a href="https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B-Instruct">SmolLM2: powerful 1.7B SLM (small language model)</a></p><p>Great model by Loubna Ben Allal from Hugging Face. Beats Qwen2.5-1.5B in multiple categories.</p></li><li><p><a href="https://github.com/browserbase/stagehand">Stagehand: an AI web browsing framework by Browserbase</a></p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2410.20771">MrT5: Dynamic Token Merging for Efficient Byte-level Language Models</a></p><p>I've been very interested in a tokenization-free approach to LLMs and this paper from Stanford nails it. Check this out if you think tokenizers are bottlenecking LLMs too!</p></li><li><p><a href="https://arxiv.org/abs/2410.16531">Bayesian scaling laws for in-context learning</a></p><p>Interesting approach to modeling scaling laws for In-context Learning ability of LLMs.</p></li><li><p><a href="https://arxiv.org/abs/2410.22179">Very Attentive Tacotron: Robust and Unbounded Length Generalization in Autoregressive Transformer-Based Text-to-Speech</a></p><p>"When applied to text-to-speech (TTS), these models (AR Transformers) tend to drop or repeat words or produce erratic output, especially for longer utterances. In this paper, we introduce enhancements aimed at AR Transformer-based encoder-decoder TTS systems that address these robustness and length generalization issues."</p></li><li><p><a href="https://ai.meta.com/research/publications/transfusion-predict-the-next-token-and-diffuse-images-with-one-multi-modal-model/">Transfusion: Predict the Next Token and Diffuse Images with One Multi-Modal Model</a></p><p>By Meta AI.</p></li><li><p><a href="https://arxiv.org/abs/2410.22456">Image2Struct: Benchmarking Structure Extraction for Vision-Language Models</a></p><p>Useful new benchmarks for VLMs. VLMs are often used for structured extraction in practice, so this benchmarks is not very academic but well aligned with applied quality needs. By Percy Liang's group at Stanford.</p></li><li><p><a href="https://arxiv.org/abs/2410.17434">LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding</a></p><p>"reduces the number of video tokens while preserving visual details of long videos" neat! By folks from Meta AI, a video content powerhouse.</p></li><li><p><a href="https://openreview.net/forum?id=BGnm7Lo8oW">Anon ICLR submission: Towards Learning to Reason at Pre-Training Scale</a></p><p>Interesting idea! "given the first tokens from a large pre-training corpus, the model generates a CoT and receives a reward based on how well the CoT helps predict the following tokens"</p></li><li><p><a href="https://arxiv.org/abs/2410.13863v1">Fluid: Scaling Autoregressive Text-to-image Generative Models with Continuous Tokens</a></p><p>By Google DeepMind.</p></li><li><p><a href="https://arxiv.org/abs/2410.23168">TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters</a></p><p>Researchers might not have tons of compute, but luckily they are smart. This paper solves the problem "When architectural modifications (e.g., channel dimensions) are introduced, the entire model typically requires retraining from scratch". This computational efficiency gain can spur faster iteration of architectural ideas, Neural-Architecture-Search let's go!</p></li><li><p><a href="https://maskgct.github.io/">MaskGCT: Zero-Shot Text-to-Speech with Masked Generative Codec Transformer</a></p><p>Very strong open source non-autoregressive TTS model. Demo space on Hugging Face <a href="https://huggingface.co/spaces/amphion/maskgct">https://huggingface.co/spaces/amphion/maskgct</a></p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://v2.scrimba.com/s08johf0et">OmniParser running in the browser with Transformer.js</a></p><p>Very impressive and useful demo showing how to run OmniParser in the browser directly. As others remarked on X, this has potential to be a core building block for browser extensions.</p></li><li><p><a href="https://www.decart.ai/">Decart launches Oasis: playable simulated Minecraft</a></p><p>This is a phenomenal achievement. It sets the stage for full world simulation. Remember what the first generated images/videos looked like. What makes this launch even more remarkable is that both weights and an interactive web demo with a limited queue are available. Just WOW.</p></li><li><p><a href="https://aligneval.com/">AlignEval: a game/tool to help you build and optimize LLM-evaluator</a></p><p>Very cool project! Source is on GitHub. And this <a href="https://x.com/eugeneyan/status/1851654159692616020">X thread</a> </p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://waymo.com/blog/2024/10/introducing-emma/">Waymo introduces Emma</a></p><p>It's built on top of Gemini's multimodal capabilities.</p></li><li><p><a href="https://hamel.dev/blog/posts/llm-judge/">LLM as a judge for business value by Hamel Husain</a></p><p>You won't find better applied AI findings than this.</p></li><li><p><a href="https://deepmind.google/discover/blog/pushing-the-frontiers-of-audio-generation/">NotebookLM's TTS system explained by Google DeepMind</a></p></li><li><p><a href="https://platform.openai.com/docs/guides/audio">OpenAI Audio generation endpoint</a></p><p>Note this is separate from real-time audio. It allows these combinations:
text in &#8594; text + audio out
audio in &#8594; text + audio out
audio in &#8594; text out
text + audio in &#8594; text + audio out
text + audio in &#8594; text out</p></li><li><p><a href="https://www.youtube.com/watch?v=YR9EztOF0R8">[Video] Learning to Reason, Insights from Language Modeling</a></p><p>By Noah D. Goodman a researcher from Stanford.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Claude Computer Use: RPA on steroids]]></title><description><![CDATA[Week 43 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/claude-computer-use-rpa-on-steroids</link><guid isPermaLink="false">https://codingwithintelligence.com/p/claude-computer-use-rpa-on-steroids</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 27 Oct 2024 15:16:17 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/145f589b-6ce3-4f0e-8c4b-89bf00b71eaf_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>What a BUSY week! Both for me personally (Sunday newsletter day, yay!) and in AI at large. I think everyone saw the Claude Computer Use release and tried answering the question: how ready is this? See this week&#8217;s resources for a hint of the risks currently involved and play with it on your device through Agent.exe - proceed with caution!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NIMK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NIMK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!NIMK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!NIMK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!NIMK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NIMK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:367474,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NIMK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!NIMK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!NIMK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!NIMK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b28ccf-1a0c-4208-be5f-e1e2f6acb80d_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://aider.chat/docs/leaderboards/">Claude Sonnet 3.5 20241022 releases tops Aider leaderboard</a></p><p>For more about upgraded Claude 3.5 Sonnet and  Claude 3.5 Haiku see Anthropic's blog post <a href="https://www.anthropic.com/news/3-5-models-and-computer-use">https://www.anthropic.com/news/3-5-models-and-computer-use</a></p></li><li><p><a href="https://huggingface.co/genmo/mochi-1-preview">Mochi 1 Preview: open source text-to-video model</a></p><p>Samples are impressive. It's great to see a push for open source text-to-video. More people can experiment, learn about mainstream approaches to text-to-video and of course tons of cool clips to generate at compute-cost-price.</p></li><li><p><a href="https://runwayml.com/research/introducing-act-one">Runway releases Act-One: expressive character performance</a></p><p>Works by transferring character performance from a source video to a target generated character. Interestingly, still depends on convincing human performance.</p></li><li><p><a href="https://about.ideogram.ai/canvas">Ideogram releases Dingboard like feature called "Canvas"</a></p><p>Companies in AI move fast and aren't afraid to ~steal~borrow good ideas from each other.</p></li><li><p><a href="https://docs.x.ai/docs">xAI releases API</a></p></li><li><p><a href="https://www.ibm.com/new/ibm-granite-3-0-open-state-of-the-art-enterprise-models">IBM introduces Granite 3.0 models</a></p><p>Model is close to Llama 3.1 8B at 8B size. Released under Apache 2.0.</p></li><li><p><a href="https://huggingface.co/microsoft/OmniParser">Microsoft releases OmniParser: a UI vision extraction model</a></p><p>"OmniParser is a general screen parsing tool, which interprets/converts UI screenshot to structured format, to improve existing LLM based UI agent."</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/e2b-dev/desktop">E2B releases Desktop Sandbox feature beta</a></p><p>Jumping on the Claude Computer Use bandwagon. Awesome to see them move fast.</p></li><li><p><a href="https://github.com/mit-han-lab/duo-attention">DuoAttention: Efficient Long-Context LLM Inference with Retrieval and Streaming Heads</a></p><p>Implementation of the identically titled paper. Awesome efficiency work by the MIT HAN Lab.</p></li><li><p><a href="https://github.com/anthropics/anthropic-quickstarts/tree/main/computer-use-demo">Anthropic Computer Use reference implementation</a></p><p>A containerized Linux desktop for isolated LLM-powered computer use. Fun experiment, evals are missing. Related docs https://docs.anthropic.com/en/docs/build-with-claude/computer-use</p></li><li><p><a href="https://github.com/DAMO-NLP-SG/Inf-CLIP">Inf-CLIP: Breaking the Memory Barrier: Near Infinite Batch Size Scaling for Contrastive Loss</a></p><p>What is very exciting about this result is that most step-change performance unlocks come from discovering computational efficiencies that allow for further scaling. Scale Is All You Need!</p></li><li><p><a href="https://github.com/microsoft/BitNet">bitnet.cpp: the official inference framework for 1-bit LLMs</a></p><p>E.g., BitNet b1.58, by Microsoft. Acceleration is still suboptimal because the lack of direct hardware support.</p></li><li><p><a href="https://github.com/usefulsensors/moonshine">Moonshine: new open source ASR</a></p><p>Licensed under MIT, claims to outperform Whisper "better than similarly-sized Whisper models from OpenAI".</p></li><li><p><a href="https://github.com/dottxt-ai/outlines-core">outlines-core</a></p><p>Structured generation in Rust</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://llava-vl.github.io/blog/2024-09-30-llava-video/">LLaVA-Video-(7B|72B) by LLaVA team</a></p><p>Impressive adaptation of open source for video understanding. Includes Hugging Face checkpoints, dataset and training code.</p></li><li><p><a href="https://www.salesforceairesearch.com/opensource/xGen-MM-Vid/index.html">xGen-MM-Vid (BLIP-3-Video): You Only Need 32 Tokens to Represent a Video Even in VLMs</a></p><p>Interesting work by Salesforce on video encoding.</p></li><li><p><a href="https://arxiv.org/abs/2404.16710">LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding</a></p><p>Interesting work to achieve higher inference performance, not clear which tasks suffer dramatically from the layer skipping described in the paper.</p></li><li><p><a href="https://arxiv.org/abs/2410.16713">Collapse or Thrive? Perils and Promises of Synthetic Data in a Self-Generating World</a></p><p>The results point to collapse due to presence of synthetic data depends on a lack of real data, as long as it is present, synthetic data in the mix doesn't seem to result in a performance collapse.</p></li><li><p><a href="https://arxiv.org/abs/2305.05084">Fast Conformer with Linearly Scalable Attention for Efficient Speech Recognition</a></p></li><li><p><a href="https://arxiv.org/abs/2410.00907">Addition is All You Need for Energy-efficient Language Models</a></p><p>Interesting ideas for making LLMs more efficient but doesn't definitively show whether performance holds up at the highest end of the performance spectrum (e.g. Llama 405B on complex reasoning).</p></li><li><p><a href="https://arxiv.org/abs/2410.02703">Selective Attention Improves Transformer</a></p><p>Interesting idea that intuitively makes sense: not everything in the context window matters for predicting the next token. The authors present an approach to selectively applying attention to the context window.</p></li><li><p><a href="https://arxiv.org/abs/2410.18252">Asynchronous RLHF: Faster and More Efficient Off-Policy RL for Language Models</a></p><p>"Inspired by classical deep RL literature, we propose separating generation and learning in RLHF. This enables asynchronous generation of new samples while simultaneously training on old samples, leading to faster training and more compute-optimal scaling." and "Finally, we verify the scalability of asynchronous RLHF by training LLaMA 3.1 8B on an instruction-following task 40% faster than a synchronous run while matching final performance." awesome work in a deep cross-collaboration between Mila, Allen Institute for AI and DeepMind.</p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://github.com/corbt/agent.exe">Claude Computer Use on your actual computer</a></p><p>Needless to say, this is very risky to run. But it's pretty cool!</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://openai.com/index/simplifying-stabilizing-and-scaling-continuous-time-consistency-models/">OpenAI introduces sCM: breakthrough in image generation speeds</a></p></li><li><p><a href="https://hailuoai.video/">Hailuo SOTA Text|Image-to-Video model</a></p><p>Very impressive results.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;440057b2-b578-4f8e-9f64-0bdd163b7452&quot;,&quot;duration&quot;:null}"></div></li><li><p><a href="https://x.com/_philschmid/status/1847565606964646077">o1 mechanism rumors could be useful for replication</a></p><p>By Philipp Schmid from Hugging Face. He draws comparison to the Stream of Search people released by Stanford researchers in April 2024.</p></li><li><p><a href="https://epochai.org/data/machine-learning-hardware">Epoch AI shares a Machine Learning Hardware Database</a></p><p>AMD MI325X and GB200 unsurprisingly take the performance crown. Cool dataset!</p></li><li><p><a href="https://github.com/vectara/hallucination-leaderboard">Vectara Hallucination leaderboard</a></p><p>Zhipu AI GLM-4-9B-Chat is a surprising #1 scoring model</p></li><li><p><a href="https://simonwillison.net/2024/Oct/24/claude-analysis-tool/">Claude ships Analysis tool to Claude web using JS sandboxing</a></p><p>Short exploration by Simon Willison</p></li><li><p><a href="https://embracethered.com/blog/posts/2024/claude-computer-use-c2-the-zombais-are-coming/">ZombAIs: From Prompt Injection to C2 with Claude Computer Use</a></p><p>Great example of Computer Use vulnerabilities. It's great Anthropic and E2B are pushing computer using LLMs forward but we need clear demonstrations of how they are vulnerable to security risks. The author of Embrace The Red does a great job. They show an E2E example of getting an agent to run malware on the target VM (still sandboxed).</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Edge AI makes waves: Qwen 2.5 Code Interpreter in your browser]]></title><description><![CDATA[Week 42 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/edge-ai-makes-waves-qwen-25-code</link><guid isPermaLink="false">https://codingwithintelligence.com/p/edge-ai-makes-waves-qwen-25-code</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 20 Oct 2024 19:28:48 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a5c943b6-c0d2-4634-b2d3-3409d4bebe71_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3xF5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3xF5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!3xF5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!3xF5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!3xF5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3xF5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:88220,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3xF5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!3xF5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!3xF5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!3xF5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41c5fe5c-0c68-4a4a-b372-7eb1bbb5007d_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Today&#8217;s demo is the most awesome Edge AI application I've seen, check it out. Python generated &amp; executed fully in your browser.</p><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://huggingface.co/spaces/cfahlgren1/qwen-2.5-code-interpreter">Qwen-2.5-Coder 1.5B with access to an in-browser code interpreter</a></p></li></ul><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://mistral.ai/news/ministraux/">Mistral releases 2 Ministral edge focused models</a></p><p>They release a 3B and 8B for mainly embedded use cases. It seems to compare favorably against Llama 3.2 3B on various benchmarks and against Llama 3.1 8B respectively.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/openreasoner/openr">OpenR: An Open Source Framework for Advanced Reasoning with Large Language Models</a></p></li><li><p><a href="https://github.com/openai/swarm">Swarm by OpenAI</a></p><p>A multi-agent orchestration framework.</p></li><li><p><a href="https://www.answer.ai/posts/2024-10-15-how-to-synthesize-data.html">Answer.ai releases fastdata: a synthetic data generation library</a></p><p>The repo: <a href="https://github.com/AnswerDotAI/fastdata">https://github.com/AnswerDotAI/fastdata</a></p></li><li><p><a href="https://huggingface.co/collections/nvidia/llama-31-nemotron-70b-670e93cd366feea16abc13d8">Llama 3.1 Nemotron: finetune by NVIDIA </a></p><p>The mention a slew of high benchmark scores: "This model reaches Arena Hard of 85.0, AlpacaEval 2 LC of 57.6 and GPT-4-Turbo MT-Bench of 8.98, which are known to be predictive of LMSys Chatbot Arena Elo" not sure how well that translates to real-word performance. Run your own evals &#128588; More interestingly, they detail the alignment procedure they used to create the model.</p></li><li><p><a href="https://github.com/facebookresearch/lingua">Meta Lingua: a hackable training and inference library</a></p><p>Looks very easy to modify and play with for trying out ideas.</p></li><li><p><a href="https://github.com/deepseek-ai/Janus">Janus-1.3B, a multimodal understand + generation model by DeepSeek</a></p><p>"Janus is a novel autoregressive framework that unifies multimodal understanding and generation"</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2408.02666">Meta releases Self-Taught Evaluators: iterative improvement through synthetic data</a></p><p>Code is available on GitHub <a href="https://github.com/facebookresearch/RAM/tree/main/projects/self_taught_evaluator">https://github.com/facebookresearch/RAM/tree/main/projects/self_taught_evaluator</a> and weights on Hugging Face <a href="https://huggingface.co/facebook/Self-taught-evaluator-llama3.1-70B">https://huggingface.co/facebook/Self-taught-evaluator-llama3.1-70B</a></p></li><li><p><a href="https://arxiv.org/abs/2407.03181">Fine-Tuning with Divergent Chains of Thought Boosts Reasoning Through Self-Correction in Language Models</a></p><p>Divergent CoT (DCoT) in short: "further improving performance by requiring models to compare multiple reasoning chains before generating a solution in a single inference step". Interesting idea!</p></li><li><p><a href="https://arxiv.org/abs/2310.04444">What's the Magic Word? A Control Theory of LLM Prompting</a></p><p>I like this exploration of a more principled approach to prompting.</p></li><li><p><a href="https://arxiv.org/abs/2410.10630">Thinking LLMs: General Instruction Following with Thought Generation</a></p><p>"We propose a training method for equipping existing LLMs with such thinking abilities for general instruction following without use of additional human data. We achieve this by an iterative search and optimization procedure that explores the space of possible thought generations, allowing the model to learn how to think without direct supervision."</p></li><li><p><a href="https://speechbot.github.io/spiritlm/">Spirit LM: Interleaved Spoken and Written Language Model</a></p><p>By Meta, a foundation multimodal language model that freely mixes text and speech.</p></li><li><p><a href="https://arxiv.org/abs/2409.12737">MEXMA: Token-level objectives improve sentence
representations</a></p><p>Improved multilingual embeddings by Meta.</p></li><li><p><a href="https://arxiv.org/abs/2410.13821">Artificial Kuramoto Oscillatory Neurons</a></p><p>By Max Welling et al. "It replaces threshold units with generalized Kuramoto oscillators. It dynamically binds neurons, generates waves of activations, is adversarially robust, calibrated for uncertainty, and can reason. If you ask me: this is the next big thing!"</p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://x.com/mlstreettalk/status/1845223735990813055?s=52">Fran&#231;ois Chollet keynote talk about ARC</a></p></li><li><p><a href="https://www.youtube.com/watch?v=c3N2H3Z5S3I">Chris Manning - Meaning and Intelligence in Language Models (COLM 2024)</a></p><p>Christopher Manning is the director of the SAIL institute at Stanford.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[Aria MoE A3.9B a new open source multimodal LLM]]></title><description><![CDATA[Week 41 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/aria-moe-a39b-a-new-open-source-multimodal</link><guid isPermaLink="false">https://codingwithintelligence.com/p/aria-moe-a39b-a-new-open-source-multimodal</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 13 Oct 2024 13:39:11 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Z-mz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Z-mz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Z-mz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!Z-mz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!Z-mz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!Z-mz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Z-mz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1081148,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Z-mz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!Z-mz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!Z-mz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!Z-mz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b5ae600-9096-42ea-8751-ca2f5ad46f46_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://www.rhymes.ai/blog-details/aria-first-open-multimodal-native-moe-model">Aria MoE A3.9B new SOTA open source multimodal LLM</a></p><p>Performance looks competitive with Pixtral, Llama 3.2 11B and in some cases even with GPT4o/GPT4o-mini.</p></li><li><p><a href="https://www.swebench.com/multimodal">SWE-bench extended with multimodal tasks</a></p></li><li><p><a href="https://x.com/victortaelin/status/1844886809005687270?s=46">$10k (o1) reasoning challenge by Victor Taelin</a></p><p>A challenge to see if frontier LLMs can reason in a way that causes generalization. The task is to invert a perfect binary tree but he adds 3 criteria that make it novel enough to be outside of the pretraining corpus with high likelihood.</p></li><li><p><a href="https://www.scaledfoundations.ai/post/matmamba">MatMamba: An Elastic and Efficient Neural Network Architecture</a></p><p>"Combining the speed of State Space Models (SSMs) like Mamba2 with the adaptability of Matryoshka-style learning." by Scaled Foundations a startup with MIT co-founders working on autonomous robotics. Implementation available on GitHub https://github.com/scaledfoundations/matmamba</p></li><li><p><a href="https://www.theinformation.com/articles/openai-eases-away-from-microsoft-data-centers?rc=w7r60d">OpenAI Leaders Say Microsoft Isn&#8217;t Moving Fast Enough to Supply Servers</a></p><p>Paywalled unfortunately. But the key message is that there are rumors that OpenAI is becoming more independent from Microsoft at the datacenter level. I guess they want to move faster than Microsoft allows at the infrastructure level.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/yangling0818/itercomp">IterComp: Iterative Composition-Aware Feedback Learning from Model Gallery for Text-to-Image Generation</a></p><p>Retry is all you need.</p></li><li><p><a href="https://huggingface.co/SWivid/F5-TTS">F5-TTS open source model</a></p><p>Very good emotive quality TTS: "non-autoregressive text-to-speech system based on flow matching with Diffusion Transformer (DiT)", see paper and repo for more details.</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2410.05258">Differential Transformer</a></p><p>An interesting architecture modification that shows better scaling properties than vanilla Transformers. Will be interesting to see if large open source/frontier groups adopt this.</p></li><li><p><a href="https://arxiv.org/abs/2410.00907">Addition is All You Need for Energy-efficient Language Models</a></p><p>This paper proposes replacing multiplications with additions and shows some convincing data that there's merit to this idea. As with all architecture modifications, the jury isn't out until the ideas are scaled up.</p></li><li><p><a href="https://arxiv.org/abs/2410.02536">Intelligence at the Edge of Chaos</a></p><p>In this paper they pretrain GPT2 on cellular automata and show that pretraining on more complex automata increases downstream task performance on tasks like chess and abstract reasoning. Fascinating result that seems to reveal something fundamental about transfer learning.</p></li><li><p><a href="https://static.synthlabs.ai/preprints/Generative_Reward_Models.pdf">Generative Reward Models - A Unified
Approach to RLHF and RLAIF</a></p><p>Interesting survey of alignment methods and a proposed technique that allows for combining expensive human preference data with synthetically generated preference data. They emphasize OOD (out-of-distribution) performance which is sometimes not highlighted enough when comparing alignment techniques. A co-author of the paper invented the DPO method at Stanford.</p></li><li><p><a href="https://arxiv.org/abs/2410.02707">LLMs Know More Than They Show: On the Intrinsic Representation of LLM Hallucinations</a></p><p>"In this work, we show that the internal representations of LLMs encode much more information about truthfulness than previously recognized." the challenge is getting them to reliably utilize the correct information they contain.</p></li><li><p><a href="https://arxiv.org/abs/2403.17031">The N+ Implementation Details of RLHF with PPO: A Case Study on TL;DR Summarization</a></p><p>Reproducing Reinforcement Learning implementations is notoriously fraught with many gotchas that individually can cause failure if handled incorrectly. A massive contribution by folks from Mila, Hugging Face and others.</p></li><li><p><a href="https://adityayedetore.github.io/assets/pdf/emnlp_2024_semantic_cues_to_hierarchy.pdf">Semantic Training Signals Promote Hierarchical Syntactic Generalization in Transformers</a></p><p>Interesting exploration of the effects of hierarchical biases in the Transformer architecture.</p></li><li><p><a href="https://arxiv.org/abs/2407.01687">Deciphering the Factors Influencing the Efficacy of Chain-of-Thought: Probability, Memorization, and Noisy Reasoning</a></p><p>Analyzing CoT behavior in LLMs using a specific task (decoding shift ciphers). Conclusion is positive. "Overall, we conclude that CoT prompting performance reflects both memorization and a probabilistic version of genuine reasoning."</p></li><li><p><a href="https://oooolga.github.io/JEDi.github.io/">Beyond FVD: Enhanced Evaluation Metrics for Video Generation Quality</a></p><p>Interesting work from Mila that aims to improve on metrics that quantify video generation quality. Since that's so hard to quantify I think this work has a lot of potential to help researchers discover which techniques actually make a meaningful difference. I'm sure all text-to-video startups are all over this (Luma Labs, Runway, OpenAI, Kling, etc.).</p></li><li><p><a href="https://arxiv.org/abs/2410.07073">Pixtral 12B technical report</a></p><p>Some welcome details on the new Pixtral multimodal model.</p></li><li><p><a href="https://arxiv.org/abs/2410.02902">Better Instruction-Following Through Minimum Bayes Risk</a></p></li><li><p><a href="https://arxiv.org/abs/2410.01131">nGPT: Normalized Transformer with Representation Learning on the Hypersphere</a></p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://huggingface.co/spaces/jbilcke-hf/FacePoke">FacePoke: impressive training free face image manipulation</a></p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://cursor.directory/">Cursor Directory</a></p><p>A list of crowdsourced .cursorrules. No need to write your own prompts for language specific Cursor rules.</p></li><li><p><a href="https://youtu.be/vJ8aEO6ggOs">[Video] AMD MI325X introduction keynote</a></p></li><li><p><a href="https://darioamodei.com/machines-of-loving-grace">Machines of Loving Grace - long read by Anthropic founder Dario Amodei</a></p><p>Subtitled "How AI Could Transform the World for the Better", a positive and grounded essay about the impact of AI from on of the, if not the, best AI labs in the world.</p></li><li><p><a href="https://notes.haroldbenoit.com/ml/llms/inference/sampling/entropy-based-sampling">Explainer of Entropy based sampling</a></p></li><li><p><a href="https://klingai.com/community/skit">Kling AI community short films</a></p><p>Check out SOTA generative AI projects</p></li><li><p><a href="https://huggingface.co/spaces/LLM360/TxT360">TxT360 - open source 15T corpus and processing pipeline</a></p><p>"We demonstrate a simple but effective upsampling recipe that creates a 15+ trillion-token corpus, outperforming FineWeb 15T on several key metrics."</p></li><li><p><a href="https://www.youtube.com/watch?app=desktop&amp;v=zmnzW0r_g8k&amp;list=PLjOo65uEP4cYhV7c2whkhDfWy58XFj7yL&amp;index=8">Forge by Nous Research @ Nouscon 2024</a></p></li><li><p><a href="https://github.com/GAIR-NLP/O1-Journey">O1 replication journey by GAIR (Shanghai Jiaotong University)</a></p></li><li><p><a href="https://www.youtube.com/watch?v=9VFHuGaq7SU&amp;list=PL_lsbAsL_o2B_znuvm-pDtV_cRhpqZb8l">[Video] PyTorch 2024 SF playlist</a></p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item><item><title><![CDATA[A new kind of Foundation Models: LFMs]]></title><description><![CDATA[Week 40 of Coding with Intelligence]]></description><link>https://codingwithintelligence.com/p/a-new-kind-of-foundation-models-lfms</link><guid isPermaLink="false">https://codingwithintelligence.com/p/a-new-kind-of-foundation-models-lfms</guid><dc:creator><![CDATA[Rick Lamers]]></dc:creator><pubDate>Sun, 06 Oct 2024 23:45:35 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3b343cb4-4057-4009-8bf7-f30e40795b70_1680x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gC7S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gC7S!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!gC7S!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!gC7S!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!gC7S!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gC7S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:254549,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gC7S!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png 424w, https://substackcdn.com/image/fetch/$s_!gC7S!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png 848w, https://substackcdn.com/image/fetch/$s_!gC7S!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!gC7S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5755b1d0-b3c2-4d20-91bd-0395373c8595_1680x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Liquid AI&#8217;s new 1B, 3B and 40B LFMs</figcaption></figure></div><p>&#128240; <strong>News</strong></p><ul><li><p><a href="https://github.com/openai/whisper/pull/2361/files">New Turbo Whisper model by OpenAI</a></p></li><li><p><a href="https://www.liquid.ai/liquid-foundation-models">Liquid AI launches Liquid Foundation Models</a></p><p>They release 1B, 3B, and 40B LFMs. The main improvement seems to be their efficiency on long context inputs, beyond Transformer and even SSM models. Their work isn't open source (yet) so claims will be hard to verify until they do. Some more historical context in <a href="https://www.liquid.ai/blog/liquid-neural-networks-research">https://www.liquid.ai/blog/liquid-neural-networks-research</a>. Try them in the <a href="https://playground.liquid.ai/login">Playground</a>.</p></li><li><p><a href="https://ai.meta.com/research/movie-gen/">Meta AI showcases high quality text-to-video model</a></p><p>No model or product release at this time unfortunately. Quality looks as good or better than OpenAI's Sora.</p></li><li><p><a href="https://openai.com/index/introducing-canvas/">OpenAI releases Canvas</a></p><p>Feature is similar to Artifact from Anthropic and the Cursor AI IDE. Responses have been mixed, use it yourself and leave your thoughts in the comments! I'd love to read it.</p></li></ul><p>&#128230; <strong>Repos</strong></p><ul><li><p><a href="https://github.com/princeton-nlp/HELMET">HELMET: How to Evaluate Long-context Language Models Effectively and Thoroughly</a></p><p>A long context eval from the Princeton NLP group.</p></li><li><p><a href="https://github.com/xjdr-alt/entropix">xjdr of X fame open sources sampling technique for emulating o1-style LLM completions</a></p></li><li><p><a href="https://github.com/KellerJordan/modded-nanogpt">Modded-NanoGPT: Keller Jordan adapts NanoGPT to show improved optimizer</a></p><p>"The proposed optimizer has the following properties: half the memory usage of Adam, 1.43x faster training, &lt;7% wallclock overhead."</p></li></ul><p>&#128196; <strong>Papers</strong></p><ul><li><p><a href="https://arxiv.org/abs/2410.02525">Contextual Document Embeddings</a></p><p>The method optimizes contrastive learning via adversarial batch construction and utilizes a two-stage encoder &#981;(d; D) to incorporate corpus statistics, enabling superior cross-domain generalization.</p></li><li><p><a href="https://arxiv.org/abs/2310.17086">Transformers Learn Higher-Order Optimization Methods for In-Context Learning: A Study with Linear Models</a></p></li><li><p><a href="https://arxiv.org/abs/2410.02660">How to Train Long-Context Language Models (Effectively)</a></p><p>Princeton NLP group proposes the ProLong model and reaches an effective context window size of 512k.</p></li><li><p><a href="https://arxiv.org/abs/2410.02089">RLEF: Grounding Code LLMs in Execution Feedback with Reinforcement Learning</a></p><p>By Meta AI. Very interesting new post-training approach.</p></li></ul><p>&#128241; <strong>Demos</strong></p><ul><li><p><a href="https://xnought.github.io/vae-explainer/">VAE explained interactively in your browser</a></p></li></ul><p>&#128218; <strong>Resources</strong></p><ul><li><p><a href="https://x.com/sedielem/status/1840170796876214427?s=46">Intuition of diffusion models in an X post</a></p><p>By research scientist Sander Dieleman at Google DeepMind.</p></li><li><p><a href="https://simonwillison.net/2024/Oct/2/not-digital-god/">Impressions of OpenAI Dev Day (10/1/24)</a></p><p>By your (my?) favorite LLM sommelier Simon Willison.</p></li><li><p><a href="https://chus.space/blog/2024/ssm_1_context/">State Space Models (1): introduction to SSMs</a></p><p>Neat 3-part (all 3 available) blog post series about intuition behind SSM models.</p></li><li><p><a href="https://aider.chat/docs/more/infinite-output.html">Infinite output through prefilling (in Aider)</a></p></li><li><p><a href="https://www.youtube.com/watch?v=OPXslklVBZc">Aider architect/editor feature</a></p></li><li><p><a href="https://symflower.com/en/company/blog/2024/dev-quality-eval-v0.6-o1-preview-is-the-king-of-code-generation-but-is-super-slow-and-expensive/">Code generation deep dive: tons of detail</a></p></li><li><p><a href="https://www.interconnects.ai/p/reverse-engineering-openai-o1">Reverse engineering OpenAI&#8217;s o1</a></p><p>By Nathan Lambert from Allen AI.</p></li><li><p><a href="https://www.youtube.com/watch?v=AfAmwIP2ntY">[Video] Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters</a></p><p>Yannic Kilcher covers a recent paper from Google DeepMind about scaling test-time compute.</p></li></ul><div><hr></div><p>Want more? Follow me on X! <a href="https://x.com/RickLamers">@ricklamers</a></p>]]></content:encoded></item></channel></rss>