<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[abstractions]]></title><description><![CDATA[perspectives on ai, tech and business]]></description><link>https://abstractions.rakshitranjan.com</link><image><url>https://substackcdn.com/image/fetch/$s_!gUW3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9670a20-4b7b-4879-b018-dcb679a5b59e_608x608.png</url><title>abstractions</title><link>https://abstractions.rakshitranjan.com</link></image><generator>Substack</generator><lastBuildDate>Sat, 09 May 2026 01:57:01 GMT</lastBuildDate><atom:link href="https://abstractions.rakshitranjan.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Rakshit Ranjan]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[rakshran@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[rakshran@substack.com]]></itunes:email><itunes:name><![CDATA[Rakshit Ranjan]]></itunes:name></itunes:owner><itunes:author><![CDATA[Rakshit Ranjan]]></itunes:author><googleplay:owner><![CDATA[rakshran@substack.com]]></googleplay:owner><googleplay:email><![CDATA[rakshran@substack.com]]></googleplay:email><googleplay:author><![CDATA[Rakshit Ranjan]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Why Google might win? (5/5)]]></title><description><![CDATA[On distribution]]></description><link>https://abstractions.rakshitranjan.com/p/why-google-might-win-55</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/why-google-might-win-55</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Mon, 27 Apr 2026 16:40:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gUW3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9670a20-4b7b-4879-b018-dcb679a5b59e_608x608.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>So far in this series, I have argued that Google has four supply-side moats.</p><ol><li><p><a href="https://abstractions.rakshitranjan.com/p/why-google-might-win">Custom TPUs</a> give it a cost edge on every token it serves.</p></li><li><p><a href="https://abstractions.rakshitranjan.com/p/why-google-might-win-25">Private infrastructure</a>&#8211;efficient, integrated, end-to-end owned&#8211;eliminates the margin leakage every other lab pays.</p></li><li><p><a href="https://abstractions.rakshitranjan.com/p/why-google-might-win-35">Proprietary data</a> that is complete, multimodal, and compounding as the public internet runs dry.</p></li><li><p><a href="https://abstractions.rakshitranjan.com/p/why-google-might-win-45">A frontier model</a>, Gemini, trained on top of all three and is fairly competitive with the other players.</p></li></ol><p>If the argument holds at four layers, the question becomes whether any of it actually reaches users. Cost advantages are inert without distribution. Better models are invisible without surfaces. The fifth layer is where structural advantage either compounds into a business or stays trapped as an interesting lab result.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://abstractions.rakshitranjan.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading abstractions! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>In software, where distribution and transaction costs are zero, Ben Thompson&#8217;s <a href="https://stratechery.com/2015/aggregation-theory/">Aggregation Theory</a>, now a decade old, is the seminal thesis on who wins:</p><blockquote><p><em>The most important factor determining success is the user experience: the best distributors/aggregators/market-makers win by providing the best experience, which earns them the most consumers/users, which attracts the most suppliers, which enhances the user experience in a virtuous cycle.</em></p></blockquote><p>The question I had when I started thinking about it was whether it would still hold in the AI era. I think it will.</p><p>I strongly believe that we are in an interim, transient state in the development of AI as a technology.</p><p>When ChatGPT broke out in late 2022, the chat interface was essentially an accident. OpenAI was sitting on GPT-3.5 and needed a way to demo it. A dialogue box was the obvious shell. What no one expected was that the shell itself would go viral, with one of the fastest consumer product adoption rates in history. Every other lab scrambled to ship a chat interface in response. That is how chat became the default.</p><p>But chat is not the end state. We are already seeing three stages unfolding in parallel.</p><ol><li><p><strong>Standalone chat:</strong> ChatGPT, Claude, Gemini as destinations. You go to the product, type, read, leave. This is where most attention currently sits.</p></li><li><p><strong>Embedded Copilot:</strong> Gemini in Docs, Copilot in Excel, Cursor in the IDE, Wispr Flow in the keyboard. You don&#8217;t go anywhere; the intelligence meets you inside the workflow.</p></li><li><p><strong>Autonomous action:</strong> Agents that take a goal and execute across products&#8211;researching, drafting, scheduling, purchasing. The user specifies outcomes; the software handles the intermediate steps. e.g., Claude Code and Cowork.</p></li></ol><p>As new long-run paradigms emerge, the constant underlying them is that users adopt technology to get a job done. Each major platform shift&#8211;the PC, the web, the smartphone&#8211;succeeded because it let people accomplish more with less friction. AI is the next layer in that sequence, and the one most likely to dissolve the boundary between &#8220;using an application&#8221; and &#8220;getting a thing done&#8221;.</p><p>Google has been laying the groundwork for this longer than any company competing in AI today. Cutting-edge machine learning has been quietly embedded across its products for over a decade, improving user experience in ways most people don&#8217;t recognise as AI at all.</p><ul><li><p>Autocomplete and spelling correction in Search.</p></li><li><p>Magic Eraser and visual search in Google Photos.</p></li><li><p>Near-native translation in Google Translate.</p></li><li><p>Smart Compose finishing sentences in Gmail.</p></li><li><p>Live traffic and rerouting in Google Maps.</p></li><li><p>Call Screening and Circle to Search on Pixel.</p></li><li><p>Autonomous driving at Waymo.</p></li></ul><p>All of it is downstream of foundational research by the Google Brain team, which has always focused on commercialisation&#8211;on embedding AI into products billions of people already use every day.</p><p>That is the point about distribution.</p><p>Getting a thing done entails three steps:</p><ol><li><p>Capture</p></li><li><p>Understand</p></li><li><p>Serve</p></li></ol><p><strong>On capture</strong>, Google operates seven products with over two billion users each&#8211;Search, Android, Chrome, YouTube, Gmail, Maps, Photos. The nearest peer is Apple, only because the iPhone is itself a universal intent-capture device.</p><p>Muscle memory already routes users: Search for facts, Maps for places, YouTube for how-tos, Gmail for correspondence. As AI advances, the range of inputs will expand&#8211;voice, text, vision, ambient sensors, eventually neural. Google already controls most of the clients through which those inputs will flow.</p><p><strong>On understanding</strong>, the proprietary data across services compounds. Location from Maps, queries from Search, calendar from Gmail, photos from Photos, viewing history from YouTube&#8211;all resolved to the same user account. That level of context is what turns a vague ask into a specific task.</p><p><strong>On serving</strong>, the products needed to act on intent&#8211;calendar, email, maps, documents, browser, storage&#8211;are ones Google already ships. Every step of an agentic workflow is likely to reduce to an existing Google surface.</p><p>While Google is broadly dominant, advances in each of these three layers are also coming and will continue to come from different companies. Wispr Flow shows what voice intent capture looks like when done natively across devices. Perplexity shows what retrieval-first search can be when built AI-first. And Claude Code shows what agentic reasoning looks like when embedded in a developer workflow.</p><p>But subsuming these layers, either through acquisition or by shipping a native feature, is not a stretch for Google. A platform advantage of this breadth can be deployed anytime.</p><p>However, caveats do exist. There are at least three places where this thesis could break.</p><p>The first is that an incumbent subsuming a challenger is not automatic. Microsoft spent years trying to absorb Slack with Teams and largely succeeded, but only because enterprise buyers chose bundling. Google spent similar years trying to subsume WhatsApp with a succession of products&#8211;Allo, Duo, Chat, Meet&#8211;and mostly failed, because consumer network effects sit with the incumbent. The bet in this essay is that AI belongs to the first pattern, not the second. The case: unlike social, where the network is the product, AI capability is increasingly a commoditised input that integrates most cleanly where the user already is. Distribution compounds.</p><p>The second is form factor. If the dominant interface of the next decade is hardware, Google does not make&#8211;OpenAI&#8217;s collaboration with Jony Ive, Meta&#8217;s Ray-Bans, Apple&#8217;s on-device intelligence&#8211;then Google&#8217;s surface area stops being universal. Android is a hedge, not a guarantee. The ambient-compute assumption in this essay depends on Google retaining at least parity at the device layer.</p><p>The third is speed. Google is a public company with a billion-user Search business to defend and active antitrust cases in multiple jurisdictions. It will never move like a startup. It compensates with scale, bundling, and the ability to absorb a category through acquisition or feature launch. That trade&#8211;slower but broader&#8211;only holds if the next two or three years don&#8217;t produce a challenger large enough to refuse acquisition.</p><p>None of these is fatal on its own. But &#8220;Google might win&#8221; was always a probabilistic claim, not a foregone one.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://abstractions.rakshitranjan.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading abstractions! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why Google might win? (4/5)]]></title><description><![CDATA[On the model]]></description><link>https://abstractions.rakshitranjan.com/p/why-google-might-win-45</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/why-google-might-win-45</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Sun, 05 Apr 2026 18:26:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Es8S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>So far in this series, I have argued that Google has three underlying moats.</p><ol><li><p><a href="https://abstractions.rakshitranjan.com/p/why-google-might-win">Custom TPUs</a> give it a cost edge on every token it serves.</p></li><li><p>Private <a href="https://abstractions.rakshitranjan.com/p/why-google-might-win-25">infrastructure</a> that was built to run Search and YouTube eliminates the margin leakage every other lab pays.</p></li><li><p>Proprietary <a href="https://abstractions.rakshitranjan.com/p/why-google-might-win-35">data</a>&#8211;complete, multimodal, and compounding as the public internet runs dry&#8211;is the layer that capital alone cannot replicate.</p></li></ol><p>None of this matters without a frontier model on top. Thankfully, Google has one.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://abstractions.rakshitranjan.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading abstractions! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>But in early 2023, that was not obvious. ChatGPT had reached 100 million users within two months of launch, Google had declared code red internally, and its first response, &#8220;Bard&#8221;, gave an erroneous answer in its launch demo and wiped $100 billion in market cap in a single day.</p><p>In April 2023, Sundar Pichai, acting like a wartime CEO, took a bold step. He merged<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> Google Brain and Google DeepMind&#8211;the two AI research labs within the company.</p><p>Google Brain was the applied research arm: obsessed with getting AI into products, optimising for what could be trained, deployed, and scaled across Search, Ads, and Translate. Its theory of intelligence was statistical: expose a model to enough data, and useful patterns emerge. Scale is the mechanism. The <a href="https://rakshitranjan.com/attention">Transformer</a> paper that every major model is built on came out of Brain.</p><p>Google DeepMind was solving a different problem: artificial general intelligence, a system that could reason through problems it had never seen. Its method was reinforcement learning: define a precise feedback signal, then train the model to maximise it. Think of it like learning a video game by trial and error. You try something, see whether the score went up, adjust, try again (except across millions of attempts).</p><p>Over the course of the past two years, this collaboration, combined with the discussed structural moats, has led to Gemini having distinctive characteristics.</p><ol><li><p>Gemini architectures allow the processing of up to 1-2 million tokens in the context window (among the highest in commercial models).</p></li><li><p>Gemini serving costs have fallen 78% year-over-year, and API pricing runs 5-10x cheaper than OpenAI&#8217;s comparable models.</p></li><li><p>Gemini is natively multimodal (a single neural network trained on text, images, video, and audio together from the start). When competing models process an image, a separate encoder translates pixels into a representation the text model can read (like handing a language-only brain a written description of a photograph). In practice, this gives Gemini an edge in tasks that require reasoning across modalities simultaneously&#8211;analysing a video&#8217;s visual content alongside its audio, or interpreting a chart embedded in a document in relation to surrounding text.</p></li></ol><p>Consequently, it is imperative to analyse what this translates to on the inference layer&#8211;in terms of utility and applications. To evaluate the models holistically, I believe three parameters matter:</p><p>The first is whether people prefer using it.</p><p>On Arena (formerly LMArena), where real users vote on two anonymous models across whatever tasks they actually care about, Gemini has two models in the top five in the Text Arena. More than any other company. Claude leads the leaderboard; Gemini trails narrowly.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Es8S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Es8S!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png 424w, https://substackcdn.com/image/fetch/$s_!Es8S!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png 848w, https://substackcdn.com/image/fetch/$s_!Es8S!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!Es8S!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Es8S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png" width="1456" height="892" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:892,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Es8S!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png 424w, https://substackcdn.com/image/fetch/$s_!Es8S!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png 848w, https://substackcdn.com/image/fetch/$s_!Es8S!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!Es8S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75eaa582-df89-4cca-b669-c237230f7c04_1844x1130.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On Text-to-Image and Text-to-Video Arena, Google is dominant with Gemini and Veo taking most of the positions.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!x4E-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!x4E-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png 424w, https://substackcdn.com/image/fetch/$s_!x4E-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png 848w, https://substackcdn.com/image/fetch/$s_!x4E-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png 1272w, https://substackcdn.com/image/fetch/$s_!x4E-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!x4E-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png" width="1456" height="867" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:867,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!x4E-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png 424w, https://substackcdn.com/image/fetch/$s_!x4E-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png 848w, https://substackcdn.com/image/fetch/$s_!x4E-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png 1272w, https://substackcdn.com/image/fetch/$s_!x4E-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48286a24-90cf-4aa2-ac93-8e00c7ee5a6a_1844x1098.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!f0I5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!f0I5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png 424w, https://substackcdn.com/image/fetch/$s_!f0I5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png 848w, https://substackcdn.com/image/fetch/$s_!f0I5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png 1272w, https://substackcdn.com/image/fetch/$s_!f0I5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!f0I5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png" width="1456" height="867" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:867,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!f0I5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png 424w, https://substackcdn.com/image/fetch/$s_!f0I5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png 848w, https://substackcdn.com/image/fetch/$s_!f0I5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png 1272w, https://substackcdn.com/image/fetch/$s_!f0I5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F255bade0-4608-4048-b6e2-dd579787567d_1844x1098.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>However, in the Code Arena, Claude is the undisputed king with Gemini coming in at seventh.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kW-1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kW-1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png 424w, https://substackcdn.com/image/fetch/$s_!kW-1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png 848w, https://substackcdn.com/image/fetch/$s_!kW-1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png 1272w, https://substackcdn.com/image/fetch/$s_!kW-1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kW-1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png" width="1456" height="867" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:867,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kW-1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png 424w, https://substackcdn.com/image/fetch/$s_!kW-1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png 848w, https://substackcdn.com/image/fetch/$s_!kW-1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png 1272w, https://substackcdn.com/image/fetch/$s_!kW-1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1af0095b-6cce-415a-8d2a-f040612a969f_1844x1098.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The second is performance on the hardest problems with canonical right answers&#8211;competition mathematics, formal science, and competitive programming. Gemini Deep Think holds gold-medal level across multiple Olympiads. On Humanity&#8217;s Last Exam<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>, which tests frontier expert knowledge across all domains, it holds the highest published score.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nrti!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nrti!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png 424w, https://substackcdn.com/image/fetch/$s_!Nrti!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png 848w, https://substackcdn.com/image/fetch/$s_!Nrti!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png 1272w, https://substackcdn.com/image/fetch/$s_!Nrti!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nrti!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png" width="1456" height="537" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:537,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Nrti!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png 424w, https://substackcdn.com/image/fetch/$s_!Nrti!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png 848w, https://substackcdn.com/image/fetch/$s_!Nrti!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png 1272w, https://substackcdn.com/image/fetch/$s_!Nrti!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6433f27-4cd4-4d56-9e03-325385b14790_2256x832.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The third is genuinely novel reasoning, measured by ARC-AGI-2. This benchmark was specifically designed to be immune to the normal tricks&#8211;it cannot be gamed by training on similar problems, because the problems are genuinely novel visual reasoning puzzles that have never appeared in any training corpus. At launch in early 2025, every major model scored near zero. Chain-of-thought reasoning, extended compute, and even brute force failed to make a mark. The benchmark was designed such that it could not be gamified.</p><p>Gemini Deep Think, however, scored 84.6%, verified independently. Gemini 3.1 Pro scored 77.1%. Going from near zero to above 80% was made possible through a reinforcement learning-based reasoning approach, applied to a problem class that scaling-only approaches cannot crack.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-GqF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-GqF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png 424w, https://substackcdn.com/image/fetch/$s_!-GqF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png 848w, https://substackcdn.com/image/fetch/$s_!-GqF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png 1272w, https://substackcdn.com/image/fetch/$s_!-GqF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-GqF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png" width="1456" height="850" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:850,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-GqF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png 424w, https://substackcdn.com/image/fetch/$s_!-GqF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png 848w, https://substackcdn.com/image/fetch/$s_!-GqF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png 1272w, https://substackcdn.com/image/fetch/$s_!-GqF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F844679e9-8ae6-4971-bd7d-18d64cdc1c5c_1928x1126.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>These artefacts prove that Google has been able to achieve best in class outcomes with Gemini. To be able to do that within two years of the Bard fiasco is a truly commendable feat.</p><p>These performances are not set in stone. And as is evident from the leaderboards above, other frontier labs are not far behind. Talent is fungible, techniques are being replicated, and access to capital is also democratised to a large extent. Hence, the only thing that matters at this layer is sustained competitiveness. Combining it with the other structural advantages and distribution (which will be discussed next), Google can ensure dominance in the time to come.</p><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Under its 2014 acquisition terms, DeepMind was guaranteed operational autonomy and an independent ethics board, allowing the London-based lab to operate in a silo that actively resisted integration with Google Brain until the competitive threat of OpenAI forced their 2023 merger.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Humanity&#8217;s Last Exam, a multi-modal benchmark at the frontier of human knowledge, designed to be the final closed-ended academic benchmark of its kind with broad subject coverage. The dataset consists of 2,500 challenging questions across over a hundred subjects. They publicly release these questions, while maintaining a private test set of held out questions to assess model overfitting.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Why Google might win? (3/5)]]></title><description><![CDATA[On data]]></description><link>https://abstractions.rakshitranjan.com/p/why-google-might-win-35</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/why-google-might-win-35</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Wed, 25 Mar 2026 17:47:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gUW3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9670a20-4b7b-4879-b018-dcb679a5b59e_608x608.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Scaling laws have shown, consistently, that model performance improves with data. The frontier labs have been scaling models and data together, and the public internet is the shared corpus everyone has been drawing from.</p><p>A data moat should largely come from having access to high volumes of proprietary data to train on. Google does sit on a huge stockpile of private data across its various services but so does every other major tech company.</p><p>Meta has the social graph. Microsoft has two decades of enterprise productivity data. Apple has billions of devices generating behavioural signals that no one else can access, and by design, cannot aggregate them centrally for training.</p><p>I believe that Google stands out because of 3 unique data properties that it possesses: completeness, multimodality, and compounding.</p><h3>1. Completeness</h3><p>Most publicly available AI training data captures a fragment of human cognition. Wikipedia captures knowledge. Reddit captures opinion. GitHub captures code. These are useful snapshots of thought but not behaviour.</p><p>Google can trace the full arc of a human decision and map it accurately to individual users.</p><ul><li><p>Search captures real-time intent in natural language.</p></li><li><p>YouTube captures what people watch after they search, revealing the interest and learning signals that sit between intent and action.</p></li><li><p>Chrome records browsing behaviour across the entire web, not just Google properties.</p></li><li><p>Google Analytics, installed on roughly 85% of the web, captures what those same people do next: what they buy, where they click, how they behave commercially.</p></li><li><p>Google Pay captures direct transaction data, closing the loop between intent and purchase.</p></li><li><p>Gmail captures the downstream evidence of decisions: order confirmations, travel bookings, subscription receipts. It sees what people did after they decided.</p></li><li><p>Nest and Google Home record domestic behaviour: when people are present, how they live, what their environment looks like.</p></li><li><p>Maps and Waymo record where they end up in the physical world.</p></li><li><p>Fitbit captures biological response.</p></li></ul><p>No other competitor comes close to this breadth.</p><p>A model trained on intent alone becomes better at predicting what people want. A model trained on intent and outcome becomes better at helping people get there.</p><h3>2. Multimodality</h3><p>Text has historically been the dominant modality of AI training and all frontier labs excel at it.</p><p>Google&#8217;s advantage is structurally significant in the other modalities.</p><ul><li><p>YouTube has billions of videos where audio, visual, and text describe the same moment simultaneously. Videos carry a spoken track, a visual frame, and an auto-generated transcript timed to the millisecond. </p></li><li><p>Street View and Waymo add spatial and physical understanding at a scale no competitor has approached. Street View is millions of miles of continuous visual imagery, each frame tagged with precise geographic coordinates. Waymo layers in LiDAR and real-world sensor data from live driving conditions. Together they give a model something close to a physical map of the world&#8211;as it actually appears.</p></li><li><p>Google Photos holds trillions of images, each tagged with time, location, and the objects and scenes they contain.</p></li><li><p>Google Translate has produced the most comprehensive multilingual parallel corpus in existence across 100+ languages, where the same meaning is expressed in structurally different linguistic forms simultaneously.</p></li></ul><p>The unique characteristic here is how the modalities align&#8211;audio, visual, and text describing the same event. That alignment is what lets a model learn the relationship between modalities.</p><p>This can&#8217;t be assembled on demand. It accumulated over decades because the products required it. That is why it is structurally different from data assets built with AI in mind.</p><h3>3. Compounding</h3><p>Google&#8217;s data grows because it is generated by use. The behavioural signals embedded in that use provide a form of continuous feedback at a scale no lab can replicate through paid annotation. Every interaction is grounded in what a real person actually did, not what a model predicted they would do.</p><p>That distinction matters because of what scaling laws tell us about data. As frontier models scale, the data requirement scales with them. The public internet, which every major lab has been drawing from, cannot keep pace. Common Crawl, Wikipedia, arXiv, GitHub are finite reservoirs, and researchers estimate they will be effectively exhausted for training purposes by 2026.</p><p>The obvious response is synthetic data. If you run out of human-generated text, you generate more using existing models. This works up to a point. Models trained on successive generations of synthetic data tend to drift from reality, a phenomenon researchers have started calling model collapse. Google&#8217;s live behavioural feed reduces this risk because a significant portion of its training signal is anchored to real-world actions. It is not immune to the problem, but it is structurally less exposed than labs whose pipelines depend primarily on synthetic generation and licensed text.</p><div><hr></div><p>The first generation of AI products, chatbots, copilots, summarisation tools etc. required models that understand language well. The next generation requires something harder. Agents that book travel, execute multi-step research, navigate physical environments, and handle purchases need models that understand how the world actually works: spatial relationships, causal chains, how decisions play out over time.</p><p>OpenAI and Anthropic are approaching this from text, trying to infer physical and causal structure from written descriptions. Google has that structure already embedded in its data, accumulated over decades of building products that operated in the physical world. These are different starting points with different ceilings.</p><p>Here is where the argument needs to be tested.</p><p>Google is not moving fast enough on agents.</p><p>Anthropic has built Claude Code, and Cowork. Just this week, they announced that Claude now has the capability to remotely trigger tasks on a computer.</p><p>OpenClaw took the internet by storm and got acquired by OpenAI.</p><p>Perplexity launched Perplexity Computer.</p><p>These are small, fast teams shipping genuine agent products at a pace that Google has not matched. Gemini is technically capable, but Google has not yet translated its structural advantages into a category-defining agent product.</p><p>While chat has low switching costs, for agents, they could be significantly high due to the complexity that they deal with. Google needs to translate this structural advantage into embedding users deeply into agentic products and workflows. And it needs to do it quickly.</p>]]></content:encoded></item><item><title><![CDATA[Why Google might win? (2/5)]]></title><description><![CDATA[On infrastructure]]></description><link>https://abstractions.rakshitranjan.com/p/why-google-might-win-25</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/why-google-might-win-25</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Wed, 04 Mar 2026 18:21:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gUW3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9670a20-4b7b-4879-b018-dcb679a5b59e_608x608.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In <a href="https://abstractions.rakshitranjan.com/p/why-google-might-win">part 1</a>, I wrote about how Google makes its own AI chips cheaper than anyone can buy Nvidia&#8217;s. That&#8217;s a real advantage, but chips are only part of the story. A chip sitting in a poorly-cooled data centre, drawing expensive power, transmitting data across rented pipes, is not cheap. The TPU advantage only compounds if the infrastructure around it is efficient.</p><p>I&#8217;m calling this the infrastructure layer. It has five components: grid access, power reliability, power efficiency, intra-data centre networking, and inter-data centre networking. On most of these, Google is strong&#8211;but not uniquely so. Other hyperscalers secured grid connections decades ago. Everyone is diversifying into nuclear. Meta is building subsea cables at a pace that rivals Google&#8217;s.</p><p>One metric, though, tells a different story.</p><p>When a data centre buys $100 of electricity, not all of it reaches the processors. A large chunk gets burned on cooling, power conversion, and overhead. The ratio of total power consumed to power that actually does useful computation is called PUE. Lower is better. A perfect score is 1.0.</p><p>The industry average is 1.56. For every $100 of electricity, $36 is wasted.</p><p>Google&#8217;s number is 1.09. Microsoft is 1.16. Amazon is 1.15.</p><p>And then there&#8217;s Meta, at 1.08&#8211;slightly better than Google.</p><p>That last number is important because it tells you why this efficiency exists. It isn&#8217;t a Google-specific secret. Meta got there the same way Google did. Both companies built their data centres to run their own products&#8211;Google for Search and YouTube, Meta for Facebook and Instagram. When you own both the hardware and the workload, you can design them as one integrated system. Cooling, power delivery, and server layout are co-optimised for a specific job. </p><p>Microsoft and Amazon built for a different purpose: selling cloud infrastructure to thousands of customers with wildly different workloads. That flexibility comes at a thermodynamic cost. Generalised infrastructure is inherently less efficient than purpose-built infrastructure.</p><p>So if Meta matches Google on efficiency, why does Google still win this layer?</p><p>Because efficiency without distribution is just a cost saving. Meta runs some of the most efficient data centres in the world, and cannot sell a single dollar of that advantage to an external customer. It has no cloud business. The efficiency stays locked inside Facebook and Instagram, unable to compound into a higher-margin revenue stream.</p><p>Google runs the same efficient infrastructure and sells it. Every marginal efficiency gain flows through to Google Cloud&#8217;s pricing and then to Google&#8217;s margins. Stack this on the chip argument from Part 1: Google pays less per chip, then pays less to power each chip. At roughly 50% gross margins for AI&#8211;considerably thinner than Search&#8217;s 89%, being structurally cheaper at both layers is the difference between a sustainable business and a money-losing one. It is also why <a href="https://docsbot.ai/models/compare/gpt-4o/gemini-2-0-flash">Gemini&#8217;s API pricing is cheaper than OpenAI&#8217;s</a> for comparable models.</p><p>No single metric gives Google a decisive moat. But completeness does.</p><p>Google designs its own chips. It runs tier-one efficient data centres. It operates a private global network. It builds a frontier model. And it sells all of this as a cloud service.</p><p>Now look at everyone else.</p><ul><li><p>Meta matches Google on efficiency and networking but doesn&#8217;t sell cloud and its models are open-sourced rather than commercially hosted.</p></li><li><p>Microsoft sells cloud at enormous scale but doesn&#8217;t make its own chips, runs less efficient data centres, and licenses its model from OpenAI without controlling the research roadmap.</p></li><li><p>Amazon sells cloud and is investing in chips (Trainium) and cables (Fastnet), but both are early, and it has no frontier model.</p></li><li><p>OpenAI and Anthropic build the models everyone talks about. But they own nothing underneath. No data centres, no chips, no cables. Every token they serve includes a landlord&#8217;s tax to whichever hyperscaler hosts them. OpenAI pays Microsoft. Anthropic pays Google and Amazon.</p></li></ul><p>Dario Amodei made this fragility vivid in a <a href="https://www.dwarkesh.com/p/dario-amodei-2">recent conversation with Dwarkesh Patel</a>. Dwarkesh pushed him hard: if you really believe we&#8217;re a few years from having &#8220;a country of geniuses in a data centre,&#8221; why aren&#8217;t you buying far more compute?</p><p>Dario&#8217;s answer was striking. He explained that Anthropic&#8217;s revenue has been growing roughly 10x per year&#8211;from zero to $100 million in 2023, to $1 billion in 2024, to $9&#8211;10 billion in 2025. So he sat down and did the math. If that 10x rate continues, revenue would be $100 billion by end of 2026 and $1 trillion by end of 2027. He could, in theory, buy $1 trillion of compute starting in 2027 to keep pace with that demand.</p><p>But data centre commitments are made years in advance. If the revenue curve is even slightly off, if growth is 5x a year instead of 10x, or if the &#8220;country of geniuses&#8221; arrives in mid-2028 instead of mid-2027&#8211;Anthropic is stuck paying for compute it can&#8217;t monetise. As Dario put it: &#8220;if you&#8217;re off by only a year, you destroy yourselves.&#8221;</p><p>This is the existential arithmetic of being a tenant. The numbers confirm it. Anthropic projected roughly 40% gross margins in 2025&#8211;sixty cents of every revenue dollar going straight to compute costs. It burned $5.6 billion in cash in 2024, expects to burn another $3 billion in 2025, and plans to spend roughly $80 billion on cloud infrastructure through 2029. Break-even isn&#8217;t projected until 2028. The revenue is real. The product is excellent. But the company is running on investor confidence that the revenue curve outpaces the cost curve before the money runs out.</p><p>Google&#8217;s position is structurally different&#8211;not risk-free, but categorically different. It plans to spend $175&#8211;185 billion on infrastructure in 2026, nearly double the $91 billion it spent in 2025. If AI demand doesn&#8217;t materialise as expected, its margins compress and shareholders are unhappy. What doesn&#8217;t happen is bankruptcy. </p><p>Alphabet earned over $130 billion in net income in 2025 on $400 billion in revenue. Search, YouTube, and Cloud keep generating cash regardless of whether frontier AI justifies its training costs on schedule. Google can be wrong by a year, or two, and come out bruised but alive. Anthropic cannot.</p><p>If scaling laws hold&#8211;if more compute continues to produce better models, then the company that trains cheapest trains most. Owning every layer means every efficiency gain compounds with no intermediary taking a cut. There&#8217;s no landlord. There&#8217;s no margin leakage. That&#8217;s a structural advantage that can&#8217;t be closed by a funding round.</p><p>And right now, only Google has no gaps in it.</p>]]></content:encoded></item><item><title><![CDATA[Why Google might win? (1/5)]]></title><description><![CDATA[On cheap chips]]></description><link>https://abstractions.rakshitranjan.com/p/why-google-might-win</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/why-google-might-win</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Tue, 17 Feb 2026 19:11:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!IyEO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In <a href="https://abstractions.rakshitranjan.com/p/the-capital-theory-of-ai">The Capital Theory of AI</a>, I argued that the AI progress is a capital allocation problem. Scale purchases capability. The winners will be those with massive capital to meet the compute scaling demands, or companies that sit on specific data moats.</p><p>While listening to <a href="https://www.acquired.fm/episodes/google-the-ai-company">this</a> amazing Acquired episode on &#8220;Google: The AI Company&#8221;, I realised that my framing was a slight oversimplification of what the AI value chain looks like.</p><p>My revised hypothesis is that it is a 5-layer structure:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IyEO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IyEO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png 424w, https://substackcdn.com/image/fetch/$s_!IyEO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png 848w, https://substackcdn.com/image/fetch/$s_!IyEO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png 1272w, https://substackcdn.com/image/fetch/$s_!IyEO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IyEO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png" width="468" height="318.87224669603523" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:928,&quot;width&quot;:1362,&quot;resizeWidth&quot;:468,&quot;bytes&quot;:214192,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://abstractions.rakshitranjan.com/i/188293768?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IyEO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png 424w, https://substackcdn.com/image/fetch/$s_!IyEO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png 848w, https://substackcdn.com/image/fetch/$s_!IyEO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png 1272w, https://substackcdn.com/image/fetch/$s_!IyEO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77bfbd9d-de77-494a-a687-c178588ef5e2_1362x928.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Google has strengths across all 5 layers, and that might be why it wins the AI race in the short/medium-term<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>.</p><ol><li><p><strong>Silicon</strong>: custom TPU (tensor processing unit) chips</p></li><li><p><strong>Infrastructure</strong>: Private fibre and cloud connecting those chips</p></li><li><p><strong>Data</strong>: YouTube, Search, Maps, Scholar</p></li><li><p><strong>Model</strong>: Gemini, built on cost-effective self-owned hardware</p></li><li><p><strong>Distribution</strong>: Consumer apps spanning almost all touchpoints of human experience</p></li></ol><p>The strategic logic is different at every level. This series examines each layer. I&#8217;m starting with silicon because it sits at the bottom of the stack, and because it&#8217;s where the most underappreciated battle is playing out.</p><div><hr></div><p>For decades, AI ran on CPUs&#8211;general-purpose processors that handle instructions one at a time.</p><p>It was only 2012 when Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton figured that the GPUs (traditionally used for rendering video game graphics) could be leveraged for training image recognition models due to the similarities in inherent computation (both required matrix multiplications).</p><p>Today, every major AI lab runs on Nvidia GPUs. OpenAI trains on Microsoft Azure&#8217;s Nvidia clusters. Meta assembled over 350,000 H100 GPUs to train Llama 4. xAI built the Colossus cluster with 200,000+ H100 and H200 GPUs.</p><p>Google took a different path. As early as 2013, they realised that given the scale of their consumer apps, relying solely on Nvidia could potentially mean buying their entire production. So, they started building their own processors&#8211;TPUs.</p><p>TPUs are purpose-built for AI math, which makes them more efficient but less flexible than Nvidia&#8217;s GPUs. This enables a unique margin structure.</p><p>Nvidia designs and sells complete chips. It controls the full stack from architecture to software ecosystem (CUDA), and prices accordingly (leading to roughly <a href="https://nvidianews.nvidia.com/news/nvidia-announces-financial-results-for-fourth-quarter-and-fiscal-2025">80% gross margins</a>).</p><p>Google&#8217;s arrangement is different. It designs the chip logic&#8211;what the chip should do, how data flows through it. Broadcom then takes that design and turns it into something that can actually be fabricated&#8211;physical transistor layouts, memory interfaces, power delivery. TSMC manufactures the final product.</p><p>Google pays Broadcom a ~50% margin for this value-added. Still significant, but it is reasonably lower than what Nvidia charges for the whole package.</p><p>When chips represent over half of data centre cost, that margin gap is a structural cost advantage on every token that Google serves. Consequently, <a href="https://intuitionlabs.ai/articles/llm-api-pricing-comparison-2025">Gemini&#8217;s API pricing is cheaper than OpenAI&#8217;s equivalent models</a>.</p><p>The clearest validation of this cost advantage comes from an unlikely source. Last October, Anthropic&#8211;arguably Google&#8217;s most direct competitor in frontier models&#8211;signed a <a href="https://www.anthropic.com/news/expanding-our-use-of-google-cloud-tpus-and-services">cloud deal</a> for up to one million TPU chips worth tens of billions of dollars. This is a commercial agreement: Anthropic pays for TPU compute through Google Cloud. It is separate from Google&#8217;s $2 billion equity investment in the company.</p><p>This is why silicon is the foundation of Google&#8217;s AI stack. AI runs at roughly 50% gross margins, considerably lower than Search&#8217;s 89%. At those margins, having low-cost compute is almost existential. And right now, no one produces it cheaper than Google.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Winning the AI race in itself doesn&#8217;t guarantee profits or long-term market dominance. The current ways of experiencing AI are at odds with Google&#8217;s dominant search business&#8211;it remains to be seen how things progress, and how Google adjusts/makes up for the alternative reality.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Structural advantage doesn&#8217;t guarantee dominance. The biggest counterforces that Google faces are organisational drag, the innovator&#8217;s dilemma, and the risk that AI would never monetise as effectively as Search. But Google has navigated well through this over the last couple of years. Pichai merged DeepMind and Google Brain&#8211;a seemingly difficult task given the circumstances of the DeepMind acquisition. And Sergey Brin returned to work on Gemini.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Vibe coding a visualisation for historical AQI trends]]></title><description><![CDATA[A research experiment with Claude Code and Codex]]></description><link>https://abstractions.rakshitranjan.com/p/vibe-coding-a-visualisation-for-historical</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/vibe-coding-a-visualisation-for-historical</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Sun, 08 Feb 2026 17:44:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gUW3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9670a20-4b7b-4879-b018-dcb679a5b59e_608x608.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2><strong>The premise</strong></h2><p>I have been a long-time admirer of <a href="https://ourworldindata.org/">Our World in Data</a> and the book <a href="https://www.amazon.in/FACTFULNESS-REASONS-WRONG-THINGS-BETTER/dp/1473637465">Factfulness</a>. While mainstream media optimises for clickbait that showcases the direness of our existence, these two make a data-driven case for how this is the best time to be alive.</p><p>I love Delhi but am disappointed every year when we don&#8217;t take any measurable steps to curb the menace of air pollution.</p><p>As a pragmatic optimist, I wanted to study how air pollution has trended historically across major cities of the world. And more importantly, what they did to fix it.</p><p>So I vibe coded a <a href="https://aqi-tracker-lilac.vercel.app/">historical AQI tracker</a> for 8 cities (across 6 pollutants) that have successfully combated air pollution. It shows how the concentration of pollutants has trended over the last several years and also plots key policy interventions that potentially drove the improvement.</p><p>This research served two purposes:</p><ol><li><p>It gave me hope that things will get better</p></li><li><p>The entire building experience was super-fun</p></li></ol><div><hr></div><h2><strong>Notes on building</strong></h2><p>Link to <a href="https://github.com/rakshran/aqi-tracker">GitHub</a></p><ol><li><p>I wanted to build a data journalism-style site that tracks historical air quality across 8 cities, 6 pollutants, and up to 70 years of data for some cities.</p></li><li><p>I vibe coded the entire thing in ~4 days with Claude Code and Codex. This was my first time using agentic AI tools to build and deploy a site end-to-end.</p></li><li><p>I have always been an amateur coder and my side projects in the pre-LLM era relied heavily on cloning GitHub repos, copying and editing source code from inspiration websites and scouring through Stack Overflow for solving maddening bugs. Most of the issues that I faced then were more to do with plumbing than the actual coding logic. Installing the right packages, structuring files correctly, and deployment overheads.</p></li><li><p>Claude Code and Codex, with their repo-level access, make this a breeze. They helped me focus more on the problem solving and the experience.</p></li><li><p>Claude Code is the better architect. It is great at taking a vision and fleshing it out. It works amazingly well in plan mode to create an elaborate context document, chart out the steps that it will take, and makes great architecture and tooling decisions.</p></li><li><p>The only constraint that I have faced with Claude Code is that it has a fairly small limit on the Pro plan, which makes it hard to work in long sessions on a single problem.</p></li><li><p>Codex came in handy when Claude Code hit its limit. Overall, I found it to be the better machinist. Precise, fast, and reliable for isolated fixes like tooltip clipping or restructuring a component without breaking six other things.</p></li><li><p>The stack is minimal. React 18, Vite, Tailwind, Recharts. Deployed on Vercel.</p></li><li><p>The design language is editorial, not dashboard. I took inspiration from Our World in Data. Everything fits in one viewport. Seeing the chart and the interventions at the same time helps establish the relationship between policy and possible outcome.</p></li><li><p>The project has 96 commits. I didn&#8217;t write a single line of code myself.</p><ul><li><p>33 of the 65 commits are authored by Claude Code. The rest were through Codex.</p></li><li><p>I was writing prompts, reviewing pull requests, pushing back on implementation choices, and deciding what &#8220;done&#8221; looked like. Claude and Codex were writing React components, building data transparency features, and fixing mobile bugs.</p></li></ul></li><li><p>The most useful thing I did was ask Claude to write a brutal critique of my own site. What would a researcher say? What would a data journalist say? What would Twitter say? The critique helped improve the site significantly. Some examples:</p><ul><li><p>Inadequate citations.</p></li><li><p>Implied causation. Markers on intervention years strongly suggest those policies caused the improvements. But air quality is multi-factorial.</p></li><li><p>Survivorship bias. I&#8217;d only picked cities where things got better.</p></li><li><p>Geographic bias. Zero cities from Africa, South America, or the Middle East.</p></li><li><p>Interpolated data displayed identically to measured data.</p></li></ul></li><li><p>Claude then built every fix into the site.</p><ul><li><p>Proper source attribution so every number is traceable.</p></li><li><p>Causation-vs-correlation disclaimers in the sidebar.</p></li><li><p>A selection bias disclosure that tells you upfront: this dataset is not representative. It lists what&#8217;s missing and warns about success-oriented bias.</p></li><li><p>Interpolated data now renders as hollow circles with dashed lines and an &#8220;Est.&#8221; badge in tooltips.</p></li></ul></li><li><p>Desktop site was fairly straightforward. But perfecting the mobile experience took a fair bit of work.</p><ul><li><p>I wrote a full CLAUDE.md spec with breakpoint architecture, desktop protection rules, and a regression checklist. I had to explicitly instruct it to consider the desktop to be frozen and focus on mobile-only.</p></li><li><p>31 pull requests later, including multiple reverts and redos, the mobile experience works.</p></li></ul></li><li><p>The experience reinforced what I wrote in <a href="https://abstractions.rakshitranjan.com/p/how-to-ask">How to ask?</a>. Now the constraint is specification and taste. Can you describe what you want with enough precision? Can you tell when the output is wrong, not syntactically, but conceptually? These are fundamentally different skills from writing code, and they draw on a different kind of expertise.</p></li></ol>]]></content:encoded></item><item><title><![CDATA[The capital theory of AI]]></title><description><![CDATA[On scaling laws and its implications]]></description><link>https://abstractions.rakshitranjan.com/p/the-capital-theory-of-ai</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/the-capital-theory-of-ai</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Sun, 01 Feb 2026 17:41:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gUW3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9670a20-4b7b-4879-b018-dcb679a5b59e_608x608.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Computer scientist Richard Sutton hypothesised the &#8220;<strong><a href="https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson.pdf">bitter truth</a></strong>&#8220; in his 2019 essay:</p><blockquote><p><em>The bitter lesson is based on the historical observations that 1) AI researchers have often tried to build knowledge into their agents, 2) this always helps in the short term, and is personally satisfying to the researcher, but 3) in the long run it plateaus and even inhibits further progress, and 4) breakthrough progress eventually arrives by an opposing approach based on scaling computation by search and learning. The eventual success is tinged with bitterness, and often incompletely digested, because it is success over a favored, human-centric approach.</em></p></blockquote><p>This was a hard-to-accept conclusion based on years of empirical evidence. Researchers have been trying to invent novel ways of embedding intelligence into machines, but these efforts have been consistently surpassed by advancements in compute tech (lower costs &amp; enhanced capabilities).</p><p>In the following year (2020), OpenAI published a paper titled &#8220;<strong><a href="https://arxiv.org/abs/2001.08361">Scaling Laws for Neural Language Models</a></strong>&#8220;, which further crystallised Sutton&#8217;s hypothesis.</p><p>The core insight was that AI performance improves with mathematical predictability, following power laws, as you scale <strong>compute, data, and model parameters</strong>. This proves that raw scale is a far more powerful driver of intelligence than specific architectural design.</p><p>Why does this matter?</p><p>It matters because technological progress is seldom predictable. The evolution of the internet, the advent of mobile, and the shift to cloud were all characterised by huge unpredictability.</p><p>Scaling laws, however, change this equation. If you know that a 10x increase in compute yields a reliable improvement in capability, then the frontier of AI is not a research problem; it&#8217;s a capital allocation problem.</p><p>The implications ripple outward from this core insight.</p><p>Consider the competitive dynamics among AI labs. If capability is purchased with compute, then the labs that can raise the most capital and secure the most chips will produce the best models. This is not a cottage industry for brilliant researchers in garages; it is a capital-intensive manufacturing process with massive economies of scale.</p><p>This is a major shift. For the past two decades, the defining logic of the technology industry has been software-centric. Software scales infinitely at zero marginal cost, which is why the aggregators&#8211;Google, Meta, Amazon&#8211;could build global monopolies with relatively modest infrastructure investments. The valuable asset was the code, the algorithm, the network effect.</p><div><hr></div><p>This doesn&#8217;t mean that there is no value to be captured for other players. There are, in fact, several layers in this emerging stack:</p><p><strong>At the bottom: compute infrastructure.</strong> Nvidia (and by extension TSMC, ASML Holdings, etc.) is the obvious winner here. Their GPUs are the de facto platform for AI training. But beyond Nvidia, the hyperscalers (Microsoft, Google, Amazon) are investing tens of billions of dollars annually in data centers.</p><p><strong>In the middle: foundation model providers.</strong> OpenAI, Anthropic, Google DeepMind, Meta&#8217;s AI labs&#8211;these are the entities actually training frontier models. Their advantage is the combination of talent, capital, and proprietary training techniques. But here&#8217;s the uncomfortable reality: if scaling laws continue to hold, the moat is not the model itself (which can be approximated by anyone with enough compute) but the rate of improvement.</p><p>There is no switching or multi-homing cost, and hence it is very hard for these model providers to build a defensible moat beyond being ahead of the curve on capabilities and perhaps experimenting with interesting subscription models (e.g., ChatGPT Go, Google&#8217;s bundling &amp; Perplexity&#8217;s Airtel Partnership).</p><p><strong>At the top: applications and integrations.</strong> This is where most businesses will play, and owning proprietary data, specific workflows, or customer relationships can be the key to differentiation.</p><p>The a16z podcast &#8220;<strong><a href="https://youtu.be/3XVDtPU8xKE?si=HqXHQZh1z9VcfI_4">The AI opportunity that goes beyond models</a></strong>&#8220; highlights three amazing examples of startups doing this:</p><ol><li><p><strong>Slingshot AI</strong>: They build an AI mental health companion. They started by taking notes for human therapists&#8212;a classic way to &#8220;do things that don&#8217;t scale.&#8221; By helping with the paperwork, they saw thousands of private sessions that aren&#8217;t on the internet. This gave them the &#8220;shadow data&#8221; needed to train a model that actually understands clinical empathy, which a general AI simply can&#8217;t learn from Reddit.</p></li><li><p><strong>FamilySearch:</strong> They run the world&#8217;s largest genealogy database. Their moat is physical friction: for decades, they sent people to remote villages to microfilm paper birth and death records. You can&#8217;t scrape what isn&#8217;t digital. They own the data because they were willing to do the tedious, manual work that Silicon Valley usually tries to avoid.</p></li><li><p><strong>vLex:</strong> This is a legal research platform that digitised fragmented, analog law records in Spain and beyond. By turning messy, offline paper trails into a &#8220;system of record,&#8221; they built a walled garden. Even the most powerful AI is useless if it can&#8217;t access the source of truth, and in this niche, vLex owns the keys to the library.</p></li></ol><div><hr></div><p>It would be a mistake to treat scaling laws as guaranteed to continue indefinitely. There are legitimate reasons for caution.</p><p>First, there is the data wall. As we exhaust the &#8220;low-hanging fruit&#8221; of high-quality human-generated internet text, the industry is forced into the expensive, artisanal world of synthetic data and human-in-the-loop labelling. There are early indications of performance degradation for models trained on outputs of other AIs.</p><p>Second, there is the distinction between loss (the statistical measure that scaling laws predict) and capability (the practical abilities we actually care about). Loss decreases smoothly, but capabilities often emerge abruptly and unpredictably. A model might go from 0% to 90% accuracy on a task with a single increment of scale. We cannot yet predict which capabilities will emerge when.</p><p>Third, AI follows power laws, where linear gains require exponential costs. We&#8217;ve reached a point where the next 1% of intelligence requires billions in hardware and massive energy, making &#8220;brute force&#8221; economically unsustainable for most.</p><div><hr></div><p>That being said, we are still far off from reaching that scaling wall. As Tomasz Tunguz writes in &#8220;<strong><a href="https://tomtunguz.com/gemini-3-proves-pretraining-scaling-laws-intact/">The scaling wall was a mirage</a></strong>&#8220;:</p><blockquote><p><em>Then Gemini 3 launched. The model has the same parameter count as Gemini 2.5, one trillion parameters, yet achieved massive performance improvements. It&#8217;s the first model to break 1500 Elo on LMArena &amp; beat GPT-5.1 on 19 of 20 benchmarks.</em></p><p><em><a href="https://x.com/OriolVinyalsML/status/1990854455802343680">Oriol Vinyals</a>, VP of Research at Google DeepMind, credited improving pre-training &amp; post-training for the gains. He continued that the delta between 2.5 &amp; 3.0 is as big as Google has ever seen with no walls in sight.</em></p><p><em>This is the strongest evidence since o1 that pre-training scaling still works when algorithmic improvements meet better compute.</em></p></blockquote><p></p><blockquote><p><em>Second, Nvidia&#8217;s earnings call reinforced the demand.</em></p><p><em>&#8220;We currently have visibility to $0.5 trillion in Blackwell and Rubin revenue from the start of this year through the end of calendar year 2026. By executing our annual product cadence and extending our performance leadership through full stack design, we believe NVIDIA will be the superior choice for the $3 trillion to $4 trillion in annual AI infrastructure build we estimate by the end of the decade.&#8221;</em></p><p><em>&#8220;The clouds are sold out and our GPU installed base, both new and previous generations, including Blackwell, Hopper and Ampere is fully utilized. Record Q3 data center revenue of $51 billion increased 66% year-over-year, a significant feat at our scale.&#8221;</em></p></blockquote><div><hr></div><p>This brings us back to Sutton&#8217;s bitter lesson. The evidence suggests we haven&#8217;t yet exhausted what scale can buy. But &#8220;not yet&#8221; is doing a lot of work in that sentence. The economics are getting harder. The next order of magnitude in compute will cost tens of billions of dollars. At some point, the capital allocation problem becomes a capital availability problem. Until then, the race continues. The winners will be those who can write the biggest checks, or those clever enough to build moats that don&#8217;t depend on writing checks at all.</p>]]></content:encoded></item><item><title><![CDATA[How to feature?]]></title><description><![CDATA[On generative engine optimisation]]></description><link>https://abstractions.rakshitranjan.com/p/how-to-feature</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/how-to-feature</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Mon, 26 Jan 2026 18:38:26 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gUW3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9670a20-4b7b-4879-b018-dcb679a5b59e_608x608.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last week, I watched the <a href="https://vryse.co/">Vryse</a> pitch on Shark Tank. It is a search engine optimisation (SEO) and generative engine optimisation (GEO) platform.</p><p>SEO concerns itself with reverse engineering Google&#8217;s search algorithm&#8211;optimising for keywords, backlinks, page structure, etc. to rank higher in search results. GEO is the natural corollary&#8211;optimising to be surfaced in LLM responses.</p><p>Fundamentally, LLM-search has two layers:</p><p>The first is pre-trained knowledge, which refers to what the model learned during training. This is a snapshot of the internet at a fixed point in time, baked into the model&#8217;s weights. If you&#8217;re a brand trying to influence this, you&#8217;re essentially trying to retroactively change history. Short of becoming Wikipedia-level ubiquitous, there&#8217;s not much you can do here.</p><p>The second is RAG&#8211;Retrieval Augmented Generation. When you ask Claude or ChatGPT a question that requires current information, the model calls a search API (Google, Bing), retrieves the top results, reads them, and synthesizes an answer. This is where optimization is actually possible.</p><p>The RAG layer still runs on the existing search infrastructure. The OpenAI web search API documentation shows that models rely on the top 10 or 20 results from traditional search engines. If you aren&#8217;t on that first page, you don&#8217;t exist in the model&#8217;s context window.</p><p>This got me thinking that if GEO for pre-trained knowledge is a black box, and GEO for RAG mostly reduces to &#8220;do good SEO,&#8221; then what&#8217;s the extent of value-add of a dedicated GEO platform?</p><p>There is not a lot of research on GEO, but a few emerging hypotheses are as follows:</p><ol><li><p>Reasoning models can go deeper into fewer pages, not wider across more pages. If your website ranks among the top few search results, the quality and structure of your content matter more than ever.</p></li><li><p>GEO rewards meaning density and semantic proximity (related concepts clustered together).</p></li><li><p>While average Google search queries are 4 words<sup>[1]</sup>, an average generative query is 23 words. There is a behavioural shift from searching to consulting in the context of LLMs. GEO implies optimising to answer complex, conversational questions.</p></li><li><p>Authority is verified through presence across key signals&#8211;communities (Reddit, Quora), YouTube transcripts, and citation-worthy journals/publications. Absence from the participatory web is treated as low-trust.</p></li><li><p>Attribution is hard. Google&#8217;s ad-supported model incentivised moving users to third-party sites to sell more ads (measurable through click-through rate). LLMs are subscription-driven. They have no financial incentive to send users away (at the time of writing this); their value lies in providing a frictionless, all-in-one answer.</p></li></ol><p>With GEO, the technical requirements of search are merging with the structural requirements of good writing. It would be interesting to see how this evolves and if we are able to deterministically optimise LLM responses just as we did with search.</p><div><hr></div><ol><li><p><a href="https://www.semrush.com/blog/chatgpt-search-insights/">Investigating ChatGPT Search: Insights from 80 Million Clickstream Records</a>&#8617;&#65038;</p></li></ol>]]></content:encoded></item><item><title><![CDATA[Attention is all you need]]></title><description><![CDATA[On the paper that started it all]]></description><link>https://abstractions.rakshitranjan.com/p/attention-is-all-you-need</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/attention-is-all-you-need</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Sun, 18 Jan 2026 18:32:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gUW3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9670a20-4b7b-4879-b018-dcb679a5b59e_608x608.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This weekend, I spent time reading Google&#8217;s now-seminal paper, <em><a href="https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf">Attention Is All You Need</a></em>, which introduced the transformer architecture that forms the foundation of large language models like ChatGPT, Claude, and Gemini.</p><p>I&#8217;ve forgotten most of the math I studied during my engineering, which made it difficult to fully understand the paper. However, I was able to build some intuition. Here are my abstracted notes:</p><ol><li><p>Prior to transformers, techniques like RNNs (Recurrent Neural Networks) were used to train models:</p><ul><li><p>RNNs process sequences one step at a time. If you&#8217;re reading a sentence, you read word 1, then word 2, then word 3. Each step depends on the previous one.</p></li><li><p>This made the process computationally slow&#8212;computing step 5 requires completing step 4, which requires step 3, and so on.</p></li><li><p>Since information has to travel far, RNNs start to fail on large contexts&#8212;just as information gets lost or modified in a long chain of Chinese whispers.</p></li></ul></li><li><p>The researchers proposed a radical new approach: instead of processing words sequentially, compute the relationships between all words in a sequence simultaneously.</p></li><li><p>There&#8217;s a bunch of math behind this. The paper formalises it using three concepts:</p><ul><li><p><strong>Query</strong>: what I&#8217;m currently looking for</p></li><li><p><strong>Key</strong>: what each word offers</p></li><li><p><strong>Value</strong>: the information each word contains</p></li></ul></li><li><p>These are used to calculate &#8220;attention.&#8221; Attention is an elegant way of assigning weights to words based on their relationship with other words. For example, in the sentence <em>&#8220;The dog sat on the chair as it was tired&#8221;</em>, the calculation will enrich the word &#8220;it&#8221; with context such that its relationship with &#8220;dog&#8221; has a higher weight than its relationship with &#8220;chair.&#8221;</p></li><li><p>Attention is content-aware but position-blind, which can be problematic. Without positional information, a transformer wouldn&#8217;t know which noun is the subject and which is the object&#8212;&#8221;man eats fish&#8221; and &#8220;fish eats man&#8221; would be indistinguishable. To solve this, the researchers added a positional encoding to each word before feeding it into the transformer. This preserves word order and meaning.</p></li><li><p>The paper had several implications:</p><ul><li><p>Training time compressed dramatically. The researchers achieved far superior results on a translation task in 3.5 days using 8 GPUs&#8212;at a fraction of the cost of the best models at that time.</p></li><li><p>Due to the parallel nature of computation, throwing more compute at the problem works well. Capital became a moat, which is why we&#8217;re seeing massive investments in GPU acquisition for training.</p></li><li><p>Transformers, due to their architecture, are extensible to multimodal inputs (voice, text, images) and diverse tasks (summarisation, text generation, translation).</p></li></ul></li></ol><p>Transformers made intelligence a scaling problem. Everything since has been a consequence of that.</p><div><hr></div><p></p>]]></content:encoded></item><item><title><![CDATA[How to ask?]]></title><description><![CDATA[On LLM prompting and getting the best out of agentic platforms]]></description><link>https://abstractions.rakshitranjan.com/p/how-to-ask</link><guid isPermaLink="false">https://abstractions.rakshitranjan.com/p/how-to-ask</guid><dc:creator><![CDATA[Rakshit Ranjan]]></dc:creator><pubDate>Sun, 11 Jan 2026 18:17:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gUW3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9670a20-4b7b-4879-b018-dcb679a5b59e_608x608.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I spent the weekend using Claude Code to reorganise my notes and update my website. Through the process, I realised that working with agentic AI tools (and, by extension, chat-based LLMs) is akin to working with engineers in a product org.</p><p>I mean the actual mechanics: writing up what I wanted before asking for it, being specific about edge cases, reviewing the output, and pushing back. The stuff that separates good product managers from bad ones.</p><p>This surprised me.</p><p>Take planning. Most people just start typing what they want. Make me a website. Write me a story. Analyse this data. And then they&#8217;re disappointed when what comes back isn&#8217;t quite right.</p><p>Good PMs don&#8217;t work this way. Before they ask anyone to build anything, they&#8217;ve already written a document explaining what they want and why. They&#8217;ve thought through the edge cases. They&#8217;ve gathered examples of what good looks like. That is how you get what you actually want instead of &#8220;what you literally asked for&#8221;.</p><p>The same thing works with LLMs. Before I ask Claude to do something complex, I dump my raw thoughts into it and ask it to write a plan. This surfaces how it&#8217;s thinking about the problem. Often, it&#8217;s different from how I thought about it. Sometimes it&#8217;s better. Sometimes it&#8217;s worse. Either way, I can correct course before any real work happens.</p><p>Then there&#8217;s precision. Bad PMs say things like &#8220;use the customer info&#8221; and &#8220;make sure it works&#8221;. Good PMs reference specific database fields and define exact test cases. The gap between vague and specific is where most miscommunication lives.</p><p>LLMs are, if anything, more sensitive to this than humans. A person will ask clarifying questions. An LLM will just guess. And its guesses, while often plausible, are frequently wrong and misaligned with your requirements.</p><p>I&#8217;ve started breaking complex tasks into steps for this reason. When I was updating my personal website, I didn&#8217;t ask Claude Code to redesign the whole thing. I asked it to audit the current structure first. Then propose changes. Then update the navigation and so on.</p><p>The last piece is a review. Code gets reviewed before it ships. Product specs get reviewed before anyone builds them. But most people treat LLM output as final the moment it appears.</p><p>One trick that works is role switching. I have the LLM do the work as a &#8220;developer&#8221; or &#8220;analyst,&#8221; then evaluate its own output as a &#8220;senior engineer&#8221; or &#8220;manager.&#8221; This forces a change in perspective and catches real problems.</p><p>In the current scenario, I believe being AI-native has nothing to do with being technical. It is about developing the ability to figure out what you want, communicate it precisely, and verify that you got it.</p><p>It can be done by anyone who has spent years translating fuzzy information into concrete instructions. Engineers. Product managers. Editors. Consultants. Anyone.</p><p>I don&#8217;t know if this will last. Maybe LLMs will get good enough at reading intent that precision won&#8217;t matter. But right now, the bottleneck isn&#8217;t the model, it&#8217;s our ability to say what we mean.</p>]]></content:encoded></item></channel></rss>