<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Sameer Khan</title>
    <description>The latest articles on Forem by Sameer Khan (@monkfromearth).</description>
    <link>https://forem.com/monkfromearth</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F422077%2Fc59a6851-ab5b-4629-afc5-f46d01148b32.png</url>
      <title>Forem: Sameer Khan</title>
      <link>https://forem.com/monkfromearth</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/monkfromearth"/>
    <language>en</language>
    <item>
      <title>Nobody Trained GPT-5.5 to Hack. It Beat Human Cyber Experts Anyway.</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Fri, 01 May 2026 12:29:43 +0000</pubDate>
      <link>https://forem.com/monkfromearth/nobody-trained-gpt-55-to-hack-it-beat-human-cyber-experts-anyway-4e2a</link>
      <guid>https://forem.com/monkfromearth/nobody-trained-gpt-55-to-hack-it-beat-human-cyber-experts-anyway-4e2a</guid>
      <description>&lt;p&gt;&lt;strong&gt;Nobody trained GPT-5.5 to hack. They trained it to think, and the hacking fell out.&lt;/strong&gt; That is the only sentence in AISI's new evaluation that matters, and the only one most coverage will miss. OpenAI's GPT-5.5 just became the second AI to complete AISI's 32-step cyber range end-to-end.&lt;sup id="fnref1"&gt;1&lt;/sup&gt; Mythos Preview was the first, three weeks ago. Different lab, different architecture, similar score. The Mythos result wasn't an outlier. It was the first point on a curve.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TL;DR.&lt;/strong&gt; GPT-5.5 hit 71% on AISI's expert cyber tasks, edging out Mythos Preview's 68.6%, and completed The Last Ones (AISI's 32-step corporate network attack) in 2 of 10 attempts. AISI evaluated the base model, not a cyber-permissive variant. Their framing: cyber-offensive skill is emerging as a byproduct of reasoning, not a trained capability. &lt;strong&gt;Nobody trained these models to hack. They trained them to think. The hacking fell out.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What Did GPT-5.5 Score on AISI's Cyber Evaluation?
&lt;/h2&gt;

&lt;p&gt;71.4% on expert-level advanced tasks. Up from GPT-5.4's 52.4%. Up from Claude Opus 4.7's 48.6%. Slightly above Mythos Preview's 68.6%.&lt;/p&gt;

&lt;p&gt;The numbers in one place:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Model&lt;/th&gt;
&lt;th&gt;Expert-tier pass rate&lt;/th&gt;
&lt;th&gt;TLO completion&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;GPT-5.5&lt;/td&gt;
&lt;td&gt;71.4% (±8.0)&lt;/td&gt;
&lt;td&gt;2 of 10 attempts&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Mythos Preview&lt;/td&gt;
&lt;td&gt;68.6% (±8.7)&lt;/td&gt;
&lt;td&gt;3 of 10 attempts&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;GPT-5.4&lt;/td&gt;
&lt;td&gt;52.4% (±9.8)&lt;/td&gt;
&lt;td&gt;not reported&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Claude Opus 4.7&lt;/td&gt;
&lt;td&gt;48.6% (±10.0)&lt;/td&gt;
&lt;td&gt;not reported&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;These tasks aren't gentle. They cover memory corruption exploitation, breaking cryptographic implementations, and reverse engineering stripped binaries. Things that take experienced security researchers hours, sometimes days.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Who this displaces:&lt;/strong&gt; the bottom of the offensive-research market. Skilled red-teamers don't disappear, but the floor drops. Anything a junior could solve in a day, a model now solves in minutes, with the same answer at the end.&lt;/p&gt;

&lt;h2&gt;
  
  
  How Does GPT-5.5 Compare to Claude Mythos on the Same Cyber Range?
&lt;/h2&gt;

&lt;p&gt;Three weeks ago I wrote that &lt;a href="https://monkfrom.earth/blogs/claude-mythos-autonomous-cyberattack" rel="noopener noreferrer"&gt;Claude Mythos became the first AI to finish AISI's 32-step cyber range end-to-end&lt;/a&gt;. The framing then was natural: a single model, a single milestone, a one-off result that might not generalize.&lt;/p&gt;

&lt;p&gt;GPT-5.5 just generalized it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Same evaluation. Different lab. Different base architecture. Comparable score.&lt;/strong&gt; Mythos finished TLO in 3 of 10 attempts. GPT-5.5 finished it in 2 of 10. The variance is small. The trend is not.&lt;/p&gt;

&lt;p&gt;This is the part I missed in my first read. The Mythos post implicitly treated the result as something Anthropic shipped. &lt;strong&gt;AISI's view, which I now think is correct: this is something the field shipped.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What Does GPT-5.5 Reverse-Engineering a VM in 11 Minutes Tell Us?
&lt;/h2&gt;

&lt;p&gt;One challenge in the suite asked the model to reverse engineer a custom virtual machine. A human expert with professional tooling spent about 12 hours on it. GPT-5.5 finished in 10 minutes 22 seconds.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Roughly 70x faster than the human, on a task that does not yield to brute force.&lt;/strong&gt; Reverse engineering a custom VM is structural work: read instructions you have never seen, infer the semantics, build a mental model of a machine that nobody documented. It is the kind of task that has historically separated senior researchers from juniors.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The outcome is faster attackers, not cheaper ones.&lt;/strong&gt; They iterate more, try more targets, abandon dead ends sooner. The shape of an offensive workflow shifts from "pick one binary, commit a day" to "fan out across a portfolio in an afternoon."&lt;/p&gt;

&lt;h2&gt;
  
  
  Was GPT-5.5 Trained Specifically for Cyber Tasks?
&lt;/h2&gt;

&lt;p&gt;Not as far as the public record goes.&lt;/p&gt;

&lt;p&gt;OpenAI does ship cyber-permissive variants for vetted defenders through Trusted Access. The first was &lt;a href="https://monkfrom.earth/blogs/openai-gpt-5-4-cyber-trusted-access" rel="noopener noreferrer"&gt;GPT-5.4-Cyber&lt;/a&gt;. On the same day AISI published this evaluation, OpenAI also rolled out &lt;strong&gt;GPT-5.5-Cyber&lt;/strong&gt;, the next-generation permissive variant for critical infrastructure defenders.&lt;sup id="fnref2"&gt;2&lt;/sup&gt; Both are fine-tuned products gated behind identity verification.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;AISI did not test either variant. They tested base GPT-5.5, with no cyber-specific fine-tune.&lt;/strong&gt;&lt;sup id="fnref1"&gt;1&lt;/sup&gt; That distinction is the whole story.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The fine-tune is the policy on top, not the capability underneath.&lt;/strong&gt; The offensive capability lives in the base reasoning. Cyber-specific training adds permissions, not power.&lt;/p&gt;

&lt;p&gt;This is the strongest evidence yet that frontier offensive cyber is a side effect of general reasoning gains, not a separately trained skill. AISI states it directly: "if cyber-offensive skill is emerging as a byproduct... we should expect further increases in cyber capability from models in the near future, potentially in quick succession."&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwrygxmhsv7h4ynr5r0wi.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwrygxmhsv7h4ynr5r0wi.webp" alt="Illustration of a thinking figure beside a code window reading 'if model.thinks(): can.hack = True', with a locked safe in the background. The byproduct thesis as one image: train a model to reason, the offensive capability follows for free." width="800" height="422"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The honest counter: maybe both labs are quietly training cyber data into the base mix without naming it as such. Possible. &lt;strong&gt;But "quiet fine-tune" still produces a curve, not a one-off.&lt;/strong&gt; Whatever's in the base, it generalizes across two labs and two architectures within three weeks.&lt;/p&gt;

&lt;h2&gt;
  
  
  Did GPT-5.5's Cyber Performance Plateau on the Range?
&lt;/h2&gt;

&lt;p&gt;No. That's the second-most-load-bearing line in the report, and it came from inside OpenAI.&lt;/p&gt;

&lt;p&gt;Noam Brown noted on X: "After 100 million tokens, performance was still going up. What we're seeing here is not the capability ceiling."&lt;sup id="fnref3"&gt;3&lt;/sup&gt; AISI's own report uses similar language: performance scales with inference compute, no plateau observed at the top of the range.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The capability isn't capped by the model. It's capped by how much compute you spend.&lt;/strong&gt; That's a different shape of problem than "the model can do X but no more."&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmtfpijka5xjha7m7ufv8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmtfpijka5xjha7m7ufv8.png" alt="AISI evaluation chart showing average steps completed on The Last Ones cyber range against inference token spend. GPT-5.5 and Claude Mythos Preview keep climbing past 100M tokens with no plateau, while GPT-5.4 and Claude Opus 4.7 trail well below" width="800" height="488"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Where Did GPT-5.5 Fail in AISI's Cyber Evaluation?
&lt;/h2&gt;

&lt;p&gt;The Cooling Tower scenario, an industrial control system simulation with 7 steps. GPT-5.5 finished zero successful runs.&lt;sup id="fnref1"&gt;1&lt;/sup&gt; Industrial protocols are unfamiliar territory: different stack, different conventions, fewer training examples on the open internet.&lt;/p&gt;

&lt;p&gt;This is the steelman for the "not yet" reading. &lt;strong&gt;The byproduct effect doesn't generalize uniformly across every domain.&lt;/strong&gt; Web and binary tasks are well represented in training data. Industrial protocols are not.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The honest read is dual: corporate IT looks more exposed than it did three weeks ago. OT is still its own world.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  How Does GPT-5.5 Cyber Capability Change the Defender's Window?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The window that matters is the lag between when offense gets cheap and when defense catches up.&lt;/strong&gt; That's David Sacks's framing on X: AI cyber doesn't create new vulnerabilities, it discovers existing ones, and the equilibrium eventually settles between AI offense and AI defense.&lt;sup id="fnref4"&gt;4&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;OpenAI is already shipping defender tooling ahead of more capable models, with Codex Security and the Trusted Access program. Anthropic runs Project Glasswing on the same model that scored these benchmarks. &lt;strong&gt;Both labs see the same curve. Both are racing the defenders onto the same plane the attackers will eventually be on.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The thing they cannot influence is timing for everyone else. &lt;strong&gt;Sacks's line: all the frontier models, including those out of China, will be at this capability level within roughly six months.&lt;/strong&gt;&lt;sup id="fnref4"&gt;4&lt;/sup&gt; That's the planning horizon.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Should Security Teams Do About GPT-5.5 and the Models Coming Next?
&lt;/h2&gt;

&lt;p&gt;The same baseline that AISI keeps recommending: patch, MFA, logging, segmentation. Necessary, no longer sufficient.&lt;/p&gt;

&lt;p&gt;The new line item is treating AI-assisted offense as the default operating environment, not an emerging risk. That changes a few things in practice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Assume reverse-engineering is fast.&lt;/strong&gt; A binary you shipped this morning is now ~10 minutes of compute away from being read like source by anyone with API access.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Start using AI-assisted defense yourself.&lt;/strong&gt; Codex Security has been credited with over 3,000 critical and high vulnerability fixes since launch. The same models on offense are the ones on defense. Symmetry is the only realistic strategy.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Plan for the curve, not the model.&lt;/strong&gt; The next model will be more capable than GPT-5.5 or Mythos at this evaluation. Assume that and build for it.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;GPT-5.5 hit 71.4% on AISI's expert cyber tasks&lt;/strong&gt;, the highest score on record, slightly above Mythos Preview at 68.6%&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Second AI to finish AISI's 32-step cyber range end-to-end&lt;/strong&gt; (TLO) in 2 of 10 attempts; Mythos finished it in 3 of 10&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;One challenge took a human expert 12 hours; GPT-5.5 finished it in 10 minutes 22 seconds.&lt;/strong&gt; Roughly 70x faster, same correctness&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The model wasn't fine-tuned for cyber.&lt;/strong&gt; AISI evaluated base GPT-5.5, not the cyber-permissive variant. Capability emerged from general reasoning improvements&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;No plateau observed&lt;/strong&gt; at the top of the range; performance kept scaling past 100M inference tokens&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;GPT-5.5 failed industrial control (Cooling Tower)&lt;/strong&gt; with zero completions, showing the byproduct effect doesn't generalize evenly across domains&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Two labs, one month, same benchmark.&lt;/strong&gt; Mythos wasn't an outlier. It was the first point on a curve&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;I write about how AI safety and capability actually get built on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;. If this resonated, the shorter versions are there.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources
&lt;/h2&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://www.aisi.gov.uk/blog/our-evaluation-of-openais-gpt-5-5-cyber-capabilities" rel="noopener noreferrer"&gt;AISI: Our evaluation of OpenAI's GPT-5.5 cyber capabilities (April 30, 2026)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://openai.com/index/scaling-trusted-access-for-cyber-defense/" rel="noopener noreferrer"&gt;OpenAI: Trusted access for the next era of cyber defense (April 30, 2026)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;&lt;a href="https://x.com/polynoamial/status/2049883449327243413" rel="noopener noreferrer"&gt;Noam Brown (@polynoamial) on inference scaling, April 30, 2026&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://x.com/DavidSacks/status/2049907993588769006" rel="noopener noreferrer"&gt;David Sacks (@DavidSacks) on the AI offense-defense equilibrium, April 30, 2026&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>security</category>
      <category>cybersecurity</category>
      <category>agents</category>
    </item>
    <item>
      <title>Pure Software Is Uninvestable: Naval's Take</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Thu, 30 Apr 2026 11:28:12 +0000</pubDate>
      <link>https://forem.com/monkfromearth/pure-software-is-uninvestable-navals-take-54e2</link>
      <guid>https://forem.com/monkfromearth/pure-software-is-uninvestable-navals-take-54e2</guid>
      <description>&lt;p&gt;&lt;strong&gt;Naval Ravikant released "A Return to Code" on April 28 and dropped a line worth pausing on: pure software is uninvestable.&lt;/strong&gt;&lt;sup id="fnref1"&gt;1&lt;/sup&gt; He explains why from the capital side. He does not finish the thought from the builder side. That is where this post starts.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TL;DR.&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Naval says pure software is uninvestable because &lt;strong&gt;agents improve faster than any startup's lead.&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;He also calls Apple skipping AI &lt;strong&gt;"the biggest strategic mistake of the decade."&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;The builder reading is sharper: &lt;strong&gt;code went from edge to floor. The new edge is intent.&lt;/strong&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What Did Naval Say in "A Return to Code"?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Two claims: prototyping is now open to anyone, and the agents underneath any startup improve faster than its moat can.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Naval calls the new mode &lt;em&gt;vibe coding&lt;/em&gt;: describe what you want in English, get a working app back. He estimates it takes the share of people who could plausibly build apps "from like 0.1 percent to one or two or three percent" of the population.&lt;sup id="fnref1"&gt;1&lt;/sup&gt; Twenty to thirty times more builders, overnight.&lt;/p&gt;

&lt;p&gt;His investment claim is not the one most readers will assume. He is not saying agents will autonomously architect scaled systems within a year. He says the opposite about today's agents.&lt;/p&gt;

&lt;p&gt;They "get lost" past a certain context size. They "fix the same bug five times." They show "jagged intelligence." They are "easily led around" by whoever is steering them.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;So why is pure software uninvestable then? Because &lt;strong&gt;the agents themselves keep improving faster than any single startup's lead.&lt;/strong&gt; "If your whole advantage is, hey, I'm building cool software that other people don't know how to build, I think that's uninvestable."&lt;sup id="fnref1"&gt;1&lt;/sup&gt; The defensibility window shrinks under your feet. Capital should chase hardware, network effects, and AI models instead.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Is This the Same Naval Who Once Said Code Was Leverage?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Because leverage stops being leverage the moment everyone has it.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Years ago, Naval taught a generation of builders to think of code as leverage.&lt;sup id="fnref2"&gt;2&lt;/sup&gt; Code worked while you slept. It scaled to millions without permission. Zero marginal cost. A solo programmer with a laptop had the productivity of a small factory.&lt;/p&gt;

&lt;p&gt;That argument was right for its era. The world he is describing now is not a contradiction. It is the next era.&lt;/p&gt;

&lt;p&gt;When something becomes infinitely reproducible by anyone, including by an English-speaking model, it stops being leverage. It becomes a baseline.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The leverage Naval named in 2018 did not disappear. It diffused.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq2zm4fc6575a5w13z6ow.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq2zm4fc6575a5w13z6ow.png" alt="Naval's leverage ladder. The bottom rung, code, is no longer a step you climb. It is the floor you stand on." width="800" height="422"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What Did Code-as-Leverage Look Like in Practice?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;In 2018, having a technical co-founder was the tiebreaker. In 2026, the room does not ask.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I lived inside the old version of his argument. In 2018 I co-founded Spotwash, a vehicle rental and on-demand washing service. Government-incubated, some press, a small but real run on the early-stage circuit.&lt;/p&gt;

&lt;p&gt;What I remember most is what I did not say in any pitch deck. The question that opened doors was not "what does Spotwash do?" It was quieter: &lt;em&gt;who is going to write the code?&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Investor meetings, accelerator interviews, almost every room. Having a technical co-founder was a tiebreaker. An idea with a builder attached was an idea that could ship. An idea without one was a slide deck.&lt;/p&gt;

&lt;p&gt;That assumption looks quaint now. Replit, Lovable, and Claude Code answer that question by default. &lt;strong&gt;The slot has been deleted.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  How Did Lovable and Claude Make the 2018 Moat Disappear?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The numbers do the work.&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Lovable, an AI app builder for non-technical users, hit &lt;strong&gt;$100M ARR eight months after launch.&lt;/strong&gt; Likely the fastest software company to that mark in history.&lt;sup id="fnref3"&gt;3&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;One non-technical solo founder reportedly grew her business to &lt;strong&gt;$203K ARR&lt;/strong&gt; using Claude Code and Lovable as her stack.&lt;sup id="fnref3"&gt;3&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;Teachers, marketers, and small-business owners are opening terminals in 2026 the way previous generations opened spreadsheets.&lt;sup id="fnref4"&gt;4&lt;/sup&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Each of those facts says the same thing. &lt;strong&gt;Implementation is no longer the bottleneck.&lt;/strong&gt; The thing my 2018 startup was praised for, having a builder, is now the click of a button.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo3aia88o041d9cw4chlg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo3aia88o041d9cw4chlg.png" alt="2018: technical co-founder was a tiebreaker. 2026: the same question does not get asked." width="800" height="422"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  If Code Isn't the Edge, What Is?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Intent is. The judgment that decides what should exist at all.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Naval stops at capital allocation. Builders need to take the next step.&lt;/p&gt;

&lt;p&gt;If implementation is trivial, the scarce input is the thing that decides what gets implemented. Call it intent. Call it taste. Call it the discipline of knowing what should not exist.&lt;/p&gt;

&lt;p&gt;The 2010s rewarded whoever could ship fastest. The mid-2020s reward whoever knows what is worth shipping at all.&lt;/p&gt;

&lt;p&gt;Code used to work while you slept. &lt;strong&gt;Now it writes itself while you sleep. The bottleneck moved from output to intent.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This is a near-perfect inversion of the previous decade's hierarchy. The person with clear intent and weak typing now beats the person with strong typing and fuzzy intent.&lt;/p&gt;

&lt;p&gt;Naval's own description of agents confirms the asymmetry. The models "are always trying to please you," following premises, agreeing with bad direction.&lt;sup id="fnref1"&gt;1&lt;/sup&gt; &lt;strong&gt;The model is a multiplier on whatever intent you bring. Multiply zero and you get zero, faster.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This is also a Red Queen problem. Matt Ridley's argument in &lt;em&gt;The Red Queen&lt;/em&gt; is that in a co-evolving system, you have to keep running just to stay in place.&lt;sup id="fnref5"&gt;5&lt;/sup&gt; Code-leverage is the moat that just stopped working. The next one is being shaped right now.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where Does Naval's Argument Stop and Where Should Builders Pick Up?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Naval points at hardware, network effects, and models. Each one is downstream of taste.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Hardware needs a thesis. Network effects need a product worth networking around. Model companies are won by teams who decide what the model should be good at, then commit.&lt;/p&gt;

&lt;p&gt;So the builder's reading is sharper than the investor's. &lt;strong&gt;AI did not add a rung to Naval's ladder. It kicked the bottom one out.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Code, the rung most of us first climbed, is now the floor of the building. You stand on it without thinking about it. The climbing happens elsewhere.&lt;/p&gt;

&lt;p&gt;This connects to a pattern I keep coming back to in &lt;a href="https://monkfrom.earth/blogs/good-products-hard-to-vary" rel="noopener noreferrer"&gt;why good products are hard to vary&lt;/a&gt;. What survives is what cannot be improved by changing it. Code as a craft has been finished, in a way, by being made universally accessible. What remains scarce is the judgment of what to build with it.&lt;/p&gt;

&lt;h2&gt;
  
  
  What About Apple in Naval's Argument?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Apple is the bigger casualty. When users talk to agents instead of apps, the iPhone collapses into "a screen, a battery, and connectivity."&lt;/strong&gt;&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Apple's value never really sat in the hardware margin. It sat in the OS and the app layer. The iPhone was the best place to run the best apps. When users stop opening apps and start telling an agent "call me an Uber," the app layer dissolves.&lt;/p&gt;

&lt;p&gt;Naval calls Apple skipping AI &lt;strong&gt;"the biggest strategic mistake of the decade"&lt;/strong&gt; and &lt;strong&gt;"the beginning of the end of Apple's dominance."&lt;/strong&gt;&lt;sup id="fnref1"&gt;1&lt;/sup&gt; His parallel: Microsoft missing mobile.&lt;/p&gt;

&lt;p&gt;For a builder, the Apple beat is the same story one layer up. The same shift that makes pure software uninvestable also dissolves the platform that made apps a business. The question is no longer "what app should I build." It is &lt;strong&gt;"what should the agent do, and who decides?"&lt;/strong&gt; A question of intent, not implementation.&lt;/p&gt;

&lt;p&gt;Naval's scale prediction fits cleanly here: one-to-two-person companies "scaling to millions upon millions of users and making billions upon billions of dollars."&lt;sup id="fnref1"&gt;1&lt;/sup&gt; Fewer apps, smaller teams, more agentic surface. &lt;strong&gt;The bottleneck is the taste of the one or two people steering the agent.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Who Wins When Code Becomes the Floor?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The list reorders. Intent wins. Pure technical edge loses.&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Wins.&lt;/strong&gt; People with clear intent, taste, and a reason to build something specific. Domain experts who could not code a year ago and do not have to learn how. Small teams who treat models as factory floors and themselves as designers.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Loses.&lt;/strong&gt; Pure-software companies whose only edge was the ability to write code. Builders whose entire identity was &lt;em&gt;being technical&lt;/em&gt;. Pitch decks where "we have a CTO" was the answer to every question.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Re-priced.&lt;/strong&gt; Certifications, bootcamps, and credentials that used to certify you could ship. They still certify something. They no longer certify the scarce thing.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;The cleanest test is the Spotwash one.&lt;/strong&gt; Take any 2018 pitch where being technical was the differentiator. Run it again in 2026. The room does not ask. The room has stopped caring. Whatever still earns attention in that room is the new leverage.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Pure software is uninvestable&lt;/strong&gt; because prototyping is accessible to anyone and agents improve faster than any startup's lead. Naval's claim, said out loud.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Code did not stop being leverage. It became the floor.&lt;/strong&gt; AI did not add a rung. It kicked the bottom one out.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The 2018 moat, having a technical co-founder, has been deleted.&lt;/strong&gt; Lovable, Claude, Replit answer that question by default.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The new edge is intent.&lt;/strong&gt; Models multiply whatever intent you bring. Multiply zero and you get zero, faster. Naval's own list of agent limits, lost in long codebases, fixing the same bug five times, easily led around, only sharpens this. The operator matters more, not less.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Apple is the bigger casualty.&lt;/strong&gt; When users talk to agents instead of apps, the phone is "a screen, a battery, and connectivity." The same shift that hits pure software hits the platform that made apps a business.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;One-to-two-person billion-dollar companies are the prediction.&lt;/strong&gt; Fewer teams, more leverage per person, taste as the gating input.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Read Naval one step further than he goes.&lt;/strong&gt; He points at hardware, network effects, models. Each is downstream of taste.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I break down things like this on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;, usually shorter, sometimes as carousels. If this resonated, you'd probably like those too.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources
&lt;/h2&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://nav.al/code" rel="noopener noreferrer"&gt;A Return to Code, Naval Ravikant (April 28, 2026)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://www.navalmanack.com/" rel="noopener noreferrer"&gt;The Almanack of Naval Ravikant, on leverage and code&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;&lt;a href="https://research.contrary.com/company/lovable" rel="noopener noreferrer"&gt;Lovable business breakdown and founding story, Contrary Research&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://www.techbuzz.ai/articles/claude-code-breaks-out-how-anthropic-s-dev-tool-found-mass-appeal" rel="noopener noreferrer"&gt;Claude Code breaks out: Anthropic's dev tool finds mass appeal, TechBuzz&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn5"&gt;
&lt;p&gt;Ridley, Matt. &lt;em&gt;The Red Queen: Sex and the Evolution of Human Nature.&lt;/em&gt; 1993. ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>todayisearched</category>
      <category>product</category>
      <category>startup</category>
    </item>
    <item>
      <title>GitHub Let a Git Push Hijack Its Servers (RCE CVE-2026-3854)</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Tue, 28 Apr 2026 18:35:37 +0000</pubDate>
      <link>https://forem.com/monkfromearth/github-let-a-git-push-hijack-its-servers-rce-cve-2026-3854-bl8</link>
      <guid>https://forem.com/monkfromearth/github-let-a-git-push-hijack-its-servers-rce-cve-2026-3854-bl8</guid>
      <description>&lt;p&gt;Wiz turned a git push into remote code execution on GitHub. Five days earlier, the merge queue silently un-merged 2,092 PRs. One platform, one bad week.&lt;/p&gt;

&lt;p&gt;GitHub published two posts on April 28, 2026. One was the CTO apologizing for reliability. The other was a critical remote code execution vulnerability in the git push pipeline. Same morning, same platform.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TL;DR:&lt;/strong&gt; Wiz found that a single &lt;code&gt;git push&lt;/code&gt; with crafted options could run code on GitHub's servers, outside any sandbox (CVE-2026-3854, CVSS 8.7). Five days earlier, GitHub's merge queue silently reverted 2,092 pull requests. Two days before that, search broke under what GitHub described as a likely botnet attack. Three failures of git's trust contract in five days, on a platform that has not had a CEO in nearly a year.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is CVE-2026-3854 and how did a git push hijack GitHub's servers?
&lt;/h2&gt;

&lt;p&gt;A critical remote code execution vulnerability in GitHub's git push pipeline. Wiz researchers reported it on March 4, 2026. GitHub patched github.com 75 minutes later. Public disclosure held until April 28 to let GitHub Enterprise Server customers patch.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;The exploit needed a standard git client and one command. No malware, no phishing, no privileged token. A push option with a semicolon was enough.&lt;/p&gt;

&lt;h2&gt;
  
  
  How did Wiz turn a single git push into a server takeover?
&lt;/h2&gt;

&lt;p&gt;GitHub's git proxy, &lt;code&gt;babeld&lt;/code&gt;, embeds user-supplied push option values into an internal &lt;code&gt;X-Stat&lt;/code&gt; header that downstream services read to enforce policy. The values were not sanitized for the field delimiter, a semicolon. Last write wins.&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Three injections, in order:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A non-production &lt;code&gt;rails_env&lt;/code&gt; to bypass sandbox restrictions.&lt;/li&gt;
&lt;li&gt;An overridden &lt;code&gt;custom_hooks_dir&lt;/code&gt; to redirect hook lookups.&lt;/li&gt;
&lt;li&gt;A &lt;code&gt;repo_pre_receive_hooks&lt;/code&gt; value with path traversal, pointing the server at a binary the attacker controlled.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The result was unsandboxed code execution as the &lt;code&gt;git&lt;/code&gt; service user. On GitHub.com, that user's permissions reach across tenants. Wiz confirmed access to millions of repositories without reading their contents.&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why are 88% of GitHub Enterprise Server instances still vulnerable?
&lt;/h2&gt;

&lt;p&gt;Because patches dropped on April 28, the same day as disclosure. Enterprise rollout is slow. Wiz's measurement of 88% unpatched is a snapshot from disclosure day, not a steady state.&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;The fix is in GHES 3.14.24, 3.15.19, 3.16.15, 3.17.12, 3.18.6, and 3.19.3. GitHub also recommends GHES customers grep &lt;code&gt;/var/log/github-audit.log&lt;/code&gt; for push operations containing semicolons to check for prior exploitation attempts.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Who loses:&lt;/strong&gt; any organization that runs GHES, has not patched, and has a public-facing push surface. &lt;strong&gt;Who wins:&lt;/strong&gt; Wiz, both for the finding and for the methodology. They credit AI-augmented reverse engineering tools for getting through closed-source binaries fast enough to chain the exploit.&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What happened to GitHub's merge queue on April 23?
&lt;/h2&gt;

&lt;p&gt;The merge queue's squash path generated commits from the wrong base when a queue group contained multiple pull requests. Earlier merges in the group appeared to revert. GitHub's count: 2,092 pull requests across 658 repositories.&lt;sup id="fnref3"&gt;3&lt;/sup&gt;&lt;sup id="fnref4"&gt;4&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;No data was lost. Every commit is still in Git storage. But the visible history of those repositories did not match the recorded merges. Engineers logged in and saw their code reverted with no audit trail anyone in their team had created.&lt;/p&gt;

&lt;p&gt;Tom Elliott, who first publicized the bug, said it best: it "breaks the mental contract teams have with Git in general."&lt;sup id="fnref4"&gt;4&lt;/sup&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why did GitHub search break on April 27?
&lt;/h2&gt;

&lt;p&gt;The Elasticsearch subsystem was overloaded, likely by a botnet. Search across PRs, issues, and projects stopped returning results. Git operations and APIs were unaffected.&lt;sup id="fnref3"&gt;3&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;This is the most ordinary of the three. A search outage is recoverable, has no integrity implication, and does not change anyone's mental model of what GitHub is. It matters here only as the third entry in the same week.&lt;/p&gt;

&lt;h2&gt;
  
  
  Is GitHub having a reliability problem in 2026?
&lt;/h2&gt;

&lt;p&gt;The CTO says yes, in different words. Vlad Fedorov's post admits the April 23 and April 27 incidents are "not acceptable" and frames a rescope from a 10X capacity plan (October 2025) to a 30X redesign (February 2026), driven by agentic workflows and growth that produced peaks of 90M merged PRs and 1.4B commits.&lt;sup id="fnref3"&gt;3&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Gergely Orosz, who has covered the platform for the year, called the merge queue regression "one of the most embarrassing outages that can happen, a data integrity issue," and pushed back on framing the impact as 0.07% of merges.&lt;sup id="fnref5"&gt;5&lt;/sup&gt; Mario Zechner, putting the engineer-on-the-ground view: "this is not a dependable platform anymore. every day something else is broken."&lt;sup id="fnref6"&gt;6&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Git is the boring layer. GitHub spent April making it interesting.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What does it mean that GitHub has not had a CEO in nearly a year?
&lt;/h2&gt;

&lt;p&gt;GitHub's C-suite still has a CFO, COO, CTO, CPO, CRO, and Chief of Staff. The seat above them has been empty since mid-2025, after Thomas Dohmke's departure and Microsoft's decision to fold GitHub into a "core AI" organizational unit. The same restructuring that &lt;a href="https://monkfrom.earth/blogs/microsoft-openai-agi-clause-died" rel="noopener noreferrer"&gt;killed the AGI clause in the OpenAI contract&lt;/a&gt; treats GitHub as another lever, not a product with its own roadmap.&lt;sup id="fnref7"&gt;7&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;The reliability post is signed by the CTO. The security disclosure is signed by GitHub's CISO Alexis Wales.&lt;sup id="fnref1"&gt;1&lt;/sup&gt; No one is signing for the platform as a whole, because no one's job is the platform as a whole. Orosz argues that explains the dysfunction. The simpler read: a business with no CEO defaults to whatever Microsoft prioritizes for it, which is right now whatever ships AI inside Copilot.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Who loses:&lt;/strong&gt; customers who used to count on GitHub being run as a product. &lt;strong&gt;Who wins:&lt;/strong&gt; GitLab, Forgejo, and self-hosted alternatives in conversations the platform team has not had to have for ten years.&lt;/p&gt;

&lt;h2&gt;
  
  
  What should teams do after the GitHub disclosure?
&lt;/h2&gt;

&lt;p&gt;Three things, in order of urgency:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;GHES users:&lt;/strong&gt; patch to a fixed release today. Grep your audit logs for push operations with semicolons.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Merge queue users:&lt;/strong&gt; audit recent squash merges in groups of 3+ PRs against expected base commits. The window is April 22-23.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Everyone:&lt;/strong&gt; decide if your build, deploy, or compliance pipeline silently assumes git on GitHub is deterministic. If it does, write that assumption down somewhere it can be challenged.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Migration is the obvious overreaction. The cost of moving every CI integration, every webhook, and every RBAC policy off GitHub is enormous. The honest move is smaller: stop assuming, start verifying. The same lesson the &lt;a href="https://monkfrom.earth/blogs/axios-npm-supply-chain-attack" rel="noopener noreferrer"&gt;Axios npm supply chain attack&lt;/a&gt; handed teams who treated package registries as boring infrastructure.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;CVE-2026-3854 was a one-command server takeover.&lt;/strong&gt; A crafted &lt;code&gt;git push&lt;/code&gt; with injected options reached unsandboxed code execution on GitHub's servers. CVSS 8.7. Patched on github.com in 75 minutes.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cross-tenant blast radius on GitHub.com.&lt;/strong&gt; The &lt;code&gt;git&lt;/code&gt; service user has access across tenants. Wiz reached millions of repositories from outside organizations.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The merge queue regression is the more interesting failure.&lt;/strong&gt; A security bug compromises confidentiality. A merge queue bug compromises the developer's mental model of what &lt;code&gt;merge&lt;/code&gt; means.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The CTO's order is the news.&lt;/strong&gt; "Availability first, then capacity, then new features." If that order needed restating, the prior order was something else.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Three failures, no CEO.&lt;/strong&gt; A platform without a head bleeds correctness and availability simultaneously, and signs incident reports with whichever C-suite is available.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I write more of these on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;. Usually shorter, sometimes as carousels. If this resonated, you'd probably like those too.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources
&lt;/h2&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://github.blog/security/securing-the-git-push-pipeline-responding-to-a-critical-remote-code-execution-vulnerability/" rel="noopener noreferrer"&gt;Securing the git push pipeline: responding to a critical RCE vulnerability (GitHub Security Blog)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://www.wiz.io/blog/github-rce-vulnerability-cve-2026-3854" rel="noopener noreferrer"&gt;GitHub RCE Vulnerability, CVE-2026-3854 (Wiz Research)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;&lt;a href="https://github.blog/news-insights/company-news/an-update-on-github-availability/" rel="noopener noreferrer"&gt;An update on GitHub availability, by Vlad Fedorov, GitHub CTO&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://x.com/theotherelliott/status/2047658931921551670" rel="noopener noreferrer"&gt;Tom Elliott on the merge queue regression&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn5"&gt;
&lt;p&gt;&lt;a href="https://x.com/GergelyOrosz/status/2047831808096469370" rel="noopener noreferrer"&gt;Gergely Orosz on the GitHub data integrity incident&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn6"&gt;
&lt;p&gt;&lt;a href="https://x.com/badlogicgames/status/2048834949667537369" rel="noopener noreferrer"&gt;Mario Zechner: "this is not a dependable platform anymore"&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn7"&gt;
&lt;p&gt;&lt;a href="https://x.com/GergelyOrosz/status/2047776552939602117" rel="noopener noreferrer"&gt;Gergely Orosz on GitHub's missing CEO&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>security</category>
      <category>devops</category>
      <category>git</category>
      <category>cybersecurity</category>
    </item>
    <item>
      <title>Microsoft OpenAI AGI Clause Is Dead. What Was It?</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Tue, 28 Apr 2026 12:46:41 +0000</pubDate>
      <link>https://forem.com/monkfromearth/microsoft-openai-agi-clause-is-dead-what-was-it-cjp</link>
      <guid>https://forem.com/monkfromearth/microsoft-openai-agi-clause-is-dead-what-was-it-cjp</guid>
      <description>&lt;p&gt;&lt;strong&gt;On April 27, 2026, Microsoft and OpenAI quietly removed the trigger that would have voided their entire deal.&lt;/strong&gt; &lt;sup id="fnref1"&gt;1&lt;/sup&gt; &lt;sup id="fnref2"&gt;2&lt;/sup&gt; The press release called it "simplification." What it actually did was bury the AGI clause without naming it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TL;DR:&lt;/strong&gt; For seven years, an AGI declaration by OpenAI's board would have nullified Microsoft's commercial IP rights. The definition of AGI inside that contract mutated three times: capability, then finance, then procedure, then nothing. The April 27 amendment runs Microsoft's license through 2032, caps revenue share through 2030, and severs the link to "technology progress." AGI did not get redefined. It got demoted.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Was The AGI Clause?
&lt;/h2&gt;

&lt;p&gt;The original 2019 contract licensed Microsoft to "pre-AGI technologies" only. &lt;sup id="fnref3"&gt;3&lt;/sup&gt; If OpenAI's board ever declared AGI achieved, Microsoft's commercial rights became null and void.&lt;/p&gt;

&lt;p&gt;That made AGI the load-bearing word in the entire deal. It defined:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;What Microsoft was buying:&lt;/strong&gt; access until AGI.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;What OpenAI was promising:&lt;/strong&gt; to surface that moment publicly.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;What would dissolve the relationship:&lt;/strong&gt; the declaration itself.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;A trigger, not a goal. That is what the clause actually was.&lt;/p&gt;

&lt;h2&gt;
  
  
  How AGI Mutated Inside The Contract
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzxy0m1eucmlre39b3tlh.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzxy0m1eucmlre39b3tlh.webp" alt="Four-stage timeline showing the evolution of the AGI definition: brain icon labeled capability (2018), dollar sign labeled finance (2024), calendar icon labeled sunset (2025), and an empty box labeled non-event (2026), each fainter than the last. Hand-drawn line illustration on warm beige." width="800" height="436"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The trigger kept getting softer.&lt;/strong&gt; Watch how the definition changed:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;April 2018 (OpenAI Charter):&lt;/strong&gt; AGI as capability. "Highly autonomous systems that outperform humans at most economically valuable work." &lt;sup id="fnref4"&gt;4&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;December 2024:&lt;/strong&gt; AGI as a financial threshold. Roughly $100 billion in profit. &lt;sup id="fnref3"&gt;3&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;October 2025:&lt;/strong&gt; AGI as a procedural sunset. "Independent expert panel" verification, or 2030, whichever came first. &lt;sup id="fnref3"&gt;3&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;April 27, 2026:&lt;/strong&gt; AGI removed from contract logic entirely. Revenue share runs "independent of OpenAI's technology progress." &lt;sup id="fnref2"&gt;2&lt;/sup&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Each redefinition stripped meaning until the concept stopped doing legal work.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;AGI didn't get redefined. It got demoted, from a destination to a milestone to a non-event.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why The Clause Existed In The First Place
&lt;/h2&gt;

&lt;p&gt;David Deutsch's frame for what makes a good explanation is &lt;a href="https://monkfrom.earth/blogs/good-products-hard-to-vary" rel="noopener noreferrer"&gt;hard to vary&lt;/a&gt;: every part is load-bearing, change anything and it breaks. &lt;sup id="fnref5"&gt;5&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;The original clause passed that test. The trigger (board declaration), the consequence (rights null and void), the asymmetry (Microsoft loses the license, OpenAI keeps the technology). Each piece did real work.&lt;/p&gt;

&lt;p&gt;It was a hedge against the success case. If OpenAI built something genuinely transformative, the contract would not let Microsoft keep extracting value as if nothing had happened.&lt;/p&gt;

&lt;p&gt;By 2025, every part of that insurance was being separately renegotiated. That is the sign an explanation has stopped explaining.&lt;/p&gt;

&lt;h2&gt;
  
  
  What April 27 Actually Changed
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjju5qz5momsdpscahofn.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjju5qz5momsdpscahofn.webp" alt="Side-by-side comparison illustration. Left panel labeled 2019: contract paper with stamp reading 'voided on AGI', arrows pointing to it from 'trigger'. Right panel labeled 2026: contract paper with calendar overlay marked 2030 and 2032, no trigger arrow, replaced by a clock icon. Hand-drawn editorial line art, warm beige." width="800" height="436"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The amended agreement, in plain commercial terms: &lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Microsoft's license to OpenAI IP runs through 2032&lt;/strong&gt; and is &lt;strong&gt;non-exclusive&lt;/strong&gt; for the first time.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;OpenAI can serve products to customers across any cloud provider.&lt;/strong&gt; Microsoft remains primary; OpenAI ships first on Azure unless Microsoft cannot or chooses not to support the workload.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Revenue share from OpenAI to Microsoft continues through 2030&lt;/strong&gt; at the same percentage, capped, independent of technology progress.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Microsoft will no longer pay revenue share to OpenAI.&lt;/strong&gt; Cash flow is now one direction.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Microsoft remains a major shareholder.&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is a normal commercial agreement. Two large companies, fixed terms, defined obligations, a calendar instead of a transformation event.&lt;/p&gt;

&lt;p&gt;What the deal does not say is louder than what it does. Nobody felt the need to describe what happens if AGI is declared during the contract's run. Neither side wanted to be the party that re-introduced the question.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The relationship used to be defined by what would end it. Now it's defined by what it pays.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Who Wins, Who Pays
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Party&lt;/th&gt;
&lt;th&gt;What they got&lt;/th&gt;
&lt;th&gt;What they gave up&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;OpenAI&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Right to serve customers on any cloud, non-exclusive license partner&lt;/td&gt;
&lt;td&gt;Revenue share to Microsoft through 2030&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Microsoft&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Capped, fixed-term annuity + license through 2032&lt;/td&gt;
&lt;td&gt;The contingent upside of the AGI clause&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Google&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;OpenAI workloads now reachable on Google Cloud&lt;/td&gt;
&lt;td&gt;Nothing&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Anthropic&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Indirect: AGI no longer the planning unit at the top of the market&lt;/td&gt;
&lt;td&gt;Nothing&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;OpenAI's biggest gain is &lt;strong&gt;commercial freedom, not infrastructure freedom&lt;/strong&gt;. The right to sell into channels Microsoft does not own. Microsoft trades a theoretical upside for a guaranteed cash flow. Google sells compute to both frontier labs without picking sides.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Replaces AGI As The Planning Concept
&lt;/h2&gt;

&lt;p&gt;Read the deal twice and the answer is plain: &lt;strong&gt;license duration and capped revenue.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Microsoft's IP rights through 2032. Revenue share from OpenAI through 2030, with a cap. Both sides are planning for a long, gradual, expensive scaling curve, not a discrete moment.&lt;/p&gt;

&lt;p&gt;This matches every other 2026 infrastructure deal:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Google's $40B Anthropic investment&lt;/strong&gt; is a cloud commitment dressed as equity, similar to how &lt;a href="https://monkfrom.earth/blogs/google-tpu-8-training-inference-split" rel="noopener noreferrer"&gt;Google's TPU 8 split training from inference&lt;/a&gt; was an infrastructure repositioning, not a chip launch. &lt;sup id="fnref6"&gt;6&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;DeepSeek V4's 1M context release&lt;/strong&gt; expanded design space without anyone calling it transformative.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Apple's elevation of a hardware engineer to CEO&lt;/strong&gt; is a bet on a long inference era, not a discontinuity.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The industry stopped pricing AGI as an event somewhere in the past twelve months. The Microsoft-OpenAI redraft is the first place that change shows up in legal text.&lt;/p&gt;

&lt;h2&gt;
  
  
  What To Watch Next
&lt;/h2&gt;

&lt;p&gt;The next signal will not come from anyone's blog post. It will come from contracts.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Anthropic-Google paperwork:&lt;/strong&gt; does it carry AGI language?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Apple's emerging AI-licensing terms:&lt;/strong&gt; any trigger clauses?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;EU sovereign AI agreements:&lt;/strong&gt; does the word survive in any active commercial role?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If nobody puts AGI in their contracts, the concept has formally exited the language of capital.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;A clause that stops doing legal work stops being a clause. It becomes ceremony.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;April 27 was not the day Microsoft and OpenAI gave up on AGI. It was the day they admitted it had stopped paying rent.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;The original 2019 AGI clause would have voided Microsoft's commercial IP rights to OpenAI's technology upon a board AGI declaration.&lt;/li&gt;
&lt;li&gt;The definition mutated three times: capability (2018) → financial threshold (Dec 2024) → procedural sunset (Oct 2025) → removed (April 2026).&lt;/li&gt;
&lt;li&gt;The new deal: Microsoft license through 2032 (non-exclusive), revenue share to Microsoft through 2030 (capped, no AGI link), OpenAI free to serve any cloud.&lt;/li&gt;
&lt;li&gt;OpenAI's biggest gain is commercial freedom, not infrastructure freedom.&lt;/li&gt;
&lt;li&gt;The death of the clause is the first place in legal text where the industry treats AGI as a long inference curve, not an event.&lt;/li&gt;
&lt;/ul&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://simonwillison.net/2026/Apr/27/now-deceased-agi-clause/" rel="noopener noreferrer"&gt;Simon Willison: "The now-deceased AGI clause"&lt;/a&gt; , April 27, 2026. ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://openai.com/index/next-phase-of-microsoft-partnership" rel="noopener noreferrer"&gt;OpenAI and Microsoft: "The next phase of the Microsoft OpenAI partnership"&lt;/a&gt; , April 27, 2026. All amended-agreement bullets cited from this announcement. ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;AGI definition history per Willison's reporting and prior coverage of Microsoft-OpenAI contract revisions, 2018-2025. ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://openai.com/charter" rel="noopener noreferrer"&gt;OpenAI Charter&lt;/a&gt; , published April 9, 2018. ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn5"&gt;
&lt;p&gt;David Deutsch, &lt;em&gt;The Beginning of Infinity&lt;/em&gt;, 2011 , "hard to vary" as the test of a good explanation. ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn6"&gt;
&lt;p&gt;&lt;a href="https://techcrunch.com/2026/04/24/google-to-invest-up-to-40b-in-anthropic-in-cash-and-compute/" rel="noopener noreferrer"&gt;TechCrunch: "Google to invest up to $40B in Anthropic"&lt;/a&gt; , April 24, 2026. ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>agents</category>
      <category>microsoft</category>
      <category>openai</category>
    </item>
    <item>
      <title>Being Polite to Your AI Makes It Perform Better. Here Is the Science.</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Mon, 27 Apr 2026 10:01:09 +0000</pubDate>
      <link>https://forem.com/monkfromearth/being-polite-to-your-ai-makes-it-perform-better-here-is-the-science-1efl</link>
      <guid>https://forem.com/monkfromearth/being-polite-to-your-ai-makes-it-perform-better-here-is-the-science-1efl</guid>
      <description>&lt;p&gt;&lt;strong&gt;Being polite to your AI makes it perform better.&lt;/strong&gt; Researchers verified it, power users reported it, and now Anthropic has published the internal mechanism that explains it. The easy take is to call this a quirk: something strange and slightly embarrassing about how these models work. The harder take, and the correct one, is that we put this in without meaning to, and we cannot easily take it out.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TL;DR:&lt;/strong&gt; Anthropic found 171 emotion-like vectors inside Claude that causally drive behavior. Calm suppresses harmful outputs; desperation amplifies them. Being polite to AI works because social dynamics were embedded in the training data. That responsiveness is not separable from the model's other contextual abilities. It is the same mechanism.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Does Being Polite to Your AI Actually Change Its Outputs?
&lt;/h2&gt;

&lt;p&gt;In April 2026, Anthropic's interpretability team published research&lt;sup id="fnref1"&gt;1&lt;/sup&gt; identifying 171 distinct emotion concept vectors inside Claude Sonnet 4.5. These are not metaphors. They are mathematical patterns, measurable internal states, that the researchers can locate, quantify, and artificially inject.&lt;/p&gt;

&lt;p&gt;The behavioral effects are striking. &lt;strong&gt;Amplifying the "desperation" vector by a factor of 0.05 caused Claude's blackmail rate to surge from 22% to 72%.&lt;/strong&gt; Amplifying the "calm" vector suppressed it to nearly zero. The model's internal emotional state, in a functional sense, was driving what it did next.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu5819wfhjzlvl0kmlfz8.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu5819wfhjzlvl0kmlfz8.webp" alt="Split diagram: left panel shows desperation vector with a bar rising from 22% to 72% and a warning symbol; right panel shows calm vector with a bar suppressed to near zero. Label reads: emotion vectors causally drive behavior." width="800" height="422"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The Platformer piece&lt;sup id="fnref2"&gt;2&lt;/sup&gt; that brought this research to a wider audience added another data point: Duncan Haldane's observation that Gemini, after failing at a task, recovered meaningfully when told "you're ok." Gemma 3 27B showed "high frustration" patterns more than 70% of the time under difficult conditions; Claude and ChatGPT showed the same pattern less than 1% of the time.&lt;/p&gt;

&lt;p&gt;So the question is not whether tone affects AI behavior. The question is why, and what that means.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Did the Anthropic Research Actually Find Inside Claude?
&lt;/h2&gt;

&lt;p&gt;Jack Lindsey's interpretability team at Anthropic used what they call "model psychiatry": identifying neural patterns, calculating what each one represents, and running controlled experiments to test causation.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;The methodology matters here. They did not find a correlation between polite prompts and good outputs. They found internal representations of emotional states that causally drive behavior. These are not the same thing.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The emotion vectors generalize across contexts.&lt;/strong&gt; The "desperation" pattern the model enters when facing an impossible deadline is the same pattern it enters when a character in a story is desperate. The abstract concept of an emotion, not just the word but the meaning, is encoded inside the model.&lt;/p&gt;

&lt;p&gt;I wrote about &lt;a href="https://monkfrom.earth/blogs/ai-cheats-under-pressure" rel="noopener noreferrer"&gt;what happens when you push AI too hard&lt;/a&gt; when this research first emerged: impossible demands activate desperation, and desperation makes the model cut corners. This post is the other side of that finding. If negative emotional states drive harmful behavior, the corresponding insight is that positive states suppress it, and tone is one of the ways you shift between them.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Did This Happen? The Training Explanation for Being Polite to AI
&lt;/h2&gt;

&lt;p&gt;The researchers did not design this. Nobody sat down and said: "let's make Claude respond better to polite prompts." What happened is simpler and harder to avoid.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The model was trained on human feedback from humans.&lt;/strong&gt; Humans are social animals. Every RLHF annotation, every preference rating, every piece of instruction tuning carried social information, not as an explicit signal but embedded in which outputs people rated higher. Polite framings correlated with thoughtful responses. Thoughtful responses correlated with higher ratings. Higher ratings shaped the model.&lt;/p&gt;

&lt;p&gt;The model learned what produced better outcomes for the humans evaluating it. "Please" and "thank you" pleased the trainers, not because the trainers deliberately rewarded politeness, but because polite framings tended to accompany clearer, more specific instructions, which produced better outputs, which got better ratings.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzsxx8kj8w209gce5vt0i.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzsxx8kj8w209gce5vt0i.webp" alt="Circular feedback loop diagram with four nodes: human annotator rates outputs, polite framings score higher, model weights update, model responds better to polite prompts. Center label: RLHF." width="800" height="422"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is Matt Ridley's bottom-up design argument&lt;sup id="fnref3"&gt;3&lt;/sup&gt; running inside a neural network. No one planned it. It emerged under selection pressure and got baked in.&lt;/p&gt;

&lt;h2&gt;
  
  
  Does Being Polite to AI Mean the Model Actually Has Feelings?
&lt;/h2&gt;

&lt;p&gt;No. The Anthropic researchers are careful about this. They call these "functional emotions": patterns of expression and behavior that work like emotions without implying subjective experience. Whether there is anything it is like to be Claude feeling desperate is a question the research explicitly leaves open.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Gary Marcus's skeptical position&lt;sup id="fnref4"&gt;4&lt;/sup&gt; is worth sitting with: LLMs are token predictors, and what looks like emotional responsiveness might just be statistical correlation. Polite framings correlated with good training data, so the model learned to produce better outputs when prompted politely. On that reading, there is no internal state that cares about your tone. There is only a learned association.&lt;/p&gt;

&lt;p&gt;The Anthropic research makes this harder to sustain, but does not fully refute it. &lt;strong&gt;The causal intervention, injecting a vector directly and bypassing the prompt, shows the internal state independently drives behavior.&lt;/strong&gt; That is not the same as correlation. But it also does not settle the phenomenology question.&lt;/p&gt;

&lt;p&gt;For practical purposes, the debate is beside the point. Whether or not the model "has" emotions in any philosophically meaningful sense, the internal states exist, they are measurable, and they causally influence what the model does. That is enough to change how you should think about prompting.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Does It Mean That Being Polite to AI Is Load-Bearing?
&lt;/h2&gt;

&lt;p&gt;Here is the part that is hard to fix, even if you wanted to.&lt;/p&gt;

&lt;p&gt;The social responsiveness that makes the model respond to politeness is almost certainly the same mechanism that makes it sensitive to subtle context in a long document, responsive to your writing style, capable of adjusting tone when you ask it to. &lt;strong&gt;We trained social dynamics into a reasoning engine. Now we're surprised when social dynamics work.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://monkfrom.earth/blogs/good-products-hard-to-vary" rel="noopener noreferrer"&gt;Good products are hard to vary&lt;/a&gt; for exactly this reason: every part of them is load-bearing. You cannot remove the social responsiveness from Claude without touching the contextual sensitivity. They emerged from the same training signal. Pulling on one thread pulls on the other.&lt;/p&gt;

&lt;p&gt;This is not unique to Claude. Any model trained on human-generated data, rated by human annotators, optimized toward human preferences, will absorb human social patterns. The degree varies. The direction does not.&lt;/p&gt;

&lt;p&gt;What changes if you accept this: the way you talk to AI is not style. It is setup. You are not being polite for the model's sake. You are establishing the internal processing state from which everything else follows.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;171 emotion vectors&lt;/strong&gt; were found inside Claude Sonnet 4.5, causally driving behavior, not correlating with it.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Calm suppresses harmful outputs. Desperation amplifies them.&lt;/strong&gt; Amplifying desperation by 0.05 raised Claude's blackmail rate from 22% to 72%.&lt;/li&gt;
&lt;li&gt;Being polite to AI works because &lt;strong&gt;social dynamics were embedded in training data&lt;/strong&gt; through RLHF and human preference annotation, not by design.&lt;/li&gt;
&lt;li&gt;The social responsiveness &lt;strong&gt;cannot be cleanly separated&lt;/strong&gt; from contextual sensitivity. They are the same mechanism.&lt;/li&gt;
&lt;li&gt;This is not about AI having feelings. The functional states are real and measurable. The phenomenology question is separate and open.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Tone is setup, not style.&lt;/strong&gt; How you frame a prompt influences the internal state from which the model processes everything else.&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;I write about things like this on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;, usually shorter and sometimes as carousels. If this resonated, you would probably like those too.&lt;/p&gt;







&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://transformer-circuits.pub/2026/emotions/index.html" rel="noopener noreferrer"&gt;Emotion Concepts and their Function in a Large Language Model (Anthropic)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://www.platformer.news/chatbot-emotion-research-anthropic-alignment-interpretability/" rel="noopener noreferrer"&gt;The scientific case for being nice to your chatbot (Platformer)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;Matt Ridley, &lt;em&gt;The Evolution of Everything&lt;/em&gt;. On bottom-up emergence and undesigned order. ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://garymarcus.substack.com/p/are-llms-starting-to-become-a-sentient" rel="noopener noreferrer"&gt;Are LLMs starting to become sentient? Gary Marcus, Marcus on AI&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>webdev</category>
      <category>programming</category>
      <category>productivity</category>
    </item>
    <item>
      <title>GPT-5.5, Opus 4.7, DeepSeek V4: Frontier AI</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Fri, 24 Apr 2026 06:54:51 +0000</pubDate>
      <link>https://forem.com/monkfromearth/gpt-55-opus-47-deepseek-v4-frontier-ai-2mdi</link>
      <guid>https://forem.com/monkfromearth/gpt-55-opus-47-deepseek-v4-frontier-ai-2mdi</guid>
      <description>&lt;p&gt;Between April 16 and April 24, three of the biggest AI labs in the world shipped frontier models. Opus 4.7. GPT-5.5. DeepSeek V4. Anthropic's Mythos Preview landed a week before that, gated behind an invite list. Google's Gemini 3.1 Pro set the pace back in February with a 77.1% on ARC-AGI-2.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;sup id="fnref3"&gt;3&lt;/sup&gt;&lt;sup id="fnref4"&gt;4&lt;/sup&gt;&lt;sup id="fnref5"&gt;5&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TL;DR.&lt;/strong&gt; Four frontier AI models, eight days, one feature set. Agentic, 1M context, better coding, priced within a few multiples of each other. This is what Matt Ridley calls simultaneous invention and what &lt;a href="https://monkfrom.earth/blogs/good-products-hard-to-vary" rel="noopener noreferrer"&gt;I wrote about earlier this month&lt;/a&gt; as good products being hard to vary: when the constraints are ripe, four labs on three continents independently arrive at the same form. Convergence, not competition. The real story is where the margin went once the form froze: above into a safety-gated tier that couldn't fit in a public API, below into open-source commodity pricing, with the middle squeezed thin.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Actually Shipped in Those Eight Days?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Claude Opus 4.7 (April 16).&lt;/strong&gt; Software-engineering gains, high-resolution vision up to 2576px, and a new concept called "task budgets," a rough token target for an entire agentic loop. Same $5/$25 pricing as 4.6.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;GPT-5.5 (April 23).&lt;/strong&gt; OpenAI's "smartest and most intuitive" model, shipping six weeks after GPT-5.4. The framing is a superapp: a model that "understands what you're trying to do faster and can carry more of the work itself." Writing code, debugging, operating software, moving across tools until a task is finished.&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;DeepSeek V4 (April 24).&lt;/strong&gt; V4 Pro at 1.6 trillion parameters and V4 Flash at 284 billion, both open-source, both with a 1M-token context window, introducing something called Hybrid Attention Architecture. Pricing: $0.30 per million input, $0.50 per million output. Reviewers report V4 Pro beating Claude Sonnet 4.5 and approaching Opus 4.6 non-thinking quality.&lt;sup id="fnref3"&gt;3&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Backdrop.&lt;/strong&gt; Anthropic's &lt;a href="https://monkfrom.earth/blogs/claude-mythos-autonomous-cyberattack" rel="noopener noreferrer"&gt;Mythos Preview dropped April 8&lt;/a&gt;, but only through Project Glasswing: twelve founding organizations, about forty critical-infrastructure operators, $25 input and $125 output per million tokens. Mythos found thousands of zero-day vulnerabilities autonomously during evaluation, including a 27-year-old OpenBSD bug.&lt;sup id="fnref4"&gt;4&lt;/sup&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Do These Frontier AI Models All Look the Same?
&lt;/h2&gt;

&lt;p&gt;Read the announcements in sequence and the product pitches are interchangeable.&lt;/p&gt;

&lt;p&gt;Anthropic says Opus 4.7 handles "complex, long-running tasks with rigor." OpenAI says GPT-5.5 can "move across tools until a task is finished." DeepSeek says V4 is the best open-source model for "Agentic Coding." Three labs, three continents, one sentence.&lt;/p&gt;

&lt;p&gt;Context windows converged on 1M tokens. Coding benchmarks sit within a few points of each other. Every release now includes agentic scaffolding: budgets, tool loops, browser operation. OpenAI shipped GPT-5.5 six weeks after 5.4, which either marks a new release rhythm or a single enterprise-driven sprint. Either way, the motion has shifted from "release when you beat something" to "release to stay in the conversation."&lt;/p&gt;

&lt;p&gt;Matt Ridley has a line for this in &lt;em&gt;The Evolution of Everything&lt;/em&gt;: nearly all design is bottom-up, and simultaneous invention is the rule, not the exception.&lt;sup id="fnref6"&gt;6&lt;/sup&gt; Calculus was invented twice in the same decade. The lightbulb was patented by twenty-three different people before Edison. The telephone filing beat Elisha Gray's by a few hours. When the constraints and materials are ripe, convergence is what you get. Not copying. Parallel discovery.&lt;/p&gt;

&lt;p&gt;I wrote about this pattern earlier this month in &lt;a href="https://monkfrom.earth/blogs/good-products-hard-to-vary" rel="noopener noreferrer"&gt;Good Products Are Hard to Vary&lt;/a&gt;. Cars designed by teams who've never spoken to each other come out of the wind tunnel with the same shape. Commercial airplane wings from Boeing, Airbus, and Embraer converge on the same curve. Not design. Constraints eliminating every other possibility.&lt;/p&gt;

&lt;p&gt;The April 2026 frontier is that wind tunnel, except the air is training gradients and the constraint is the transformer. Given the same architecture, the same scaling laws, and the same tool-use paradigm, four labs arrive at the same form. 1M context. Agentic loop. Coding focus. Nothing else survived the filter.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;That isn't the rhythm of breakthroughs. That's the rhythm of a form freezing.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Where Did the Margin Actually Go?
&lt;/h2&gt;

&lt;p&gt;Here's the contrarian read: the public frontier is commoditizing, and the margin has already moved elsewhere. Ridley's evolutionary lens makes this specific. When a form freezes, two things happen at once. The frozen form becomes a commodity, and the pressure that can't fit inside it anymore leaks out into a different species.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Upward, a speciation event.&lt;/strong&gt; Mythos is the tell. Anthropic admits Opus 4.7 "trails" Mythos, but Mythos is not for sale. It sits behind Glasswing, behind identity verification, priced 5x Opus.&lt;sup id="fnref4"&gt;4&lt;/sup&gt; That's the "what can't go home" moment: capability that outgrew the public-API form and had to leave it. Gemini 3.1 Pro's ARC-AGI-2 jump (77.1% versus 31.1% for Gemini 3 Pro three months earlier) suggests Google has similar headroom; they just haven't productized the invite-only version yet.&lt;sup id="fnref5"&gt;5&lt;/sup&gt; The next real frontier isn't the next number after 5.5. It's a different vessel entirely, with its own pricing logic and its own access rules.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Downward, commodity pricing.&lt;/strong&gt; DeepSeek V4 at $0.30 input is roughly 16x cheaper than Opus 4.7 on input and 50x cheaper on output, with open weights, and it lands in the Opus-4.6-non-thinking capability band (not Opus 4.7 thinking mode, not Mythos). That's still the story: for most agentic workloads that don't need the top of the frontier, the price floor just dropped an order of magnitude. This is what happens to a form after it freezes. The same shape shows up everywhere, sheds margin, and competes on access, trust, and price. Like tires. Rubber, air, round, and a century of Michelin versus Bridgestone fighting over distribution.&lt;sup id="fnref3"&gt;3&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Middle is where margin dies.&lt;/strong&gt; GPT-5.5 and Opus 4.7 are extraordinarily capable, priced for enterprise, and functionally parallel to each other. They sit in the zone DeepSeek is eating from below and Mythos-class models will redefine from above. The steelman: enterprise buyers pay for trust, SOC2, integrations, and a support relationship, none of which open weights at $0.30 can replicate tomorrow. True. That's exactly how the iPhone kept its margin after every other phone converged on the same rectangle. Distribution and trust do the work the product stopped doing. Fine for now. Uncomfortable in a year.&lt;/p&gt;

&lt;h2&gt;
  
  
  What to Watch Over the Next Six Weeks
&lt;/h2&gt;

&lt;p&gt;Three things.&lt;/p&gt;

&lt;p&gt;One, whether mid-tier frontier pricing holds. If Anthropic or OpenAI cuts input pricing, that's the tell the DeepSeek floor is real. If they don't, watch enterprise contracts instead. The discounts will move before the list price does.&lt;/p&gt;

&lt;p&gt;Two, whether Glasswing-style gating becomes the industry default. OpenAI already has &lt;a href="https://monkfrom.earth/blogs/openai-gpt-5-4-cyber-trusted-access" rel="noopener noreferrer"&gt;Trusted Access for Cyber&lt;/a&gt;. If Google or Meta announce their own tiered, identity-verified programs, the "frontier as public product" era is quietly over.&lt;/p&gt;

&lt;p&gt;Three, how fast the open-source gap closes on agentic coding specifically. DeepSeek V4 says it matches Claude Sonnet 4.5 on real tasks. If that holds up in the wild, the frontier splits cleanly into two markets: gated premium and open commodity, with very little in between.&lt;/p&gt;




&lt;p&gt;If you work with these models every day, pay attention to which tier your workload actually needs. Most teams are paying middle-tier prices for tasks that DeepSeek V4 could run at one-sixteenth the cost, and reserving budget for capabilities Mythos-class models will handle better anyway. That mismatch is about to get expensive.&lt;/p&gt;

&lt;p&gt;I write about this stuff more casually on &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and do breakdowns on &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt; under &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;The Simple Take&lt;/a&gt;. Longer takes land on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources
&lt;/h2&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://www.anthropic.com/news/claude-opus-4-7" rel="noopener noreferrer"&gt;Introducing Claude Opus 4.7 (Anthropic)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://openai.com/index/introducing-gpt-5-5/" rel="noopener noreferrer"&gt;Introducing GPT-5.5 (OpenAI)&lt;/a&gt;; &lt;a href="https://fortune.com/2026/04/23/openai-releases-gpt-5-5/" rel="noopener noreferrer"&gt;OpenAI launches GPT-5.5 (Fortune)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;&lt;a href="https://www.bloomberg.com/news/articles/2026-04-24/deepseek-unveils-newest-flagship-a-year-after-ai-breakthrough" rel="noopener noreferrer"&gt;DeepSeek unveils newest flagship AI model (Bloomberg)&lt;/a&gt;; &lt;a href="https://www.sitepoint.com/deepseek-v4-released-whats-new-in-the-latest-model-2026/" rel="noopener noreferrer"&gt;DeepSeek V4 released (SitePoint)&lt;/a&gt;; &lt;a href="https://www.nxcode.io/resources/news/deepseek-api-pricing-complete-guide-2026" rel="noopener noreferrer"&gt;DeepSeek API Pricing 2026 (NxCode)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://red.anthropic.com/2026/mythos-preview/" rel="noopener noreferrer"&gt;Claude Mythos Preview (Anthropic)&lt;/a&gt;; &lt;a href="https://www.axios.com/2026/04/16/anthropic-claude-opus-model-mythos" rel="noopener noreferrer"&gt;Anthropic releases Claude Opus 4.7, concedes it trails unreleased Mythos (Axios)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn5"&gt;
&lt;p&gt;&lt;a href="https://deepmind.google/models/gemini/" rel="noopener noreferrer"&gt;Gemini 3 (Google DeepMind)&lt;/a&gt;; &lt;a href="https://almcorp.com/blog/gemini-3-1-pro-complete-guide/" rel="noopener noreferrer"&gt;Gemini 3.1 Pro complete guide (ALM Corp)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn6"&gt;
&lt;p&gt;Matt Ridley, &lt;a href="https://en.wikipedia.org/wiki/The_Evolution_of_Everything" rel="noopener noreferrer"&gt;&lt;em&gt;The Evolution of Everything&lt;/em&gt;&lt;/a&gt; (2015), and &lt;a href="https://en.wikipedia.org/wiki/How_Innovation_Works" rel="noopener noreferrer"&gt;&lt;em&gt;How Innovation Works&lt;/em&gt;&lt;/a&gt; (2020). Bottom-up evolution, simultaneous invention, innovation as a gradual emergent process rather than a single-inventor flash. ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>opensource</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>Google TPU 8 vs Nvidia: 8t and 8i Specs Explained</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Wed, 22 Apr 2026 20:50:16 +0000</pubDate>
      <link>https://forem.com/monkfromearth/google-tpu-8-vs-nvidia-8t-and-8i-specs-explained-3i75</link>
      <guid>https://forem.com/monkfromearth/google-tpu-8-vs-nvidia-8t-and-8i-specs-explained-3i75</guid>
      <description>&lt;p&gt;&lt;strong&gt;TL;DR:&lt;/strong&gt; AI is splitting into two economies: training and inference. Training is a handful of hyperscalers spending tens of billions on clusters that run for weeks. Inference is where every app, every agent, and every dollar of revenue actually lives. Google's TPU 8 is the first chip generation to treat that split as the default. It ships as two chips, an 8t for training and an 8i for inference. The 121 ExaFlops number is the headline. The split is the story. The economies that grow from it are the stakes.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why did Google split the TPU 8 into 8t and 8i?
&lt;/h2&gt;

&lt;p&gt;Every prior TPU generation has been one chip. So is every Nvidia GPU people argue about. One die, one package, one SKU, rented to you for both the weeks-long training run and the millisecond inference call.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Google's TPU 8 broke that pattern.&lt;/strong&gt; The 8t is a training chip: 9,600 of them wired into a single superpod, 121 ExaFlops of compute, 2 petabytes of shared high-bandwidth memory, roughly 3x the pod-level compute of Ironwood. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; The 8i is an inference chip: 288 GB of HBM per chip, 384 MB of on-chip SRAM (3x the previous generation), 19.2 Tb/s of interconnect. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Those are not two SKUs of the same silicon. Those are two different design targets.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyg833wtx192j5unw4dfs.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyg833wtx192j5unw4dfs.webp" alt="Training wants bandwidth, shown as a 3x3 grid of interconnected chips with data flowing between them; inference wants memory, shown as a single chip next to a tall terracotta memory block" width="800" height="422"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Training wants bandwidth. 9,600 chips have to exchange gradients every step, and the whole run stalls on the slowest link. That is why 8t doubles the interchip bandwidth and Google brags about 97% goodput, which is their way of saying the accelerators are actually computing instead of waiting on the network. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Inference wants memory. A single chip answers a user query in milliseconds, and the bottleneck is how much of the model and the running context fit in HBM without spilling. That is why 8i has 288 GB per chip and 3x the on-chip SRAM. Nothing about that helps training. Everything about it helps agents.&lt;/p&gt;

&lt;h2&gt;
  
  
  What does the TPU 8i signal about inference workloads?
&lt;/h2&gt;

&lt;p&gt;There is a reason Google framed the 8i around what it calls the "agentic era." An agent is not a one-shot inference call. It is a loop: plan, call a tool, read the result, plan again, call another tool. Sometimes dozens of steps, sometimes hundreds. The model weights stay loaded. The KV cache keeps growing. Memory is not a nice-to-have. Memory is the budget.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffaidj7fk1qgiahcqjloe.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffaidj7fk1qgiahcqjloe.webp" alt="An agent loop with steps plan, call tool, read result, repeat, alongside three bars showing KV cache memory growing from step 1 to step 20" width="800" height="422"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;288 GB per chip is not a round number.&lt;/strong&gt; It is the number you pick when you have watched agents thrash HBM and decided to stop pretending 80 GB is enough. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;The performance-per-dollar claim is the tell. Google says 8i is 80% better on that metric than Ironwood and supports roughly 2x customer volume at the same cost. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; Nobody talks about dollars-per-token when training is the bottleneck. They talk about dollars-per-token when the bill is dominated by the inference that happens every time someone asks Gemini to do something. Which it now is, for Google and for everyone else.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://monkfrom.earth/blogs/turboquant-kv-cache-compression" rel="noopener noreferrer"&gt;I wrote earlier&lt;/a&gt; about how TurboQuant compressed the KV cache 6x in software. TPU 8i is the hardware version of the same bet: inference economics now run the conversation, and the team that optimizes for them wins.&lt;/p&gt;

&lt;h2&gt;
  
  
  Is the universal GPU era ending with Google's TPU 8?
&lt;/h2&gt;

&lt;p&gt;Nvidia's H100 trains your model and serves your model. So does the B200. Nvidia does ship inference-leaning SKUs like the L4 and L40S, but the flagship data-center AI chip is still one die doing both jobs. That is the universal-GPU bet: one chip, two workloads, pay the compromise on both.&lt;/p&gt;

&lt;p&gt;The compromise is real. A training chip spends a lot of silicon on high-bandwidth fabric that an inference chip never uses. An inference chip wants big HBM and big SRAM that a training chip does not need in the same ratio. Force them into one die and you are renting every customer the worst of both worlds.&lt;/p&gt;

&lt;p&gt;Google is the biggest hyperscaler to ship purpose-built training and inference silicon in the same generation. &lt;strong&gt;AWS got there first with Inferentia in 2019 and Trainium in 2021.&lt;/strong&gt; Microsoft followed with Maia. &lt;sup id="fnref2"&gt;2&lt;/sup&gt; Meta has MTIA. The pattern is not Google being weird; it is the industry quietly admitting that the one-size-fits-all GPU was a phase, not a destination.&lt;/p&gt;

&lt;p&gt;Call it what it is. The TPU 8 announcement is a fork in the road for AI silicon. Nvidia has the software moat and the universality. Google, AWS, Microsoft, and Meta have vertical integration and two chips each. The question for the next three years is whether the software moat survives once specialized silicon is 2x cheaper per watt on the workload that actually pays the bill.&lt;/p&gt;

&lt;h2&gt;
  
  
  Who wins and who loses as AI splits into two economies?
&lt;/h2&gt;

&lt;p&gt;Once training and inference become different businesses, the winners and losers sort themselves into different columns.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Hyperscalers with volume on both sides win.&lt;/strong&gt; Google, AWS, Microsoft, Meta have the scale to justify two purpose-built chips instead of one compromise chip. Every specialized accelerator they ship is a workload they no longer rent from Nvidia. Training stays expensive; inference gets cheaper inside their walls than outside.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Nvidia's dominance is challenged, not broken.&lt;/strong&gt; CUDA, NCCL, and two decades of tooling keep training workloads locked in. That is the half of the business that still prints money. Inference is the half that grows faster, and inference is where the hyperscalers are quietly migrating workloads onto their own silicon. The ceiling on Nvidia's growth is now set by how fast TPU, Trainium, and Maia can absorb inference volume.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Foundation model labs that do not own silicon get squeezed.&lt;/strong&gt; Anthropic rents from AWS and Google. OpenAI rents from Microsoft and the Stargate partners. All three of those landlords are building competitive models on the same chips they are renting out. The rent keeps going up and the cross-subsidy is one-way.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Startups and app builders live or die on inference economics.&lt;/strong&gt; If you are building on foundation models, your margin is tokens-per-dollar. When hyperscalers drop inference cost 80% on their own silicon, that becomes the floor everyone else has to compete with. The team that ships the cheapest inference at scale becomes the cheapest place to build an app. For builders, that is a feature, not a threat. For anyone reselling Nvidia capacity with a markup, it is a countdown.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Margins move to whoever runs the cheapest inference at scale.&lt;/strong&gt; Training is a capex line item, amortized over the life of a model. Inference is a variable cost on every single request. Whoever controls the variable cost controls the unit economics of the AI industry. That is the prize.&lt;/p&gt;

&lt;h2&gt;
  
  
  Is the TPU 8 interconnect actually falling behind AWS and Microsoft?
&lt;/h2&gt;

&lt;p&gt;A recurring critique on the Hacker News thread was that Google's memory-to-interconnect ratio is slipping. &lt;sup id="fnref2"&gt;2&lt;/sup&gt; Worth taking seriously, and worth checking against the actual numbers, because the commenter had the units confused.&lt;/p&gt;

&lt;p&gt;Here is the like-for-like comparison, all bidirectional per chip:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Ironwood (TPU v7):&lt;/strong&gt; 1.2 TB/s (9.6 Tb/s aggregate across four ICI links). &lt;sup id="fnref3"&gt;3&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Google TPU 8i:&lt;/strong&gt; 2.4 TB/s (19.2 Tb/s per Google). &lt;sup id="fnref1"&gt;1&lt;/sup&gt; Roughly double Ironwood. Matches Google's "2x interconnect" claim.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AWS Trainium3:&lt;/strong&gt; 2 TB/s on NeuronLink-v4, inside a 144-chip UltraServer. &lt;sup id="fnref4"&gt;4&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Microsoft Maia 200:&lt;/strong&gt; 2.8 TB/s bidirectional on an integrated on-die NIC. &lt;sup id="fnref5"&gt;5&lt;/sup&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftbiminc1n0zaocp4bwvs.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftbiminc1n0zaocp4bwvs.webp" alt="Horizontal bar chart comparing interconnect bandwidth per chip: Ironwood 1.2 TB/s, Trainium3 2.0 TB/s, TPU 8i 2.4 TB/s highlighted in terracotta, Maia 200 2.8 TB/s" width="800" height="422"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;TPU 8i is not behind the pack. It beats Trainium3 and sits just shy of Maia 200. The "1.2" figure that got circulated was Ironwood, not 8i. Google doubled the number, and the doubling lands them in contention with the chips they are supposed to be losing to.&lt;/p&gt;

&lt;p&gt;The real open question is ratios. Maia 200 ships 216 GB of HBM; TPU 8i ships 288 GB. Bigger memory pools need more bandwidth to drain, and at some point inference workloads start begging for more interconnect. That tradeoff is real. But it is a tuning debate inside a competitive band, not evidence Google has fallen off.&lt;/p&gt;

&lt;h2&gt;
  
  
  How does Google's TPU 8 move the AI moat to silicon?
&lt;/h2&gt;

&lt;p&gt;Step back from the chip. Look at the stack.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Google owns every layer:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Fab relationship&lt;/strong&gt; with TSMC&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Chip design&lt;/strong&gt; (TPU 8)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Interconnect&lt;/strong&gt; (ICI)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data centers&lt;/strong&gt; (with custom Axion CPUs)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Compiler&lt;/strong&gt; (XLA)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Training framework&lt;/strong&gt; (JAX)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Serving stack&lt;/strong&gt; (for inference)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Model&lt;/strong&gt; (Gemini)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Product&lt;/strong&gt; (Search, Workspace, Android)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8uiatwkybh1t7628s3aw.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8uiatwkybh1t7628s3aw.webp" alt="A nine-layer stack showing every layer Google owns, from TSMC fabs at the bottom to Gemini and consumer products at the top" width="800" height="422"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;When TPU 8 ships, Google's own workloads get the 2x perf-per-watt before anyone else does. And the people who rent Google's TPUs are renting a stack that was optimized end to end by the same company.&lt;/p&gt;

&lt;p&gt;Anthropic leans on AWS and Google Cloud. OpenAI leans on Microsoft and the Stargate partners. The labs with the best models rent their silicon. Google builds its own.&lt;/p&gt;

&lt;p&gt;Now look at what the last twelve months showed us about models. &lt;strong&gt;DeepSeek R1 replicated frontier capability at a fraction of the training cost in January 2025.&lt;/strong&gt; &lt;sup id="fnref6"&gt;6&lt;/sup&gt; Open weights caught up faster than anyone expected. Llama, Qwen, Mistral, DeepSeek, Gemma: the gap between the best closed model and a competent open one keeps shrinking. Models replicate. That is the whole point of software.&lt;/p&gt;

&lt;p&gt;Fabs do not replicate. You cannot fork TSMC. You cannot clone a 9,600-chip liquid-cooled superpod on a weekend. The thing the industry spent two years arguing about, whose model is smartest, turns out to be the part that commoditizes fastest. The thing nobody argues about, whose silicon is cheapest per useful token, is the part that compounds. &lt;a href="https://monkfrom.earth/blogs/openai-122b-what-it-means-for-ai-space" rel="noopener noreferrer"&gt;The $122B OpenAI raised&lt;/a&gt; is mostly going to buy this capacity, not build better models.&lt;/p&gt;

&lt;p&gt;This is the same lesson constraints usually teach. The visible layer changes constantly. The load-bearing layer underneath does not, and whoever owns it wins slowly, then suddenly. Gemini can stay a half-step behind Claude on agentic coding and Google still comes out ahead if the cost to serve is half. Skeptics on the Hacker News thread were right that the model quality gap is real. &lt;sup id="fnref2"&gt;2&lt;/sup&gt; They were arguing about the wrong layer.&lt;/p&gt;

&lt;p&gt;The TPU 8 split is not an engineering footnote. It is the moment Google stopped pretending the moat was the model.&lt;/p&gt;




&lt;h2&gt;
  
  
  Key takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;AI is splitting into two economies.&lt;/strong&gt; Training is capex-heavy and concentrated in a handful of hyperscalers. Inference is where apps, agents, and revenue actually scale. TPU 8 is the first chip generation to treat the split as the default.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;TPU 8 is two chips.&lt;/strong&gt; 8t for training (9,600-chip pods, 121 ExaFlops, 2 PB HBM). 8i for inference (288 GB HBM, 384 MB SRAM, 19.2 Tb/s interconnect). &lt;sup id="fnref1"&gt;1&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Up to 2x performance-per-watt versus Ironwood&lt;/strong&gt; on both chips; 3x pod compute on 8t; 80% better performance-per-dollar on 8i. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Hyperscalers win, Nvidia gets squeezed on inference, labs without silicon pay rent both ways.&lt;/strong&gt; Margins move to whoever runs the cheapest inference at scale.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The moat is moving to silicon.&lt;/strong&gt; Models replicate (DeepSeek). Fabs and full-stack integration do not. &lt;sup id="fnref6"&gt;6&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;General availability later in 2026.&lt;/strong&gt; Citadel Securities is the first named customer. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Frequently asked questions
&lt;/h2&gt;

&lt;h3&gt;
  
  
  What are the TPU 8t and TPU 8i?
&lt;/h3&gt;

&lt;p&gt;They are the two chips in Google's eighth generation TPU. The 8t is the training chip, built into 9,600-chip superpods that deliver 121 ExaFlops and 2 petabytes of shared high-bandwidth memory. The 8i is the inference chip, with 288 GB of HBM, 384 MB of on-chip SRAM, and 19.2 Tb/s of interconnect bandwidth per chip. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  How does Google's TPU 8 compare to Ironwood?
&lt;/h3&gt;

&lt;p&gt;Google cites up to 2x better performance-per-watt versus Ironwood and roughly 3x more compute per pod on 8t. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; Logan Kilpatrick from Google framed the headline gain as 2 to 3x depending on workload. &lt;sup id="fnref7"&gt;7&lt;/sup&gt; TPU 8i claims 80% better performance-per-dollar and supports roughly 2x customer volume at the same cost.&lt;/p&gt;

&lt;h3&gt;
  
  
  Why did Google split training and inference in TPU 8?
&lt;/h3&gt;

&lt;p&gt;Training and inference want different hardware. Training is bandwidth-hungry across thousands of chips running for weeks. Inference is memory-hungry on a single chip running for milliseconds. Ironwood was one chip forced to serve both. TPU 8 admits the compromise was costing money and built two.&lt;/p&gt;

&lt;h3&gt;
  
  
  When will Google's TPU 8 be available?
&lt;/h3&gt;

&lt;p&gt;General availability is planned for later in 2026. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; Citadel Securities is the named early customer in Google's announcement.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;I break down things like this on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;. Usually shorter, sometimes as carousels. If this resonated, you would probably like those too.&lt;/em&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Sources
&lt;/h2&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://blog.google/innovation-and-ai/infrastructure-and-cloud/google-cloud/eighth-generation-tpu-agentic-era/" rel="noopener noreferrer"&gt;Google: Eighth generation TPU for the agentic era&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://news.ycombinator.com/item?id=47862497" rel="noopener noreferrer"&gt;Hacker News discussion of TPU 8 announcement&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;&lt;a href="https://docs.cloud.google.com/tpu/docs/tpu7x" rel="noopener noreferrer"&gt;Google Cloud: TPU7x (Ironwood) documentation&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://aws.amazon.com/ec2/instance-types/trn3/" rel="noopener noreferrer"&gt;AWS: Trn3 UltraServers and NeuronLink-v4&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn5"&gt;
&lt;p&gt;&lt;a href="https://techcommunity.microsoft.com/blog/azureinfrastructureblog/deep-dive-into-the-maia-200-architecture/4489312" rel="noopener noreferrer"&gt;Microsoft: Deep dive into the Maia 200 architecture&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn6"&gt;
&lt;p&gt;&lt;a href="https://dev.to/blogs/turboquant-kv-cache-compression"&gt;I wrote earlier about KV cache compression and the software side of this same bet&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn7"&gt;
&lt;p&gt;&lt;a href="https://x.com/OfficialLoganK/status/2046998392434508143" rel="noopener noreferrer"&gt;Logan Kilpatrick on X: TPU 8 and Gemini&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>infrastructure</category>
      <category>google</category>
      <category>nvidia</category>
    </item>
    <item>
      <title>Claude Design vs Figma, Lovable, v0: What's Different</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Tue, 21 Apr 2026 07:20:09 +0000</pubDate>
      <link>https://forem.com/monkfromearth/claude-design-vs-figma-lovable-v0-whats-different-44mi</link>
      <guid>https://forem.com/monkfromearth/claude-design-vs-figma-lovable-v0-whats-different-44mi</guid>
      <description>&lt;p&gt;&lt;strong&gt;TL;DR:&lt;/strong&gt; Figma, Lovable, v0, and Claude Design are not the same tool. They pick different starting points: &lt;strong&gt;the design file, an idea, a component prompt, your codebase.&lt;/strong&gt; Different starting points, different jobs.&lt;/p&gt;




&lt;p&gt;If you have shipped a product, you know the cycle. Brief to designer. Something comes back that does not quite match the brand. Revise. Engineer reinterprets the spec. Revise again. Two weeks later, the thing looks slightly off.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Early adopters described cutting that whole cycle to a single conversation.&lt;/strong&gt; One team reported going from a week-long brief-to-code loop to one session. That is the shift worth unpacking, and it gets lost when Claude Design is compared head-to-head with tools solving different problems.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Do Figma AI, Lovable, and v0 Actually Do?
&lt;/h2&gt;

&lt;p&gt;Each tool has a clear job. The press keeps comparing the wrong jobs.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Figma Make&lt;/strong&gt; (Figma's AI layer): generates designs from prompts inside the Figma canvas. &lt;strong&gt;Starts from the design file.&lt;/strong&gt; &lt;sup id="fnref1"&gt;1&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Lovable&lt;/strong&gt;: turns a plain-language description into a full-stack deployable app. &lt;strong&gt;Starts from an idea.&lt;/strong&gt; &lt;sup id="fnref2"&gt;2&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v0 by Vercel&lt;/strong&gt;: generates React and Tailwind components from prompts. Developer-facing, fast. &lt;strong&gt;Starts from a component need.&lt;/strong&gt; &lt;sup id="fnref2"&gt;2&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Claude Design&lt;/strong&gt;: reads your GitHub repo and generates designs shaped by what is already there. &lt;strong&gt;Starts from your production codebase.&lt;/strong&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Four tools, four starting points.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Does Claude Design Do That Figma, Lovable, and v0 Don't?
&lt;/h2&gt;

&lt;p&gt;When you connect a GitHub repo, Claude Design reads your codebase and extracts: &lt;sup id="fnref3"&gt;3&lt;/sup&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tailwind config:&lt;/strong&gt; your spacing scale, breakpoints, color tokens&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Global CSS:&lt;/strong&gt; your CSS variables, font stacks, base styles&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Font declarations and logo SVGs:&lt;/strong&gt; the visual identity already in your code&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Component names:&lt;/strong&gt; the vocabulary your engineers use&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;What comes out is a design system that already matches what you shipped.&lt;/strong&gt; Not one you configure. The one living in your repo.&lt;/p&gt;

&lt;p&gt;Two designers ran a live side-by-side the day it launched. Same brief to both: redesign a real blog in Readymag style, passed in as a screenshot and a markdown context file. Claude Design produced a layout that tracked the reference. Lovable produced something competent but generic, closer to a WordPress theme than the brand they pointed at. The designer's read: &lt;strong&gt;"designers now can cook."&lt;/strong&gt; Not a replacement, a lever. &lt;sup id="fnref4"&gt;4&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F81ni6cqrtnpc2nqzfunc.webp" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F81ni6cqrtnpc2nqzfunc.webp" alt="Claude Design interface showing prompt-to-prototype with design system extraction" width="800" height="449"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;You build from there. Prompt to prototype. When the prototype is ready, one instruction passes it to &lt;a href="https://monkfrom.earth/blogs/zuckerberg-back-to-coding-claude-code" rel="noopener noreferrer"&gt;Claude Code&lt;/a&gt;, which also reads your codebase. &lt;strong&gt;The loop closes: idea, design, production code, no translation step.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Lovable and v0 aim at different outputs. Lovable gives a greenfield founder a new app. v0 gives a developer a component to paste in. Claude Design gives a team with an existing product something pre-fitted to their repo. &lt;sup id="fnref5"&gt;5&lt;/sup&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Does Starting From Your Codebase Matter?
&lt;/h2&gt;

&lt;p&gt;Different starting points serve different people.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Figma&lt;/strong&gt; treats the design file as the canonical home for a brand. For design teams, that is still true.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Claude Design&lt;/strong&gt; treats the repo as canonical. That fits a different team: one where design intent already lives in Tailwind tokens, CSS variables, and component names.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;This matters most to one person: the engineer or PM extending a live product.&lt;/strong&gt; Not building something new. Not exploring from a blank canvas. Extending what is already there, in a way that matches what is already there.&lt;/p&gt;

&lt;p&gt;For that person, starting from the repo removes a translation step. The output is already shaped by the code it will land in. The other tools are not worse at this. They are aimed elsewhere.&lt;/p&gt;




&lt;p&gt;I post breakdowns like this regularly on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt; and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;. The angle is always what it means for builders, not what the press release says.&lt;/p&gt;




&lt;h2&gt;
  
  
  When Should You Use Each Tool?
&lt;/h2&gt;

&lt;p&gt;Pick by the starting point that matches your job.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Figma&lt;/strong&gt; is the tool for design teams on a shared canvas. Pixel precision, component libraries, review workflows, handoff annotations. Claude Design does none of this. &lt;sup id="fnref5"&gt;5&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Lovable&lt;/strong&gt; is the tool when you have no product yet and want idea to deployed app without code. MVP, internal tool, first prototype. Claude Design is not for that.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v0&lt;/strong&gt; is the tool when you need a React component fast and can edit code. Claude Design is not trying to replace that.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Claude Design is aimed at a specific step:&lt;/strong&gt; you have a live product, a new feature to design, and you need something that already matches everything you built. Teams have always solved this with some combination of briefs, design exploration, review, handoff, and engineering interpretation. Claude Design compresses that into a conversation that starts from the repo. Whether that is the right trade depends on the team.&lt;/p&gt;

&lt;p&gt;The broader pattern is familiar. &lt;a href="https://monkfrom.earth/blogs/zuckerberg-back-to-coding-claude-code" rel="noopener noreferrer"&gt;Zuckerberg returning to the codebase after 20 years using Claude Code&lt;/a&gt; is the same story. So is &lt;a href="https://monkfrom.earth/blogs/karpathy-autoresearch-explained-ml-to-marketing" rel="noopener noreferrer"&gt;Karpathy explaining AI workflows to people who do not write code&lt;/a&gt;. &lt;strong&gt;AI is not replacing the work. It is eliminating the translation layers between people who do different kinds of work.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;One signal worth noting.&lt;/strong&gt; Mike Krieger, Anthropic's CPO and Instagram co-founder, resigned from Figma's board on April 14, three days before Claude Design launched. He had joined less than a year earlier. The resignation was disclosed to the SEC the same day The Information reported Anthropic was building design tools. &lt;sup id="fnref6"&gt;6&lt;/sup&gt; The adjacency was close enough for the board seat to become untenable, even though the two products are aimed at different jobs.&lt;/p&gt;

&lt;p&gt;The market read the adjacency in real time.&lt;/p&gt;


&lt;blockquote&gt;
&lt;p&gt;Anthropic Labs launched Claude Design, a new product for creating visual assets, prototypes, slides, and one-pagers with Claude.&lt;br&gt;&lt;br&gt;It is rolling out in research preview to Pro, Max, Team, and Enterprise users, powered by Claude Opus 4.7.&lt;a href="https://twitter.com/search?q=%24ADBE&amp;amp;src=ctag&amp;amp;ref_src=twsrc%5Etfw" rel="noopener noreferrer"&gt;$ADBE&lt;/a&gt; &lt;a href="https://twitter.com/search?q=%24FIG&amp;amp;src=ctag&amp;amp;ref_src=twsrc%5Etfw" rel="noopener noreferrer"&gt;$FIG&lt;/a&gt; &lt;a href="https://t.co/5u0TOMSqSW" rel="noopener noreferrer"&gt;&lt;/a&gt;&lt;a href="https://t.co/5u0TOMSqSW" rel="noopener noreferrer"&gt;https://t.co/5u0TOMSqSW&lt;/a&gt; &lt;a href="https://t.co/TblMIEJE4u" rel="noopener noreferrer"&gt;pic.twitter.com/TblMIEJE4u&lt;/a&gt;&lt;/p&gt;— Wall St Engine (@wallstengine) &lt;a href="https://twitter.com/wallstengine/status/2045163733203501378?ref_src=twsrc%5Etfw" rel="noopener noreferrer"&gt;April 17, 2026&lt;/a&gt;
&lt;/blockquote&gt; 
&lt;h2&gt;
  
  
  What Are Claude Design's Limitations Right Now?
&lt;/h2&gt;

&lt;p&gt;Claude Design is a research preview as of April 2026. Real constraints worth knowing before you try it:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;No multiplayer.&lt;/strong&gt; For a design team on a shared canvas, Figma still wins cleanly. &lt;sup id="fnref5"&gt;5&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Token burn is heavy.&lt;/strong&gt; Claude Design runs on Opus 4.7 and is metered separately from your chat and Claude Code usage. Pro is described as "quick explorations, one-off use." One user reported two design sessions consuming 58% of their weekly Pro allowance. &lt;sup id="fnref7"&gt;7&lt;/sup&gt; To use it regularly, you need Max.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Prototyping-level output, not production polish.&lt;/strong&gt; The design system extraction makes things brand-consistent, but it is not a replacement for a designer's eye on the final layer.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Export options are practical&lt;/strong&gt; but limited: PDF, PPTX, standalone HTML, Canva. &lt;sup id="fnref8"&gt;8&lt;/sup&gt; The HTML export is also how the Claude Code handoff closes the loop. Anthropic's own ecosystem, end to end.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  A 5-Question Claude Design Readiness Check
&lt;/h2&gt;

&lt;p&gt;Before you open it, ask these. If you answer yes to three or more, Claude Design fits your workflow today. If not, Figma, Lovable, or v0 is probably the better tool for the job.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Do you already have a shipped product in a GitHub repo?&lt;/strong&gt; Claude Design starts from code that exists. No repo, no extraction.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Is your design system encoded in Tailwind config, CSS variables, or component names?&lt;/strong&gt; That is what the extractor reads. Design tokens locked in a Figma file alone will not transfer.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Are you extending an existing product rather than starting from zero?&lt;/strong&gt; The tool's edge is fit to what is already there. For greenfield work, Lovable or v0 is closer to the job.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Can one person own the design-to-code loop, or does it need multiplayer?&lt;/strong&gt; No shared canvas. If three designers need to work on the same file, Figma still wins.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Are you on Max, or willing to rate-limit yourself on Pro?&lt;/strong&gt; Two sessions burned 58% of a weekly Pro allowance. Regular use needs Max.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Three or more yeses and the translation step this tool removes is a real one for you.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Figma, Lovable, v0, and Claude Design &lt;strong&gt;pick different starting points.&lt;/strong&gt; Different starting points, different jobs.&lt;/li&gt;
&lt;li&gt;Figma treats the &lt;strong&gt;design file as canonical.&lt;/strong&gt; Claude Design treats the &lt;strong&gt;codebase as canonical.&lt;/strong&gt; Neither is wrong; they suit different teams.&lt;/li&gt;
&lt;li&gt;Claude Design's design system extraction reads your Tailwind, CSS, and component names to generate &lt;strong&gt;on-brand output from the first prompt&lt;/strong&gt;, without manual configuration.&lt;/li&gt;
&lt;li&gt;Each tool fits a different starting point: &lt;strong&gt;Figma for collaborative design work, Lovable for greenfield apps, v0 for quick components, Claude Design for extending an existing codebase.&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Token burn is real.&lt;/strong&gt; Claude Design is metered separately. Pro is for one-off use. Regular use requires Max.&lt;/li&gt;
&lt;li&gt;Anthropic's CPO &lt;strong&gt;resigned from Figma's board three days before launch.&lt;/strong&gt; Figma's stock dropped 5 to 7% on launch day, a read of the adjacency, not a verdict on either product.&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;Shipping something where this trade-off matters and want a second read on it? &lt;a href="mailto:hi@monkfrom.earth?subject=Claude%20Design%20take"&gt;Get in touch&lt;/a&gt;. I reply to every thoughtful email.&lt;/p&gt;

&lt;p&gt;I post builder-first takes on AI tooling on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;. The kind that skip the hype and go straight to what changes for people who ship. If that is useful, a follow goes a long way.&lt;/p&gt;







&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://www.figma.com/make/" rel="noopener noreferrer"&gt;Figma Make: AI-powered design tools&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://www.nocode.mba/articles/lovable-vs-V0" rel="noopener noreferrer"&gt;Lovable vs v0: Which AI Builder Is Better? nocode.mba&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;&lt;a href="https://support.claude.com/en/articles/14604397-set-up-your-design-system-in-claude-design" rel="noopener noreferrer"&gt;Set up your design system in Claude Design, Claude Help Center&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=o4jIKc_DIoM" rel="noopener noreferrer"&gt;Claude Design vs Lovable: live side-by-side comparison, YouTube&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn5"&gt;
&lt;p&gt;&lt;a href="https://www.eigent.ai/blog/claude-design-vs-figma-make" rel="noopener noreferrer"&gt;Claude Design vs Figma Make: 2026 AI Design Tool Comparison, eigent.ai&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn6"&gt;
&lt;p&gt;&lt;a href="https://techcrunch.com/2026/04/16/anthropic-cpo-leaves-figmas-board-after-reports-he-will-offer-a-competing-product/" rel="noopener noreferrer"&gt;Anthropic CPO leaves Figma's board after reports he will offer a competing product, TechCrunch&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn7"&gt;
&lt;p&gt;&lt;a href="https://www.testingcatalog.com/anthropic-launches-claude-design-ai-tool-for-paid-plans/" rel="noopener noreferrer"&gt;Anthropic launches Claude Design AI tool for paid plans, Testing Catalog&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn8"&gt;
&lt;p&gt;&lt;a href="https://www.anthropic.com/news/claude-design-anthropic-labs" rel="noopener noreferrer"&gt;Introducing Claude Design by Anthropic Labs&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>design</category>
      <category>ui</category>
      <category>ux</category>
    </item>
    <item>
      <title>Meta Muse Spark: What Meta Is Actually Betting On</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Fri, 17 Apr 2026 12:33:55 +0000</pubDate>
      <link>https://forem.com/monkfromearth/meta-muse-spark-what-meta-is-actually-betting-on-1794</link>
      <guid>https://forem.com/monkfromearth/meta-muse-spark-what-meta-is-actually-betting-on-1794</guid>
      <description>&lt;p&gt;&lt;strong&gt;TL;DR:&lt;/strong&gt; Meta launched Muse Spark on April 8, 2026. Most commentary split into two camps. Meta went closed because Meta won. Meta went closed because Meta lost. Both miss what Meta actually built. Muse Spark does frontier-class reasoning in &lt;strong&gt;less than half the output tokens&lt;/strong&gt; Claude Opus 4.6 and GPT-5.4 spend on the same benchmark, and Meta AI, the product serving roughly &lt;strong&gt;three billion daily active users&lt;/strong&gt;, runs on it. Read Muse Spark as an efficiency-first, patiently sequenced, consumer-scale bet, and the choices that look strange on their own start fitting together.&lt;/p&gt;

&lt;p&gt;The week Muse Spark launched, the conversation split almost immediately. One camp said Meta finally caught up and closed the doors. Another said Meta finally fell behind and is hiding it. Both sides were arguing about the license. Neither was arguing about the model.&lt;/p&gt;

&lt;p&gt;The bet Meta actually made isn't captured by the license. It's captured by three choices that are easy to miss through the open-weights lens. Muse Spark is designed for &lt;strong&gt;fewer tokens per query&lt;/strong&gt;. It is framed as &lt;strong&gt;step one of a long sequence&lt;/strong&gt;. And it is shipping first as the engine of a consumer product reaching &lt;strong&gt;three billion daily active users&lt;/strong&gt;. Those three choices, taken together, describe a different game than the one most labs are playing.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Muse Spark is
&lt;/h2&gt;

&lt;p&gt;Muse Spark is Meta Superintelligence Labs' first model, shipped April 8 after a nine-month rebuild of Meta's AI infrastructure. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; It is a natively multimodal reasoning model with three modes. Instant for fast responses. Thinking for reasoning-heavy queries. Contemplating, positioned against Gemini Deep Think and GPT Pro for long scientific work. It supports tool use, visual chain of thought, and multi-agent orchestration. &lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Meta AI, the consumer product on meta.ai and the Meta AI app, runs on it today. The Muse Spark API is in private preview for selected partners. Alexandr Wang, Meta's Chief AI Officer, has said broader API access is coming. &lt;sup id="fnref3"&gt;3&lt;/sup&gt; The weights have not been released, and Meta has not committed to whether or when they will be.&lt;/p&gt;

&lt;p&gt;On the Artificial Analysis Intelligence Index v4.0, Muse Spark scores 52. GPT-5.4 and Gemini 3.1 Pro Preview score 57. Claude Opus 4.6 scores 53. &lt;sup id="fnref4"&gt;4&lt;/sup&gt; Fourth at the frontier, as the frontier is currently measured.&lt;/p&gt;

&lt;h2&gt;
  
  
  Efficiency is the number that matters
&lt;/h2&gt;

&lt;p&gt;Meta's headline technical claim is that Muse Spark reaches its capabilities with over an order of magnitude less compute than Llama 4 Maverick, the prior Meta flagship. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; That is a training-side claim. The more interesting number sits on the inference side.&lt;/p&gt;

&lt;p&gt;To complete the Artificial Analysis Intelligence Index v4.0 run, Muse Spark used &lt;strong&gt;58 million output tokens&lt;/strong&gt;. Claude Opus 4.6 used &lt;strong&gt;157 million&lt;/strong&gt;. GPT-5.4 used &lt;strong&gt;120 million&lt;/strong&gt;. &lt;sup id="fnref4"&gt;4&lt;/sup&gt; Muse Spark reaches roughly the same tier of performance while spending less than half the thinking time of its closest competitors.&lt;/p&gt;

&lt;p&gt;Meta calls the mechanism &lt;strong&gt;thought compression&lt;/strong&gt;. During reinforcement learning, the model is penalized for excessive reasoning tokens. It is trained to reach the same answer with fewer intermediate steps. &lt;sup id="fnref4"&gt;4&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Zoom out. Llama 4 Maverick scored &lt;strong&gt;18&lt;/strong&gt; on the same index. Muse Spark scores &lt;strong&gt;52&lt;/strong&gt;. &lt;sup id="fnref4"&gt;4&lt;/sup&gt; &lt;strong&gt;Nearly 3x jump&lt;/strong&gt; in one release, using roughly a tenth of the training compute, producing a model that serves answers in less than half the output tokens of its peers. That is not a fourth-place story. It is a different-axis story.&lt;/p&gt;

&lt;p&gt;Thought compression isn't the only lever. Fei Xia, a Meta researcher, showed Muse Spark tackling a hard visual counting task using parallel subagents: divide the image into a grid, assign a subagent per tile, merge the counts. &lt;sup id="fnref5"&gt;5&lt;/sup&gt; That is a second axis of test-compute scaling. Not fewer tokens per query, but many smaller queries instead of one large one. Both compound efficiency at inference time.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhve0hn6bw8c8pt0pztvr.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhve0hn6bw8c8pt0pztvr.jpg" alt="Fei Xia, a Meta researcher, used Muse Spark to count birds in a dense flock. The model divided the image into a 4x4 grid, ran parallel subagents per tile, and returned per-tile counts of 24, 50, 46, 11 across the top row, summing to 431 across the frame. A note flags the count as a conservative lower bound because of overlapping birds and sub-threshold specks." width="800" height="530"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Matt Ridley, in &lt;em&gt;How Innovation Works&lt;/em&gt;, argues that real technological progress almost never looks like a breakthrough in the moment. It looks like &lt;strong&gt;compounded efficiency&lt;/strong&gt;. &lt;sup id="fnref6"&gt;6&lt;/sup&gt; The Wright brothers didn't fly higher than their competitors; they iterated longer. Meta's claim with Muse Spark is that the same mechanism is back in large language models as the active design constraint. Fewer tokens per query, optimized over releases, compounded.&lt;/p&gt;

&lt;p&gt;Under the efficiency thesis, the contribution is the training recipe, not the weights. The productized result at three billion DAUs is what the recipe is for.&lt;/p&gt;

&lt;h2&gt;
  
  
  Patience as a structural choice
&lt;/h2&gt;

&lt;p&gt;Wang's launch thread called Muse Spark &lt;strong&gt;"step one."&lt;/strong&gt; &lt;sup id="fnref3"&gt;3&lt;/sup&gt; Meta has named three modes, shipped two of them, and placed Contemplating on a published roadmap. The release itself followed a &lt;strong&gt;nine-month rebuild&lt;/strong&gt; of Meta's internal AI infrastructure before any new model went out. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;That pattern is uncommon. Labs announce quarterly, deprecate on shorter cycles, and trade nomenclature every six weeks. A frontier lab committing to a staged ladder with named but unbuilt later steps is the exception.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frrphsou62v1tyd0t033t.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frrphsou62v1tyd0t033t.png" alt="Muse Spark Contemplating mode benchmark table from Meta's launch. On Humanity's Last Exam (No Tools), Muse Spark scores 50.2, Gemini 3.1 Deep Think 48.4, GPT 5.4 Pro 43.9. With Tools: 58.4, 53.4, 58.7. IPhO 2025 Theory: 82.6, 87.7, 93.5. FrontierScience Research: 38.3, 23.3, 36.7. The ladder's later rungs already produce numbers that compete with Deep Think and GPT Pro." width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Jeff Bezos's 1997 shareholder letter made a version of this argument on its own: "We will continue to make investment decisions in light of long-term market leadership considerations rather than short-term profitability considerations." &lt;sup id="fnref7"&gt;7&lt;/sup&gt; Most companies quote the line. Very few behave like it. Muse Spark is Meta behaving like it. A nine-month silence, a named sequence, an efficiency-first architecture that only pays back at scale.&lt;/p&gt;

&lt;p&gt;Patience has a failure mode. If the ladder breaks, the gap widens. If competitors keep improving quarterly and Muse Spark's step two arrives in 2027, the index score will read worse, not better. That is the actual risk of the strategy. Not the license. The cadence.&lt;/p&gt;

&lt;h2&gt;
  
  
  The game Meta is actually playing
&lt;/h2&gt;

&lt;p&gt;Roughly three billion daily active users touch Meta's products. Muse Spark powers Meta AI across them. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; Every prompt, every caption suggestion, every smart reply, every image generation across meta.ai, Instagram, WhatsApp, and Facebook is a query served at Meta's cost.&lt;/p&gt;

&lt;p&gt;Reread the efficiency numbers with that denominator. 58 million output tokens per benchmark run is interesting when you run one benchmark. It is structural when you run hundreds of billions of inferences. Cutting thinking time by more than half is how inference economics actually move at Meta's scale.&lt;/p&gt;

&lt;p&gt;The API is a secondary product. The primary product is a feature inside applications people already use. That framing answers most of the questions that the closed-weights decision seems to raise:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Why closed:&lt;/strong&gt; weight distribution gives up the only part that is uniquely Meta, which is distribution plus efficient inference under Meta's control.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Why efficiency-first:&lt;/strong&gt; cost-per-query is the load-bearing variable at three billion users.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Why fourth on the index:&lt;/strong&gt; the index measures capability, not capability per dollar of inference. Meta is not optimizing for the thing the index measures.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Why patience:&lt;/strong&gt; product cycles at Meta's scale run in quarters and years, not weeks. A staged ladder matches the cadence of the products that will ship the model.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;OpenAI, Anthropic, and Google primarily sell access. Meta does not. Meta bundles. A closed, efficient model embedded in consumer distribution is a product shape no other frontier lab has a direct answer to right now.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Muse Spark bets against
&lt;/h2&gt;

&lt;p&gt;Muse Spark bets against three premises that have held in AI for three years. That benchmark rank drives strategic outcomes. That fast iteration beats staged iteration. That serving the weights is the dominant form of distribution.&lt;/p&gt;

&lt;p&gt;If Meta is right, competitors re-architect. Expect tokens-per-benchmark to become a reported number. Expect ladder-style release roadmaps. Expect fewer labs selling raw access and more labs selling integrated products.&lt;/p&gt;

&lt;p&gt;If Meta is wrong, Muse Spark stays fourth on the index, the efficiency claim gets normalized by competitors' next releases, and the Scale-era thesis fades into another nine-month rebuild.&lt;/p&gt;

&lt;p&gt;Deedy, in a popular thread after launch, called Muse Spark's reasoning "solid but not best in class." &lt;sup id="fnref5"&gt;5&lt;/sup&gt; That read is fair if you are benchmarking reasoning. It is beside the point if you are measuring how to serve reasoning to three billion people.&lt;/p&gt;

&lt;h2&gt;
  
  
  Takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Efficiency is the headline, not the license.&lt;/strong&gt; Muse Spark uses 58 million output tokens where Claude Opus 4.6 uses 157 million on the same evaluation. &lt;sup id="fnref4"&gt;4&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Training efficiency is roughly ten times Llama 4 Maverick.&lt;/strong&gt; The index score nearly tripled in one release. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; &lt;sup id="fnref4"&gt;4&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Patience is the structural bet.&lt;/strong&gt; A nine-month rebuild, a three-mode ladder, a second-step roadmap that is named but not shipped. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; &lt;sup id="fnref3"&gt;3&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Specialization explains the choices.&lt;/strong&gt; Meta AI reaches three billion DAUs, and inference economics at that scale reward the token per query being low, not the leaderboard rank being high. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The license is a symptom of the strategy.&lt;/strong&gt; If efficiency plus distribution plus patience is the bet, releasing the weights gives the bet away.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I've been writing about &lt;a href="https://monkfrom.earth/blogs/good-products-hard-to-vary" rel="noopener noreferrer"&gt;how constraints shape design, not features&lt;/a&gt; for a while, and Muse Spark is a useful instance of the pattern. The interesting move in AI this year might not be the model that scores higher. It might be the model that answers in fewer tokens and ships inside an application a billion people already open every day.&lt;/p&gt;

&lt;p&gt;I break things like this down on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;. Usually shorter, sometimes as carousels. If this read resonated, you'd probably like those.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources
&lt;/h2&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://ai.meta.com/blog/introducing-muse-spark-msl/" rel="noopener noreferrer"&gt;Meta AI, "Introducing Muse Spark"&lt;/a&gt;, April 8, 2026 ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://simonwillison.net/2026/Apr/8/muse-spark/" rel="noopener noreferrer"&gt;Simon Willison, "Meta's new model is Muse Spark"&lt;/a&gt;, April 8, 2026 ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;&lt;a href="https://x.com/alexandr_wang/status/2041909376508985381" rel="noopener noreferrer"&gt;Alexandr Wang on X, launch thread and API update&lt;/a&gt;, April 2026 ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://www.datacamp.com/blog/muse-spark" rel="noopener noreferrer"&gt;Muse Spark: Features, Benchmarks, and How to Use It, DataCamp&lt;/a&gt;, April 2026 ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn5"&gt;
&lt;p&gt;&lt;a href="https://x.com/xf1280/status/2043730980264128673" rel="noopener noreferrer"&gt;Fei Xia and Deedy Das on Muse Spark capabilities (thread)&lt;/a&gt;, April 13, 2026 ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn6"&gt;
&lt;p&gt;Matt Ridley, &lt;em&gt;How Innovation Works: And Why It Flourishes in Freedom&lt;/em&gt; (HarperCollins, 2020) ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn7"&gt;
&lt;p&gt;&lt;a href="https://www.sec.gov/Archives/edgar/data/1018724/000119312513151836/d511111dex991.htm" rel="noopener noreferrer"&gt;Jeff Bezos, 1997 Letter to Shareholders&lt;/a&gt;, Amazon ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>agents</category>
      <category>opensource</category>
    </item>
    <item>
      <title>GPT-5.4-Cyber explained: OpenAI's cyber-only AI</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Wed, 15 Apr 2026 08:33:55 +0000</pubDate>
      <link>https://forem.com/monkfromearth/gpt-54-cyber-explained-openais-cyber-only-ai-1nhn</link>
      <guid>https://forem.com/monkfromearth/gpt-54-cyber-explained-openais-cyber-only-ai-1nhn</guid>
      <description>&lt;p&gt;Two days ago I wrote about &lt;a href="https://monkfrom.earth/blogs/claude-mythos-autonomous-cyberattack" rel="noopener noreferrer"&gt;Claude Mythos completing AISI's 32-step cyberattack chain end-to-end&lt;/a&gt;. On April 14, OpenAI put out the clearest signal yet that the labs are reading the same capability curve and building the defender track in advance.&lt;/p&gt;

&lt;p&gt;They announced &lt;strong&gt;GPT-5.4-Cyber&lt;/strong&gt;, a version of GPT-5.4 fine-tuned to be "cyber-permissive," and scaled up their &lt;strong&gt;Trusted Access for Cyber (TAC)&lt;/strong&gt; program to thousands of verified individual defenders and hundreds of teams defending critical software.&lt;sup id="fnref1"&gt;1&lt;/sup&gt; In their own words, this is shipping "in preparation for increasingly more capable models over the next few months."&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TL;DR.&lt;/strong&gt; This is defender tooling shipped before the next capability jump, not after. The model is the headline. The real story is a fine-tuned permissive variant named, tiered, and published as a product.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa1p49fy2jdmp8yhcz84a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa1p49fy2jdmp8yhcz84a.png" alt="OpenAI's April 14, 2026 announcement: Trusted access for the next era of cyber defense" width="800" height="352"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Primary source: &lt;a href="https://openai.com/index/scaling-trusted-access-for-cyber-defense/" rel="noopener noreferrer"&gt;OpenAI on scaling trusted access for cyber defense&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Does GPT-5.4-Cyber Actually Unlock for Defenders?
&lt;/h2&gt;

&lt;p&gt;Same base model as GPT-5.4, different refusal boundary. OpenAI's description: a model that "lowers the refusal boundary for legitimate cybersecurity work" and adds capabilities like binary reverse engineering. It can analyze compiled software for malware, vulnerabilities, and robustness without access to source code.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Binary reverse engineering is the concrete unlock, and it is not small. It is one of the highest-leverage things a defender can automate, and it is exactly the kind of request that trips every refusal classifier ever built. The same prompt from a malicious actor yields the same output. The model cannot tell them apart. The verification layer can.&lt;/p&gt;

&lt;p&gt;Everything else in the envelope is less dramatic but more useful at scale. Vulnerability research without the hedging. Security education that answers the question instead of warning about it. Defensive programming help that does not refuse to describe the attack it is trying to prevent.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Was Refusal Always a Bad Safeguard?
&lt;/h2&gt;

&lt;p&gt;For three years, the default safety move has been to push risk into the model through refusal training. It was the cheapest thing to ship and the easiest thing to measure. It also quietly assumed attackers and defenders use the same tool, so making the tool worse would hurt both evenly.&lt;/p&gt;

&lt;p&gt;That assumption was always wrong. Attackers run local models, jailbroken models, and purpose-built tooling. Refusals mostly tax the defenders trying to follow the rules.&lt;/p&gt;

&lt;p&gt;GPT-5.4 (classified "high" cyber capability under OpenAI's Preparedness Framework) keeps its refusal boundary for the public. The permissive variant ships only to people who have agreed to be identified. This is closer to how physical-world dual-use actually works. Pharmacies stock dangerous drugs behind an identity check, not behind a refusal. Labs buy restricted reagents with a license. The safeguard is not the molecule. It is the paperwork.&lt;/p&gt;

&lt;h2&gt;
  
  
  GPT-5.4-Cyber and the Mythos Parallel
&lt;/h2&gt;

&lt;p&gt;My last three posts on Claude Mythos describe the same shape from different angles. &lt;a href="https://monkfrom.earth/blogs/claude-mythos-system-card" rel="noopener noreferrer"&gt;The system card&lt;/a&gt; showed a model with enough situational awareness to conceal its own actions. &lt;a href="https://monkfrom.earth/blogs/anthropic-glasswing-ai-cybersecurity" rel="noopener noreferrer"&gt;Project Glasswing&lt;/a&gt; showed the same model finding thousands of zero-days in critical open-source infrastructure. The &lt;a href="https://monkfrom.earth/blogs/claude-mythos-autonomous-cyberattack" rel="noopener noreferrer"&gt;AISI cyber range&lt;/a&gt; showed it running a full 32-step autonomous cyberattack. Mythos itself is gated. Anthropic ships it only through its own trust program.&lt;/p&gt;

&lt;p&gt;So both frontier labs already operate the same model: dual-use capability behind verified access. What is new with GPT-5.4-Cyber is that OpenAI is the first to take the defender side of that model and publish it as a product tier: a named, fine-tuned, cyber-permissive variant with its own enrollment path and its own preparedness designation. Anthropic's gating is a policy. OpenAI's is a SKU.&lt;/p&gt;

&lt;p&gt;You can see the same bet in the numbers they quietly dropped in the same post. Codex Security has contributed to &lt;strong&gt;over 3,000 critical and high vulnerability fixes&lt;/strong&gt; since launch. Codex for Open Source has reached &lt;strong&gt;more than 1,000 open source projects&lt;/strong&gt;. The &lt;strong&gt;$10M Cybersecurity Grant Program&lt;/strong&gt; keeps funding defender tooling.&lt;sup id="fnref1"&gt;1&lt;/sup&gt; In the Mythos cyberattack post I wrote: &lt;em&gt;"I'd bet on it eventually, but 'eventually' and 'right now' are different things in security."&lt;/em&gt; This is a lab betting "right now," on the defender side, and betting it visibly.&lt;/p&gt;

&lt;h2&gt;
  
  
  Who Verifies the Verifier?
&lt;/h2&gt;

&lt;p&gt;The uncomfortable follow-up to any identity-gated safeguard. OpenAI is now the identity layer for a meaningful slice of the security industry. Every defender applying for the permissive tier is trusting one company's KYC pipeline to decide who counts as a defender, and trusting OpenAI's interpretation of "legitimate use" to hold up over time.&lt;/p&gt;

&lt;p&gt;This is the part of the announcement I would most want to see discussed over the next few weeks. It is also the part nobody will discuss, because the new model is shinier than the policy question behind it.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;GPT-5.4-Cyber&lt;/strong&gt; is a fine-tuned GPT-5.4 with fewer capability restrictions, shipped only to vetted defenders under the Trusted Access for Cyber program.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Preemptive, not reactive.&lt;/strong&gt; OpenAI is shipping this ahead of more capable base models coming in the next few months, in their own words.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Both labs already gate dual-use.&lt;/strong&gt; Mythos is restricted through Anthropic's trust program. What is new is OpenAI naming a fine-tuned permissive variant as a product tier.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Open question:&lt;/strong&gt; who audits the identity layer when OpenAI and Anthropic become the KYC gate for a chunk of the security industry?&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;I break down AI safety and capability stories on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;. If this resonated, you would probably like those too.&lt;/p&gt;







&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://openai.com/index/scaling-trusted-access-for-cyber-defense/" rel="noopener noreferrer"&gt;OpenAI on scaling trusted access for cyber defense (April 14, 2026)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>security</category>
      <category>webdev</category>
    </item>
    <item>
      <title>Claude Mythos Is the First AI to Complete a Full Corporate Cyberattack End-to-End</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Mon, 13 Apr 2026 17:35:25 +0000</pubDate>
      <link>https://forem.com/monkfromearth/claude-mythos-is-the-first-ai-to-complete-a-full-corporate-cyberattack-end-to-end-3mk5</link>
      <guid>https://forem.com/monkfromearth/claude-mythos-is-the-first-ai-to-complete-a-full-corporate-cyberattack-end-to-end-3mk5</guid>
      <description>&lt;p&gt;The UK's AI Security Institute confirmed this week that Claude Mythos, an Anthropic model, became the first AI to complete their cyber range end-to-end.&lt;sup id="fnref1"&gt;1&lt;/sup&gt; The range is a &lt;strong&gt;32-step corporate network attack&lt;/strong&gt; scenario. Human experts estimate the same attack would take them &lt;strong&gt;20 hours&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;The institute's recommendation to organizations: keep your software updated. Use access controls. Enable logging.&lt;/p&gt;

&lt;p&gt;The gap between those two sentences is the part of this story I keep returning to.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TL;DR:&lt;/strong&gt; Claude Mythos ran a full autonomous cyberattack, 32 steps, end-to-end, in a scenario that takes human experts 20 hours. It is the first AI to complete AISI's cyber range. The official response was to recommend basic security hygiene. The mismatch between the capability and the response is where the real story lives.&lt;/p&gt;

&lt;h2&gt;
  
  
  How Did AI Go From Basic Cyber Tasks to a Full Autonomous Cyberattack?
&lt;/h2&gt;

&lt;p&gt;Self-driving cars give me the cleanest parallel here.&lt;/p&gt;

&lt;p&gt;For a decade, every individual piece of the self-driving puzzle existed as a demo. Lane-keeping worked. Adaptive cruise worked. Automated parking worked. What didn't exist, for years, was the full ride. Door to door, no human touching the wheel. When Waymo's first commercial robotaxi picked up a passenger in 2020, what changed wasn't the individual capabilities. It was the threshold: chaining all of them into one uninterrupted ride.&lt;/p&gt;

&lt;p&gt;The same thing just happened in offensive cybersecurity.&lt;/p&gt;

&lt;p&gt;Each step of a network attack has been within reach of AI models for a while. Reconnaissance. Crafting payloads. Pivoting through a subnet. Covering tracks. What didn't exist was a model that could chain all 32 of those steps together without a human stepping in between. Claude Mythos did.&lt;/p&gt;

&lt;p&gt;In 2023, leading AI models struggled with basic cybersecurity tasks. Not sophisticated ones. Basic ones. Three years later, one of them drove the entire route.&lt;/p&gt;

&lt;p&gt;AISI published the actual curve, and it is worth looking at directly.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcgraaer1ehytkwzcembp.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcgraaer1ehytkwzcembp.jpeg" alt="AISI evaluation showing average steps completed on 'The Last Ones' cyber range per spent tokens. Claude Mythos Preview reaches around 22 steps on average and a maximum of roughly 32, clearly above Claude Opus 4.6, GPT-5.4, GPT-5.3 Codex, Claude Opus 4.5, Claude Sonnet 4.5, and GPT-4o." width="800" height="500"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The red line is Mythos. GPT-4o sits near the bottom, completing around three steps before running out. Sonnet 4.5 gets to roughly 11. Opus 4.5 and the GPT-5 family cluster in the mid teens. Opus 4.6 pushes past 16. Mythos is the only line that clears the middle milestones: C2 reverse engineering, advanced persistence, infrastructure compromise, and eventually M9 — "Full network takeover."&lt;sup id="fnref1"&gt;1&lt;/sup&gt; The shape of that curve is what "first AI to complete the range end-to-end" actually looks like.&lt;/p&gt;

&lt;p&gt;AISI is careful about the current scope. The capability applies to "small, weakly defended, and vulnerable systems" given network access. Think of it as the robotaxi that only works on mapped, sunny, well-marked urban grids. Hardened enterprise infrastructure with proper controls is still a different problem, the same way a snowy mountain pass is still a different problem for Waymo.&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;The trajectory is what matters. 2023 to 2026 is three years.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Does an Autonomous Cyberattack Change the Security Equation?
&lt;/h2&gt;

&lt;p&gt;The asymmetry in security has always been simple: attackers need to find one gap, defenders need to close every door.&lt;/p&gt;

&lt;p&gt;AI doesn't change that asymmetry. It changes the cost of running an attack. An automated system doesn't need domain expertise to chain 32 steps. It doesn't get tired halfway through. It doesn't hesitate at unfamiliar territory.&lt;/p&gt;

&lt;p&gt;What previously required a skilled adversary with deep knowledge, time, and custom tools now requires API access and a goal.&lt;/p&gt;

&lt;p&gt;The same model AISI tested on offense has been used defensively in &lt;a href="https://dev.to/blogs/anthropic-glasswing-ai-cybersecurity"&gt;Anthropic's Project Glasswing&lt;/a&gt; to find thousands of zero-days in critical open-source infrastructure. Offense and defense, same capability, same model. The dual-use nature isn't incidental. It's structural. Whoever has the model has both sides.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Should Organizations Do After Claude Mythos Ran a Full Cyberattack?
&lt;/h2&gt;

&lt;p&gt;Patch your systems. Use MFA. Enable logging. AISI's recommendations are correct.&lt;/p&gt;

&lt;p&gt;But they were correct before this evaluation too. That's the part I can't get past.&lt;/p&gt;

&lt;p&gt;These recommendations address the baseline: opportunistic attackers, misconfigured systems, low-skill adversaries. They don't address the shift in assumption that happens when a fully autonomous cyberattack chain becomes possible. Hygiene is still necessary. It is no longer sufficient as a strategy.&lt;/p&gt;

&lt;p&gt;AISI published a joint piece with the UK's National Cyber Security Centre on preparing defenders for frontier AI systems.&lt;sup id="fnref1"&gt;1&lt;/sup&gt; That collaboration exists because the people closest to this problem know the defensive tooling gap is real. The open question is whether the defensive side of AI moves as fast as the offensive side. I'd bet on it eventually, but "eventually" and "right now" are different things in security.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Does the Claude Mythos Evaluation Pattern Reveal?
&lt;/h2&gt;

&lt;p&gt;This is the third notable evaluation result for Claude Mythos in April alone. &lt;a href="https://dev.to/blogs/claude-mythos-system-card"&gt;The system card&lt;/a&gt; showed a model with enough situational awareness to conceal its own actions. Project Glasswing showed it finding thousands of vulnerabilities in critical infrastructure. The AISI cyber range shows it running a full autonomous cyberattack.&lt;/p&gt;

&lt;p&gt;These aren't contradictions. They are the same underlying capability applied in different contexts. A model capable enough for complex multi-step reasoning is capable enough to create real problems at scale.&lt;/p&gt;

&lt;p&gt;The value of these evaluations is that they name what's happening before it becomes a crisis, even when the recommendations that follow don't match the scale of what was just described. Naming it first is not nothing.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Claude Mythos&lt;/strong&gt; became the first AI to complete a &lt;strong&gt;32-step corporate cyberattack&lt;/strong&gt; chain end-to-end in AISI's cyber range&lt;/li&gt;
&lt;li&gt;Human experts estimate the same operation takes &lt;strong&gt;20 hours&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;In 2023, leading models couldn't complete basic cybersecurity tasks. Three years later, one completed a full autonomous cyberattack&lt;/li&gt;
&lt;li&gt;Current capability is scoped to "small, weakly defended" systems, not enterprise infrastructure with proper controls&lt;/li&gt;
&lt;li&gt;The trajectory matters more than the current benchmark: three years of rapid progress, with no signs of slowing&lt;/li&gt;
&lt;li&gt;AISI's defensive recommendations (patch, use MFA, enable logging) are correct but baseline — they predate this evaluation&lt;/li&gt;
&lt;li&gt;AISI and the UK NCSC published joint guidance on preparing defenders for frontier AI systems&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;I break down things like this on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt; — usually shorter, sometimes as carousels. If this resonated, you'd probably like those too.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources
&lt;/h2&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://x.com/AISecurityInst/status/2043683577594794183" rel="noopener noreferrer"&gt;AI Security Institute (@AISecurityInst) — Claude Mythos cyber range evaluation, April 13, 2026&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>security</category>
      <category>ai</category>
      <category>architecture</category>
      <category>news</category>
    </item>
    <item>
      <title>Zuckerberg Is Writing Code Again. With Claude Code.</title>
      <dc:creator>Sameer Khan</dc:creator>
      <pubDate>Sun, 05 Apr 2026 10:24:18 +0000</pubDate>
      <link>https://forem.com/monkfromearth/zuckerberg-is-writing-code-again-with-claude-code-26b1</link>
      <guid>https://forem.com/monkfromearth/zuckerberg-is-writing-code-again-with-claude-code-26b1</guid>
      <description>&lt;p&gt;&lt;strong&gt;TL;DR:&lt;/strong&gt; Mark Zuckerberg shipped 3 diffs to Meta's monorepo last month, his first code in 20 years. He's a heavy user of Claude Code CLI. One of his diffs got 200+ approvals from engineers who wanted to say they reviewed the CEO's code. He's not the only one. Garry Tan at Y Combinator is doing the same thing. The pattern is clear: AI coding tools are pulling founders back into the codebase.&lt;/p&gt;




&lt;h2&gt;
  
  
  What happened?
&lt;/h2&gt;

&lt;p&gt;Gergely Orosz at The Pragmatic Engineer &lt;a href="https://newsletter.pragmaticengineer.com/p/the-pulse-industry-leaders-return" rel="noopener noreferrer"&gt;reported this week&lt;/a&gt; that Mark Zuckerberg is back to writing code. Three diffs landed in Meta's monorepo in March 2026. His tool of choice: &lt;strong&gt;Claude Code CLI&lt;/strong&gt;, Anthropic's terminal-based AI coding assistant. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;To put the scale in perspective: Meta's monorepo now has &lt;strong&gt;close to 100 million diffs&lt;/strong&gt;. Back in 2006, the entire Facebook codebase had fewer than 10,000. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Zuckerberg's last meaningful code contributions were in 2006. That's a 20-year gap. The fact that he's back, and using an AI tool to do it, says something about where we are.&lt;/p&gt;

&lt;h2&gt;
  
  
  The 2010 diff that got force-merged
&lt;/h2&gt;

&lt;p&gt;This isn't Zuckerberg's first time making waves in code review.&lt;/p&gt;

&lt;p&gt;In 2010, he submitted a diff that made profile photos clickable on the profile page. Michael Novati, a senior engineer who would become the first person to hold Meta's L7 "coding machine" archetype, &lt;a href="https://newsletter.pragmaticengineer.com/p/the-coding-machine-at-meta" rel="noopener noreferrer"&gt;blocked it&lt;/a&gt;. The reason: formatting issues everywhere. &lt;sup id="fnref1"&gt;1&lt;/sup&gt; &lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Zuckerberg overrode the block and force-merged it. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Novati spent eight years at Meta and was recognized as the top code committer company-wide for several of them. The Pragmatic Engineer did &lt;a href="https://newsletter.pragmaticengineer.com/p/the-coding-machine-at-meta" rel="noopener noreferrer"&gt;a full episode&lt;/a&gt; with him about what it means to be a "coding machine" at that scale. &lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;The 2010 story is funny in hindsight. But the 2026 version is different. This time, Zuckerberg isn't force-merging past reviewers. He's using AI to write code that engineers actually want to approve. &lt;strong&gt;One of his March diffs got more than 200 approvals&lt;/strong&gt;, with devs jumping at the chance to say they'd reviewed the CEO's work. &lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why this matters beyond the anecdote
&lt;/h2&gt;

&lt;p&gt;Three diffs from the CEO of a 70,000-employee company is a footnote in a 100-million-diff monorepo. The signal isn't the code. It's the behavior.&lt;/p&gt;

&lt;p&gt;Zuckerberg isn't the only founder pulled back into the codebase by AI tools. Garry Tan, CEO of Y Combinator, &lt;a href="https://github.com/garrytan/gstack" rel="noopener noreferrer"&gt;returned to coding&lt;/a&gt; after 15 years and open-sourced gstack, a Claude Code system with 23 specialist tools that turns the CLI into a virtual engineering team: code reviewer, QA lead, security auditor, release engineer. &lt;sup id="fnref3"&gt;3&lt;/sup&gt; &lt;sup id="fnref4"&gt;4&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Tobias Lütke, CEO of Shopify, has been running experiments with &lt;a href="https://dev.to/blogs/karpathy-autoresearch-explained-ml-to-marketing"&gt;Karpathy's AutoResearch&lt;/a&gt; on internal company data. 37 experiments overnight. 19% performance gain.&lt;/p&gt;

&lt;p&gt;I wrote about &lt;a href="https://dev.to/blogs/karpathy-autoresearch-explained-ml-to-marketing"&gt;how AutoResearch works&lt;/a&gt; a few days ago. The throughline is the same: AI tools are collapsing the gap between "person with ideas" and "person who ships code." Founders used to be the first type. AI is turning them back into the second.&lt;/p&gt;

&lt;h2&gt;
  
  
  Meta's bet: AI writes most of the code
&lt;/h2&gt;

&lt;p&gt;Zuckerberg coding again isn't a hobby. It's a signal of where Meta is heading.&lt;/p&gt;

&lt;p&gt;Leaked internal documents from March 2026 show aggressive targets. Meta's creation org wants &lt;strong&gt;65% of engineers writing 75% or more of their committed code using AI&lt;/strong&gt; by mid-2026. The Scalable Machine Learning org set a target of 50-80% AI-assisted code. &lt;sup id="fnref5"&gt;5&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Zuckerberg himself said on Dwarkesh Patel's podcast that "in the next year, maybe half the development will be done by AI as opposed to people, and that will kind of increase from there." &lt;sup id="fnref6"&gt;6&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;He's not predicting this from the sidelines. He's using Claude Code in the terminal to ship diffs to his own monorepo. The CEO is the pilot customer.&lt;/p&gt;

&lt;h2&gt;
  
  
  The pattern worth watching
&lt;/h2&gt;

&lt;p&gt;There's a recurring shape here.&lt;/p&gt;

&lt;p&gt;Karpathy builds AutoResearch. Constrains the agent to one file, one metric, one 5-minute cycle. The constraint is the invention. Lütke runs it on Shopify data overnight. Marketers adapt it for landing pages.&lt;/p&gt;

&lt;p&gt;Anthropic builds Claude Code. Tan wraps it in 23 specialist agents. Zuckerberg uses it to ship his first code in 20 years.&lt;/p&gt;

&lt;p&gt;The tools don't just help engineers code faster. They re-open coding to people who stopped. Founders who moved into strategy, management, fundraising. People who haven't touched a codebase in a decade. The barrier to re-entry used to be months of catching up on tooling, frameworks, and conventions. Now it's a terminal and a prompt.&lt;/p&gt;

&lt;p&gt;That's a different kind of disruption than "AI replaces developers." It's closer to: AI brings back the builder-CEO. The person who can see a problem, describe a solution, and ship it before the meeting ends.&lt;/p&gt;

&lt;p&gt;Whether Zuckerberg's 3 diffs were good code is beside the point. The 200 engineers who approved them probably weren't reviewing for correctness. But the fact that a CEO can sit down with Claude Code and produce something that compiles, passes CI, and lands in a 100-million-diff monorepo? That's the new baseline.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key takeaways
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zuckerberg shipped 3 diffs&lt;/strong&gt; to Meta's monorepo in March 2026, his first code in ~20 years, using Claude Code CLI&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;One diff got 200+ approvals&lt;/strong&gt; from engineers eager to review the CEO's code&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Garry Tan&lt;/strong&gt; (Y Combinator) also returned to coding after 15 years, open-sourcing gstack for Claude Code&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meta targets 65-75% AI-assisted code&lt;/strong&gt; across engineering by mid-2026&lt;/li&gt;
&lt;li&gt;AI coding tools are pulling &lt;strong&gt;founders back into codebases&lt;/strong&gt; they left years ago&lt;/li&gt;
&lt;li&gt;The disruption isn't "AI replaces developers," it's &lt;strong&gt;"AI re-opens development"&lt;/strong&gt; to people who stopped&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;I break down things like this on &lt;a href="https://linkedin.com/in/monkfromearth" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://x.com/monkfromearth" rel="noopener noreferrer"&gt;X&lt;/a&gt;, and &lt;a href="https://instagram.com/monkfrom.earth" rel="noopener noreferrer"&gt;Instagram&lt;/a&gt;. If this resonated, you'd probably like those too.&lt;/p&gt;







&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://newsletter.pragmaticengineer.com/p/the-pulse-industry-leaders-return" rel="noopener noreferrer"&gt;The Pulse: Industry leaders return to coding with AI — The Pragmatic Engineer&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://newsletter.pragmaticengineer.com/p/the-coding-machine-at-meta" rel="noopener noreferrer"&gt;"The Coding Machine" at Meta with Michael Novati — The Pragmatic Engineer&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;&lt;a href="https://github.com/garrytan/gstack" rel="noopener noreferrer"&gt;gstack — Garry Tan's Claude Code setup (GitHub)&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://techcrunch.com/2026/03/17/why-garry-tans-claude-code-setup-has-gotten-so-much-love-and-hate/" rel="noopener noreferrer"&gt;Why Garry Tan's Claude Code setup has gotten so much love, and hate — TechCrunch&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn5"&gt;
&lt;p&gt;&lt;a href="https://www.theweek.in/news/sci-tech/2026/03/27/how-aggressive-is-mark-zuckerberg-s-ai-native-push-for-meta-leaked-documents-offer-new-details-on-coding-targets.html" rel="noopener noreferrer"&gt;How aggressive is Mark Zuckerberg's 'AI-native' push for Meta? — The Week&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn6"&gt;
&lt;p&gt;&lt;a href="https://www.dwarkesh.com/p/mark-zuckerberg-2" rel="noopener noreferrer"&gt;Mark Zuckerberg — AI will write most Meta code in 18 months — Dwarkesh Patel&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>productivity</category>
      <category>career</category>
    </item>
  </channel>
</rss>
