<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Daniel King</title>
    <description>The latest articles on Forem by Daniel King (@danfking).</description>
    <link>https://forem.com/danfking</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3892219%2Fbc627dae-2978-4190-b262-46d328d917e7.jpg</url>
      <title>Forem: Daniel King</title>
      <link>https://forem.com/danfking</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/danfking"/>
    <language>en</language>
    <item>
      <title>I trained a sprite model with agents. The data was the bottleneck.</title>
      <dc:creator>Daniel King</dc:creator>
      <pubDate>Wed, 06 May 2026 09:47:04 +0000</pubDate>
      <link>https://forem.com/danfking/i-trained-a-sprite-model-with-agents-the-data-was-the-bottleneck-16k6</link>
      <guid>https://forem.com/danfking/i-trained-a-sprite-model-with-agents-the-data-was-the-bottleneck-16k6</guid>
      <description>&lt;p&gt;I just published &lt;a href="https://github.com/danfking/pixel-llm" rel="noopener noreferrer"&gt;pixel-llm&lt;/a&gt;, a small autoregressive transformer that generates 32x32 pixel art sprites of reef sea creatures. About 2.9 million parameters, a 64-colour palette, runs on consumer hardware. Built end to end through agent sessions, with me steering rather than typing.&lt;/p&gt;

&lt;p&gt;The output is sub-par. I am sharing it anyway, because the way it failed taught me something I did not expect.&lt;/p&gt;

&lt;p&gt;The setup was narrow on purpose. I picked sea creatures because the visual vocabulary is constrained: a few zones (shallows, twilight, midnight, abyss, hadal) and a few categories (reef fish, grazer, coral, jellyfish, cephalopod, plus an abyssal catch-all). A small, well-defined domain felt like the right shape for a small model. Six categories, five zones, thirty cells in the grid. Tractable on paper.&lt;/p&gt;

&lt;p&gt;The model itself fell out fast. Agents wrote the transformer, the KV-cache inference loop, the sprite breeding via partial completion, and the post-process palette-aware shader. That last piece is the strongest output. The model produces flat colour-indexed sprites and a separate procedural shader applies directional light and ambient occlusion, staying inside the 64-colour palette by walking pre-computed luminance ramps.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Froycgkk8xa0s3rxydrvh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Froycgkk8xa0s3rxydrvh.png" alt="Flat versus shaded sprites across zones and categories" width="800" height="463"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;When the categories worked, you can see what I was after. When they did not, you can see that too: two of the six categories (cephalopod and one abyssal column) never converged. Pure noise, regardless of sampling temperature.&lt;/p&gt;

&lt;p&gt;I iterated the training data four times. A procedural synthetic generator. Wikimedia Commons photographs, downloaded and palette-quantised. Sprite sheet extraction from OpenGameArt. A mixed corpus stitched together from all three. The validation loss kept going down. The samples for those two categories kept looking wrong. The other four held up well enough to look at.&lt;/p&gt;

&lt;p&gt;That is the part I want to flag. Loss is not taste. The agentic loop has a fast, local correctness signal for the code: does it run, does the loss go down, does it not crash. It does not have a corresponding signal for the data. Whether a corpus is the right shape for a problem is a slow, aesthetic judgment that arrives after a training run, after staring at sample grids, after a cycle measured in hours rather than seconds. Agents cannot close that loop on their own yet.&lt;/p&gt;

&lt;p&gt;So the work split cleanly. The model code, training scaffold, sampler, breeder, and shader were straightforward agent output. The data choices were the part where I had to keep showing up.&lt;/p&gt;

&lt;p&gt;This connects back to something I wrote about in April: when agents take over execution, the premium activity is the layer above. For a coding agent that layer is verification. For a research-flavoured agent loop, it is data curation: deciding what the model should see, recognising when the existing corpus is wrong, and recognising when the iteration has hit its ceiling.&lt;/p&gt;

&lt;p&gt;Knowing when to stop is itself the call. After the fourth dataset I judged that the agentic loop had run out of useful moves for this architecture. The next step would not be more data, it would be a different model shape. I called time, wrote the README honestly, and shipped.&lt;/p&gt;

&lt;p&gt;The repo is up at &lt;a href="https://github.com/danfking/pixel-llm" rel="noopener noreferrer"&gt;github.com/danfking/pixel-llm&lt;/a&gt; with the sample images and a fuller writeup. The interesting thing in there is not the trained model. It is the trail.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>agents</category>
      <category>pixelart</category>
    </item>
    <item>
      <title>Verification is the expensive thing now</title>
      <dc:creator>Daniel King</dc:creator>
      <pubDate>Thu, 23 Apr 2026 09:59:06 +0000</pubDate>
      <link>https://forem.com/danfking/verification-is-the-expensive-thing-now-2lag</link>
      <guid>https://forem.com/danfking/verification-is-the-expensive-thing-now-2lag</guid>
      <description>&lt;p&gt;Martin Fowler's &lt;a href="https://martinfowler.com/fragments/2026-04-02.html" rel="noopener noreferrer"&gt;latest fragments post&lt;/a&gt; collects several ideas about how AI is reshaping software development. The one that stuck with me is Ajey Gore's argument: as coding agents take over execution, verification becomes the premium activity.&lt;/p&gt;

&lt;p&gt;Gore puts it bluntly. Instead of ten engineers building, you might have three engineers plus seven people defining acceptance criteria and designing tests. The bottleneck moves from "can we write the code?" to "do we know whether the code is right?"&lt;/p&gt;

&lt;p&gt;This matches what I see daily. I run multiple Claude Code sessions in parallel, each producing working code at a pace I couldn't match alone. The hard part is never the generation. The hard part is knowing whether what came out actually does what I intended, handles the edges I care about, and doesn't quietly break something else. And it's not just me who needs to know. My team members need to look at that same output and reach the same confidence, often without the context I had when I prompted it.&lt;/p&gt;

&lt;p&gt;The cultural shift Gore describes is the part most teams will struggle with. Your Monday standup changes. Instead of "what did we ship?" the question becomes "what did we validate?" Instead of tracking output, you're tracking whether the output was right. That reframes what it means to be productive. An engineer who catches a subtle misalignment in generated code before it ships has done more valuable work than one who prompted three features into existence without checking them.&lt;/p&gt;

&lt;p&gt;This connects to something else Fowler highlights in the same post: Margaret-Anne Storey's concept of "intent debt," where the goals guiding a system are poorly documented or maintained. If you can't clearly articulate what the system should do, you can't verify that it does it. Intent debt was always a problem, but it was partially hidden when the same person writing the code also held the intent in their head. When an agent writes the code, that implicit knowledge gap becomes a concrete failure mode.&lt;/p&gt;

&lt;p&gt;I think the teams that figure out verification workflows early will have a real advantage. Not just automated tests (though those matter), but the whole practice of clearly stating intent, reviewing output critically, and building confidence that what shipped is what was meant.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agentdev</category>
      <category>productivity</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>🔔 Small productivity hack that's changed how I work with Claude Code</title>
      <dc:creator>Daniel King</dc:creator>
      <pubDate>Wed, 22 Apr 2026 10:05:16 +0000</pubDate>
      <link>https://forem.com/danfking/small-productivity-hack-thats-changed-how-i-work-with-claude-code-5014</link>
      <guid>https://forem.com/danfking/small-productivity-hack-thats-changed-how-i-work-with-claude-code-5014</guid>
      <description>&lt;p&gt;I typically have half a dozen Claude Code sessions running at once, spread across different terminals and monitors, some hidden behind other windows. The visual "done" indicator is easy to miss when you're not looking at the right terminal.&lt;/p&gt;

&lt;p&gt;About a month ago I added a global hook that plays a short chime whenever any session finishes a task. Two minutes to configure. Can't live without it now.&lt;/p&gt;

&lt;p&gt;The difference is about flow. Before, I'd either stare at a terminal waiting, or context-switch and then keep interrupting myself to check which session was ready. Now the chime pulls me back at exactly the right moment. I stay in whatever I'm doing until I hear it, then go find the session that needs me. It's kept me in a flow state in a way I genuinely didn't expect from something so simple.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to set it up
&lt;/h2&gt;

&lt;p&gt;Drop this into your &lt;code&gt;~/.claude/settings.json&lt;/code&gt;:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight json"&gt;&lt;code&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"hooks"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
    &lt;/span&gt;&lt;span class="nl"&gt;"Notification"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="w"&gt;
      &lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
        &lt;/span&gt;&lt;span class="nl"&gt;"matcher"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;""&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
        &lt;/span&gt;&lt;span class="nl"&gt;"hooks"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="w"&gt;
          &lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
            &lt;/span&gt;&lt;span class="nl"&gt;"type"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"command"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
            &lt;/span&gt;&lt;span class="nl"&gt;"command"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"powershell.exe -NoProfile -Command &lt;/span&gt;&lt;span class="se"&gt;\"&lt;/span&gt;&lt;span class="s2"&gt;(New-Object Media.SoundPlayer 'C:&lt;/span&gt;&lt;span class="se"&gt;\\&lt;/span&gt;&lt;span class="s2"&gt;Windows&lt;/span&gt;&lt;span class="se"&gt;\\&lt;/span&gt;&lt;span class="s2"&gt;Media&lt;/span&gt;&lt;span class="se"&gt;\\&lt;/span&gt;&lt;span class="s2"&gt;chimes.wav').PlaySync()&lt;/span&gt;&lt;span class="se"&gt;\"&lt;/span&gt;&lt;span class="s2"&gt;"&lt;/span&gt;&lt;span class="w"&gt;
          &lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
        &lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;&lt;span class="w"&gt;
      &lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
    &lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;That's it. Windows has the sound file built in. For Mac/Linux, swap the command for &lt;code&gt;afplay&lt;/code&gt; or &lt;code&gt;paplay&lt;/code&gt; with a sound file of your choice.&lt;/p&gt;

&lt;p&gt;The empty matcher means it fires on every notification, regardless of which project or session triggered it. Claude Code sends a notification whenever a session finishes and is waiting for input, which is exactly the moment you want to know about.&lt;/p&gt;

</description>
      <category>claudecode</category>
      <category>productivity</category>
      <category>ai</category>
      <category>developer</category>
    </item>
  </channel>
</rss>
