<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Benedict L</title>
    <description>The latest articles on Forem by Benedict L (@ailnk0).</description>
    <link>https://forem.com/ailnk0</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3555390%2Fd7480b51-a8eb-491a-9873-7a388a9f6cd5.jpg</url>
      <title>Forem: Benedict L</title>
      <link>https://forem.com/ailnk0</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/ailnk0"/>
    <language>en</language>
    <item>
      <title>Claude Code stuck in a failure loop? I escaped with multi-agent.</title>
      <dc:creator>Benedict L</dc:creator>
      <pubDate>Tue, 06 Jan 2026 12:46:35 +0000</pubDate>
      <link>https://forem.com/ailnk0/claude-code-stuck-in-a-failure-loop-i-escaped-with-multi-agent-16di</link>
      <guid>https://forem.com/ailnk0/claude-code-stuck-in-a-failure-loop-i-escaped-with-multi-agent-16di</guid>
      <description>&lt;p&gt;I tried to improve benchmark scores on my open-source project with Claude Code. Failed miserably.&lt;/p&gt;

&lt;p&gt;It would try approach A, hit a wall, switch to B, then circle back to A. A classic failure loop.&lt;/p&gt;

&lt;p&gt;Too much data and work to fit in one agent's context.&lt;/p&gt;

&lt;p&gt;And every time Auto Compact ran, earlier decisions and work got diluted.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2xazjxtjjtyj93n8wzu7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2xazjxtjjtyj93n8wzu7.png" alt=" " width="800" height="435"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;So I switched to a multi-agent architecture.&lt;/p&gt;

&lt;p&gt;Redesigned the entire task into 12 Tasks across 5 Phases.&lt;br&gt;
Each Task is sized so one agent can complete it within its context window.&lt;/p&gt;

&lt;p&gt;Sub-agents run in parallel or sequence depending on dependencies.&lt;/p&gt;

&lt;p&gt;The key is knowledge transfer between agents.&lt;/p&gt;

&lt;p&gt;For example, the triage optimization routine from Task 10 → Task 11 ran 5 iterations:&lt;br&gt;
Round 1: Precision 27%, 110 false positives.&lt;br&gt;
Round 2: Removed Y-overlap. Precision up to 36%, FP down to 72.&lt;br&gt;
...&lt;br&gt;
Round 5: Added image ratio conditions. Hit 100% Recall.&lt;br&gt;
Each experiment result gets recorded for the next agent to read.&lt;br&gt;
New agents read previous results and don't retry failed approaches.&lt;br&gt;
When an agent runs a new experiment, I make the call on what to focus on next.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fue206svl3iffvig34ot2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fue206svl3iffvig34ot2.png" alt=" " width="800" height="768"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Results:&lt;/p&gt;

&lt;p&gt;✅ The A-B-A failure loop disappeared.&lt;br&gt;
✅ Table inference score improved from 0.49 to 0.93—nearly 2x.&lt;br&gt;
✅ Beyond expectations: achieved #1 among open-source solutions.&lt;/p&gt;

&lt;p&gt;Why did multi-agent work?&lt;/p&gt;

&lt;p&gt;The bottleneck for today's AI agents isn't intelligence—it's memory.&lt;/p&gt;

&lt;p&gt;Context windows are small. The longer they get, the more early decisions get buried.&lt;/p&gt;

&lt;p&gt;You have to engineer both the size and structure of memory.&lt;br&gt;
If Claude Code is stuck in a failure loop, run through this checklist.&lt;/p&gt;

&lt;p&gt;📋 Checklist:&lt;/p&gt;

&lt;p&gt;Size (task decomposition)&lt;br&gt;
☐ Is each task focused on a single goal?&lt;br&gt;
☐ Can one agent complete it within context limits?&lt;br&gt;
☐ Are task dependencies clearly defined?&lt;/p&gt;

&lt;p&gt;Structure (knowledge transfer)&lt;br&gt;
☐ Are previous agent results documented?&lt;br&gt;
☐ Are required specs/docs loaded into context?&lt;br&gt;
☐ Can the agent self-verify completion criteria?&lt;/p&gt;

</description>
    </item>
    <item>
      <title>AI Agents aren't genies that grant wishes</title>
      <dc:creator>Benedict L</dc:creator>
      <pubDate>Tue, 06 Jan 2026 12:43:58 +0000</pubDate>
      <link>https://forem.com/ailnk0/ai-agents-arent-genies-that-grant-wishes-2l20</link>
      <guid>https://forem.com/ailnk0/ai-agents-arent-genies-that-grant-wishes-2l20</guid>
      <description>&lt;p&gt;"Plan and build a dashboard for me." Sounds like it should work. It doesn't.&lt;/p&gt;

&lt;p&gt;Planners and developers have different goals.&lt;/p&gt;

&lt;p&gt;A planner's goal is delivering value. "What will the customer be able to decide from this?"&lt;/p&gt;

&lt;p&gt;A developer's goal is shipping code. "Does it run?"&lt;/p&gt;

&lt;p&gt;Give one agent both jobs, and it prioritizes whichever is easier to measure.&lt;/p&gt;

&lt;p&gt;"Is the planning good?" Hard to judge.&lt;/p&gt;

&lt;p&gt;"Does the code run?" Immediate feedback.&lt;/p&gt;

&lt;p&gt;So the agent rushes through planning and jumps straight to implementation.&lt;/p&gt;

&lt;p&gt;"Plan and build this" becomes "half-ass the plan and build this."&lt;/p&gt;

&lt;p&gt;That's why role separation matters.&lt;/p&gt;

&lt;p&gt;1️⃣ Researcher role → Investigate customer interests, gather industry trends and benchmarks&lt;/p&gt;

&lt;p&gt;2️⃣ Planner role → Redefine customer value based on research, select key KPIs, design screen structure&lt;/p&gt;

&lt;p&gt;3️⃣ Developer role → Implement based on the planning doc&lt;br&gt;
We review each agent's output, iterate with feedback and direction until it hits the right level, then move to the next stage.&lt;/p&gt;

&lt;p&gt;Now each agent can focus on its specialty.&lt;/p&gt;

&lt;p&gt;Once the planning agent finishes the "why," the dev agent only worries about the "how."&lt;/p&gt;

&lt;p&gt;Separate the roles. Define each role's expertise and success criteria.&lt;/p&gt;

&lt;p&gt;That's context engineering.&lt;/p&gt;

&lt;p&gt;What's interesting is how much context engineering resembles leadership.&lt;/p&gt;

&lt;p&gt;A leader isn't a superhero who's best at everything.&lt;/p&gt;

&lt;p&gt;A leader sets a valuable goal and assembles a team of specialists who are better than them.&lt;/p&gt;

&lt;p&gt;AI Agents aren't genies—but with a good leader, they become an incredible team.&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>productivity</category>
      <category>systemdesign</category>
    </item>
    <item>
      <title>I thought plugging in LangSmith would solve agentic AI monitoring</title>
      <dc:creator>Benedict L</dc:creator>
      <pubDate>Tue, 06 Jan 2026 12:42:34 +0000</pubDate>
      <link>https://forem.com/ailnk0/i-thought-plugging-in-langsmith-would-solve-agentic-ai-monitoring-5fng</link>
      <guid>https://forem.com/ailnk0/i-thought-plugging-in-langsmith-would-solve-agentic-ai-monitoring-5fng</guid>
      <description>&lt;p&gt;I thought plugging in LangSmith would solve agentic AI monitoring&lt;br&gt;
In my last post, I shared the "costs are a black box" problem.&lt;/p&gt;

&lt;p&gt;Some issues used 10,000 tokens, others 1,000,000. A 100x difference with no explanation.&lt;/p&gt;

&lt;p&gt;I figured switching to LangGraph and adding LangSmith would fix it.&lt;/p&gt;

&lt;p&gt;Wrong.&lt;/p&gt;

&lt;p&gt;First wall: CLI calls can't be traced&lt;/p&gt;

&lt;p&gt;The original setup ran Claude Code CLI as a subprocess from GitHub Actions.&lt;/p&gt;

&lt;p&gt;GitHub Actions → npx claude-code → (black box) → result&lt;/p&gt;

&lt;p&gt;Even with LangSmith connected, LLM calls inside the CLI were invisible. The problem wasn't the tooling—it was that the architecture wasn't observable in the first place.&lt;/p&gt;

&lt;p&gt;So I switched from CLI to the Anthropic SDK.&lt;/p&gt;

&lt;p&gt;——&lt;br&gt;
Second wall: tokens don't show up in LangSmith&lt;/p&gt;

&lt;p&gt;Switched to SDK. Still no token counts.&lt;/p&gt;

&lt;p&gt;After debugging, I learned:&lt;/p&gt;

&lt;p&gt;You need run_type=llm for tokens to be tracked&lt;br&gt;
input_cost and output_cost must be manually added as metadata&lt;/p&gt;

&lt;p&gt;I assumed an "observability tool" would show everything automatically. Turns out, defining what you want to see is your job.&lt;/p&gt;

&lt;p&gt;Third wall: the trap of over-engineering&lt;/p&gt;

&lt;p&gt;I tried to build for scale and include LangGraph V1's new features.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Durable State &amp;amp; Built-in Persistence&lt;/li&gt;
&lt;li&gt;Scoring, observability, model config&lt;/li&gt;
&lt;li&gt;Fallback API logic&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In the end, I deleted all of it.&lt;/p&gt;

&lt;p&gt;Once I actually ran it, none of that was needed. Most settings were better hardcoded. Unnecessary logic just added complexity.&lt;/p&gt;

&lt;p&gt;Result: now I can see the costs&lt;/p&gt;

&lt;p&gt;The cause of that 100x difference finally became visible:&lt;/p&gt;

&lt;p&gt;Triage → input 1.5K / output 102 tokens → $0.002&lt;br&gt;
Analyze → input 18K / output 703K tokens → $0.105&lt;br&gt;
Fix → input 50K / output 3,760K tokens → $0.512&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdyzd4oriquxrvqzphk8s.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdyzd4oriquxrvqzphk8s.png" alt=" " width="725" height="561"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fue206svl3iffvig34ot2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fue206svl3iffvig34ot2.png" alt=" " width="800" height="768"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The shift from "running on gut feel" to "running on numbers."&lt;/p&gt;

&lt;p&gt;Plugging in an observability tool is easy. Building an observable architecture is the real work.&lt;/p&gt;

&lt;p&gt;Next: a dashboard that shows whether these numbers actually deliver value—productivity metrics in real time.&lt;/p&gt;

</description>
      <category>agents</category>
      <category>architecture</category>
      <category>llm</category>
      <category>monitoring</category>
    </item>
    <item>
      <title>The real problems I faced after deploying agentic AI to production</title>
      <dc:creator>Benedict L</dc:creator>
      <pubDate>Tue, 06 Jan 2026 12:40:27 +0000</pubDate>
      <link>https://forem.com/ailnk0/the-real-problems-i-faced-after-deploying-agentic-ai-to-production-2b0l</link>
      <guid>https://forem.com/ailnk0/the-real-problems-i-faced-after-deploying-agentic-ai-to-production-2b0l</guid>
      <description>&lt;p&gt;In my last post, I shared that agentic AI cut issue handling time by 66%.&lt;br&gt;
The numbers looked great on paper. But the moment I asked "can we actually keep running this?"—unexpected problems started piling up.&lt;/p&gt;

&lt;p&gt;Costs are a black box&lt;/p&gt;

&lt;p&gt;Agents make multiple LLM calls internally.&lt;br&gt;
Reading code, analyzing, deciding, sometimes double-checking.&lt;/p&gt;

&lt;p&gt;But I had no way to track how many tokens any of this used.&lt;br&gt;
Some issues consumed 10,000 tokens. Others burned 1,000,000.&lt;br&gt;
A 100x difference—and I couldn't tell why.&lt;/p&gt;

&lt;p&gt;The current setup runs Claude Code CLI inside GitHub Actions.&lt;/p&gt;

&lt;p&gt;There's no system for systematically tracking and analyzing token usage.&lt;/p&gt;

&lt;p&gt;I discovered that tools like LangSmith are built to solve exactly this. Next step: either parse CLI logs into structured data, or integrate a dedicated monitoring tool.&lt;/p&gt;

&lt;p&gt;The workflow only exists as code&lt;/p&gt;

&lt;p&gt;Right now the workflow is 4 YAML files. Each one is 100-300 lines, and the conditional branches keep getting more complex. When a teammate asks "what exactly is the AI doing right now?"—I have to open the code and explain line by line.&lt;/p&gt;

&lt;p&gt;Trust is everything for agentic AI.&lt;/p&gt;

&lt;p&gt;To explain "why did the AI make this decision?", you need to see the intermediate steps visually.&lt;/p&gt;

&lt;p&gt;I'm looking into low-code platforms like LangFlow and n8n. If the visual workflow actually reflects what's running, with logs at each node, that would help a lot.&lt;/p&gt;

&lt;p&gt;No real-time KPIs&lt;/p&gt;

&lt;p&gt;"AI saved 66% of developer time" came from manually analyzing 10 issues.&lt;/p&gt;

&lt;p&gt;But I have no idea how well the AI is performing right now.&lt;br&gt;
Metrics I need:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;AI filtering rate: % of issues AI resolved without developer involvement&lt;/li&gt;
&lt;li&gt;Average first response time: how fast did we reply after issue creation&lt;/li&gt;
&lt;li&gt;Auto-resolution rate: % where AI fixed code and opened a PR&lt;/li&gt;
&lt;li&gt;Daily/weekly/monthly trends: is performance improving or degrading&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Without a dashboard, I'm left judging by feel—"seems to be working fine."&lt;/p&gt;

&lt;p&gt;Agentic AI isn't "build and done"—it's "the beginning of operations."&lt;/p&gt;

&lt;p&gt;Results came fast, but making this a sustainable system is a completely different challenge.&lt;/p&gt;

&lt;p&gt;Next time I'll share how I'm tackling these problems one by one. Starting with cost monitoring and real-time dashboards.&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>llm</category>
      <category>monitoring</category>
    </item>
    <item>
      <title>What developers do is changing after agentic AI</title>
      <dc:creator>Benedict L</dc:creator>
      <pubDate>Tue, 06 Jan 2026 10:17:22 +0000</pubDate>
      <link>https://forem.com/ailnk0/what-developers-do-is-changing-after-agentic-ai-4lh5</link>
      <guid>https://forem.com/ailnk0/what-developers-do-is-changing-after-agentic-ai-4lh5</guid>
      <description>&lt;p&gt;Before, when developers handled issues directly, even a simple question meant re-reading context, crafting a response, double-checking nothing was missed. Trying to remember if a similar issue came up before, digging through old threads to see how it was handled. This triage and analysis work quietly, steadily drains time and focus—regardless of how hard the actual issue is.&lt;/p&gt;

&lt;p&gt;I thought: maybe AI should be doing this, not developers.&lt;/p&gt;

&lt;p&gt;So I applied an agentic AI workflow to real work—analyzing GitHub issues, responding, and in some cases, even fixing code.&lt;/p&gt;

&lt;p&gt;I compared the old human-driven workflow against the agentic AI workflow across 10 GitHub issues.&lt;/p&gt;

&lt;p&gt;The criteria were simple.&lt;/p&gt;

&lt;p&gt;From the organization's perspective:&lt;br&gt;
→ How much time does it actually take to resolve one issue?&lt;/p&gt;

&lt;p&gt;From the user's perspective:&lt;br&gt;
→ How fast do they get a response after filing an issue?&lt;/p&gt;

&lt;p&gt;The results were clear.&lt;/p&gt;

&lt;p&gt;The old workflow took an average of 2.4 developer-days per issue.&lt;/p&gt;

&lt;p&gt;After agentic AI handled triage and first-pass resolution, that dropped to 0.8 days.&lt;/p&gt;

&lt;p&gt;That's a 66.7% cost reduction for the organization.&lt;/p&gt;

&lt;p&gt;The change from the user's side was even more dramatic.&lt;/p&gt;

&lt;p&gt;Previously, first response took over 3 days on average.&lt;/p&gt;

&lt;p&gt;With agentic AI, first response came in 4-5 seconds.&lt;/p&gt;

&lt;p&gt;Users now get near-instant feedback after filing an issue.&lt;/p&gt;

&lt;p&gt;But what impressed me most: even with just simple policies and a basic domain knowledge structure, the agentic AI workflow stayed consistent in knowing what to do and what not to do.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Simple questions got answered and closed&lt;/li&gt;
&lt;li&gt;Small fixable bugs got PRs created directly&lt;/li&gt;
&lt;li&gt;Risky or complex issues got handed to humans without hesitation&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Bottom line: after applying agentic AI, only 4 out of 10 issues actually reached developers.&lt;/p&gt;

&lt;p&gt;About 60% were handled without any developer involvement.&lt;/p&gt;

&lt;p&gt;Developers no longer need to spend hours on repetitive triage, responses, and context reconstruction.&lt;/p&gt;

&lt;p&gt;Instead, they define the boundaries of automation and design how AI and humans should collaborate.&lt;/p&gt;

&lt;p&gt;I believe that's becoming the new role of developers.&lt;/p&gt;

&lt;p&gt;Next up: building a dashboard to see "how is agentic AI performing right now" at a glance.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fewyu48b9xdk0k3y9a3ay.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fewyu48b9xdk0k3y9a3ay.png" alt=" " width="399" height="706"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>automation</category>
      <category>productivity</category>
    </item>
    <item>
      <title>Why being smart isn't enough for agentic AI</title>
      <dc:creator>Benedict L</dc:creator>
      <pubDate>Tue, 06 Jan 2026 10:15:06 +0000</pubDate>
      <link>https://forem.com/ailnk0/why-being-smart-isnt-enough-for-agentic-ai-54mb</link>
      <guid>https://forem.com/ailnk0/why-being-smart-isnt-enough-for-agentic-ai-54mb</guid>
      <description>&lt;p&gt;Agentic AI isn't just about generating answers.&lt;/p&gt;

&lt;p&gt;It judges, acts, and actually affects your systems.&lt;/p&gt;

&lt;p&gt;The moment that happens, "sounding smart" stops mattering.&lt;/p&gt;

&lt;p&gt;What matters is making the same decision every time, given the same situation.&lt;/p&gt;

&lt;p&gt;Without an explicit domain knowledge model, agents make different calls depending on the prompt and context.&lt;/p&gt;

&lt;p&gt;The same issue might:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Get auto-fixed one day&lt;/li&gt;
&lt;li&gt;Be routed to manual handling the next&lt;/li&gt;
&lt;li&gt;End up as just a question another time&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That might look like automation, but you can't call it production-ready agentic AI.&lt;/p&gt;

&lt;p&gt;So instead of burying decision criteria inside prompts, I separated them into an explicit domain knowledge model.&lt;/p&gt;

&lt;p&gt;Now the agent can explain why it made a decision—and make the same call in the same situation, consistently.&lt;/p&gt;

&lt;p&gt;Palantir Ontology and Claude Skills both tackle this problem. Palantir Ontology impressed me by defining what AI should judge against at the system level. Claude Skills solves the same problem in a simpler form.&lt;/p&gt;

&lt;p&gt;For this project, I didn't need enterprise-scale ontology design. I needed to quickly define and validate the decision criteria that mattered right now.&lt;/p&gt;

&lt;p&gt;So I chose Claude Skills. Policies, actions, judgment criteria—all separated into explicit skill documents.&lt;br&gt;
But defining criteria isn't the end.&lt;/p&gt;

&lt;p&gt;An AI that takes action needs guardrails.&lt;/p&gt;

&lt;p&gt;My agentic AI adds labels, closes issues, creates PRs, changes code.&lt;/p&gt;

&lt;p&gt;It has real permissions in a real system.&lt;/p&gt;

&lt;p&gt;So alongside the domain knowledge model, I built a workflow testing framework.&lt;/p&gt;

&lt;p&gt;For each action, I defined test cases: "Given this input, this judgment must follow." Then I run actual Claude API responses to verify.&lt;/p&gt;

&lt;p&gt;Now the agentic workflow isn't "seems to be working"—it's continuously validated against intended criteria.&lt;/p&gt;

&lt;p&gt;Next step: a dashboard that shows the value of agentic AI in numbers, not features.&lt;/p&gt;

&lt;p&gt;How many issues did AI handle? How much time did it save?&lt;/p&gt;

&lt;p&gt;Without that visibility, you can't convince the team or scale further.&lt;/p&gt;

&lt;p&gt;Moving agentic AI from "fun experiment" to "system the team can trust."&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhsvgmvgj24ykghxc09ih.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhsvgmvgj24ykghxc09ih.png" alt=" " width="627" height="321"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>architecture</category>
    </item>
    <item>
      <title>Can I build an agentic AI that handles issues and fixes code for me?</title>
      <dc:creator>Benedict L</dc:creator>
      <pubDate>Tue, 06 Jan 2026 10:13:20 +0000</pubDate>
      <link>https://forem.com/ailnk0/can-i-build-an-agentic-ai-that-handles-issues-and-fixes-code-for-me-5e06</link>
      <guid>https://forem.com/ailnk0/can-i-build-an-agentic-ai-that-handles-issues-and-fixes-code-for-me-5e06</guid>
      <description>&lt;p&gt;GitHub issues were piling up on my open-source project, and it was becoming a real burden.&lt;br&gt;
I needed automation. My first attempt was n8n + LLM.&lt;br&gt;
When an issue came in, the LLM would decide:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Spam or not?&lt;/li&gt;
&lt;li&gt;Bug / feature request / question?&lt;/li&gt;
&lt;li&gt;Need more info?&lt;/li&gt;
&lt;li&gt;Can AI handle this directly?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Triage and auto-commenting worked surprisingly well.&lt;br&gt;
Then I hit a wall.&lt;/p&gt;

&lt;p&gt;The real core of this agentic workflow wasn't classification—it was the AI actually understanding code and taking action.&lt;/p&gt;

&lt;p&gt;Reading code, leaving informed comments, making fixes and opening PRs... For any of that, the agent needs to run where the code lives.&lt;/p&gt;

&lt;p&gt;That's when I landed on GitHub Actions.&lt;/p&gt;

&lt;p&gt;Free compute for AI agents, and the GitHub integration is seamless.&lt;/p&gt;

&lt;p&gt;I designed the architecture with Claude Code and documented all decision policies and role definitions in markdown.&lt;br&gt;
Prototype done in 4 hours.&lt;/p&gt;

&lt;p&gt;Next week: apply improvements, build test cases, deploy to the real project.&lt;/p&gt;

&lt;p&gt;Curious to see how much time this actually saves the team.&lt;br&gt;
What I'm experimenting with next:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Structuring domain knowledge with Claude Skills (policies, metrics, builds, tests)&lt;/li&gt;
&lt;li&gt;Better AI Q&amp;amp;A interactions&lt;/li&gt;
&lt;li&gt;Test cases that validate LLM and agent behavior&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg0q3zargiczaffoafam1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg0q3zargiczaffoafam1.png" alt=" " width="480" height="895"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>automation</category>
      <category>github</category>
    </item>
  </channel>
</rss>
