<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: nostalgicskinco</title>
    <description>The latest articles on Forem by nostalgicskinco (@nostalgicskinco).</description>
    <link>https://forem.com/nostalgicskinco</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3778573%2Fd150ac2f-e4d0-4d51-bd0c-87274ab3c459.png</url>
      <title>Forem: nostalgicskinco</title>
      <link>https://forem.com/nostalgicskinco</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/nostalgicskinco"/>
    <language>en</language>
    <item>
      <title>AI Agents Are Making Decisions Nobody Can Audit</title>
      <dc:creator>nostalgicskinco</dc:creator>
      <pubDate>Sat, 21 Feb 2026 06:08:04 +0000</pubDate>
      <link>https://forem.com/nostalgicskinco/ai-agents-are-making-decisions-nobody-can-audit-2gpn</link>
      <guid>https://forem.com/nostalgicskinco/ai-agents-are-making-decisions-nobody-can-audit-2gpn</guid>
      <description>&lt;p&gt;Last month, a developer posted on Reddit about an AI agent that got stuck in a loop and fired off 50,000 API requests before anyone noticed. Production was down. The bill was ugly. And the worst part? Nobody could tell exactly what the agent had been doing or why.&lt;/p&gt;

&lt;p&gt;This isn't an edge case anymore. It's Tuesday.&lt;/p&gt;

&lt;h2&gt;
  
  
  The problem nobody wants to talk about
&lt;/h2&gt;

&lt;p&gt;AI agents are everywhere now. They're calling APIs, querying databases, executing code, and in some cases, spending real money — all autonomously. The frameworks for building them are incredible. CrewAI, LangChain, AutoGen, OpenAI's Agents SDK — they make it shockingly easy to stand up an agent that can do real work.&lt;/p&gt;

&lt;p&gt;But here's what none of these frameworks give you: visibility into what your agent actually did.&lt;/p&gt;

&lt;p&gt;No audit trail. No kill switch. No way to replay what happened after something goes wrong. No policy enforcement before a dangerous action executes. And perhaps most concerning — no PII redaction. Every prompt and completion your agent generates ships directly to your observability backend with customer data, API keys, and internal information fully intact.&lt;/p&gt;

&lt;p&gt;Every team I've talked to handles this differently. Most don't handle it at all.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why this is an infrastructure problem, not an application problem
&lt;/h2&gt;

&lt;p&gt;Think about TLS. Nobody implements TLS differently in every microservice. It's a standardized layer that sits below application code and handles encryption for everything above it.&lt;/p&gt;

&lt;p&gt;Agent safety needs to work the same way.&lt;/p&gt;

&lt;p&gt;If every team builds their own logging, their own kill switches, their own policy checks — you get inconsistency, gaps, and the kind of "we'll deal with it later" approach that leads to the 50,000-request incident above.&lt;/p&gt;

&lt;p&gt;The safety layer needs to be:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Framework-agnostic&lt;/strong&gt; — works whether you're using CrewAI, LangChain, AutoGen, or something custom&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Infrastructure-level&lt;/strong&gt; — operates in the network path and telemetry pipeline, not inside agent code&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Standardized&lt;/strong&gt; — uses OpenTelemetry so it plugs into whatever observability stack you already have&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What I built
&lt;/h2&gt;

&lt;p&gt;I've been working on an open-source project called &lt;strong&gt;AIR Blackbox&lt;/strong&gt; — think of it like a flight recorder for AI agents. It sits between your agents and your LLM providers and captures everything.&lt;/p&gt;

&lt;p&gt;The architecture is straightforward:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Your Agent ──→ Gateway ──→ Policy Engine ──→ LLM Provider
                 │               │
                 ▼               ▼
           OTel Collector   Kill Switches
                 │          Trust Scoring
                 ▼          Risk Tiers
           Episode Store
           Jaeger · Prometheus
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;One line change — swap your &lt;code&gt;base_url&lt;/code&gt; — and every agent call flows through it. No SDK changes. No code refactoring.&lt;/p&gt;

&lt;p&gt;Here's what each piece does:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Gateway&lt;/strong&gt; — An OpenAI-compatible reverse proxy written in Go. It intercepts all LLM traffic, emits structured OpenTelemetry traces, and checks policies before forwarding requests. Any OpenAI-compatible client works without modification.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Policy Engine&lt;/strong&gt; — Evaluates requests against YAML-defined policies in real time. Risk tiers (low, medium, high, critical), trust scoring, programmable kill switches, and human-in-the-loop gates for high-risk operations. This isn't monitoring after the fact — it's governance before the action happens.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;OTel Collector&lt;/strong&gt; — A custom processor for gen_ai telemetry. PII redaction using hash-and-preview (48-character preview + hash, so you can debug without exposing full data). Cost metrics. And loop detection — the thing that would have caught that 50,000-request incident before it became a disaster.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Episode Store&lt;/strong&gt; — Groups individual traces into task-level episodes you can replay. When something goes wrong, you don't sift through raw logs — you replay the episode like rewinding a tape.&lt;/p&gt;

&lt;h2&gt;
  
  
  The part I didn't expect
&lt;/h2&gt;

&lt;p&gt;When I started building this, I thought the hard problem would be the technical architecture. It wasn't. OpenTelemetry gives you a solid foundation. Go is great for proxies. The plumbing was actually the straightforward part.&lt;/p&gt;

&lt;p&gt;The hard problem is convincing people they need it before the incident happens.&lt;/p&gt;

&lt;p&gt;Every team I talk to says some version of: "We're being careful." "Our agents are simple." "We'll add monitoring later."&lt;/p&gt;

&lt;p&gt;And then later arrives as a production incident, a leaked API key, or an auditor asking questions nobody prepared for.&lt;/p&gt;

&lt;p&gt;The companies that are thinking about this — the ones deploying agents in regulated industries, in healthcare, in finance — they already know. They're the ones asking: "Can we prove what our agent did? Can we shut it down instantly? Can we guarantee PII doesn't leak into our trace backend?"&lt;/p&gt;

&lt;p&gt;These aren't hypothetical questions. ISO 27001 auditors are starting to ask them. SOC 2 reviewers are starting to ask them. And if your answer is "we log stuff to CloudWatch," that's not going to cut it.&lt;/p&gt;

&lt;h2&gt;
  
  
  What's next
&lt;/h2&gt;

&lt;p&gt;AIR Blackbox is fully open source under Apache 2.0. It's 21 repositories, fully modular — you can use the whole stack or just the pieces you need.&lt;/p&gt;

&lt;p&gt;There are trust plugins for CrewAI, LangChain, AutoGen, and OpenAI's Agents SDK. A five-minute quickstart gets you the full stack running locally with &lt;code&gt;make up&lt;/code&gt;.&lt;/p&gt;

&lt;p&gt;If you're deploying AI agents in production — or planning to — I'd genuinely appreciate your feedback. What gaps are you seeing? What keeps you up at night?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;GitHub:&lt;/strong&gt; &lt;a href="https://github.com/airblackbox" rel="noopener noreferrer"&gt;github.com/airblackbox&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;There are interactive demos in the README if you want to explore without installing anything.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;I'm building AIR Blackbox because I think agent safety shouldn't be an afterthought bolted on after the first incident. It should be infrastructure — boring, reliable, and already running when the 50,001st request tries to fire.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>security</category>
      <category>devops</category>
    </item>
    <item>
      <title>Why Your AI Agents Need a Black Box</title>
      <dc:creator>nostalgicskinco</dc:creator>
      <pubDate>Wed, 18 Feb 2026 03:09:21 +0000</pubDate>
      <link>https://forem.com/nostalgicskinco/why-your-ai-agents-need-a-black-box-i12</link>
      <guid>https://forem.com/nostalgicskinco/why-your-ai-agents-need-a-black-box-i12</guid>
      <description>&lt;p&gt;My AI agents went rogue.&lt;/p&gt;

&lt;p&gt;I run an e-commerce store. A few months ago, I deployed AI agents to handle customer emails — returns, refund requests, product questions. It worked great, until it didn't.&lt;/p&gt;

&lt;p&gt;The agents started making promises we couldn't keep: wrong refund amounts, unauthorized discounts, completely fabricated policies. "Sure, we'll refund your shipping even though our policy says otherwise." "Yes, you can return that item after 90 days." None of it was true.&lt;/p&gt;

&lt;p&gt;The worst part wasn't that they failed. That's fixable. The worst part was that &lt;strong&gt;I couldn't prove what they actually said.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;When customers disputed AI responses, I had nothing. Logs were scattered across three different services. They were mutable — anyone (or any process) could quietly change them after the fact. And they were incomplete. Half the tool calls weren't captured at all.&lt;/p&gt;

&lt;p&gt;I had no audit trail. No accountability. No evidence.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Gap That Nobody Talks About
&lt;/h2&gt;

&lt;p&gt;When I went looking for a solution, I found plenty of observability tools. Langfuse. Helicone. LangSmith. They're all excellent at showing you &lt;em&gt;what happened&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;But I didn't need to see what happened. I needed to &lt;strong&gt;prove&lt;/strong&gt; what happened.&lt;/p&gt;

&lt;p&gt;That distinction sounds subtle. It isn't.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Observability&lt;/strong&gt; answers: "What did the agent do?"&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Accountability&lt;/strong&gt; answers: "What did the agent do, and can you prove it wasn't changed after the fact?"&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In a regulated world — and we're entering one fast — that difference is everything. The EU AI Act is partially in force. The Colorado AI Act hits in June 2026. Texas TRAIGA is live now. The SEC has made AI governance its top priority for 2026.&lt;/p&gt;

&lt;p&gt;Companies deploying AI agents for anything consequential (approving loans, handling complaints, writing medical summaries, processing transactions) are going to need tamper-evident records of what their AI said and did. Not logs. Proof.&lt;/p&gt;

&lt;h2&gt;
  
  
  What I Built
&lt;/h2&gt;

&lt;p&gt;So I built AIR — the open-source black box for AI agents.&lt;/p&gt;

&lt;p&gt;Like the flight recorder on an aircraft, AIR captures every decision, every interaction, every tool call your AI agents make. But unlike scattered logs, AIR creates &lt;strong&gt;cryptographic chains&lt;/strong&gt;: HMAC-SHA256 proof that records haven't been modified after the fact. Change one record and the entire chain breaks.&lt;/p&gt;

&lt;p&gt;Three lines of code to wrap your existing OpenAI app:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="kn"&gt;from&lt;/span&gt; &lt;span class="n"&gt;openai&lt;/span&gt; &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;OpenAI&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;air&lt;/span&gt;

&lt;span class="n"&gt;client&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;air&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;air_wrap&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="nc"&gt;OpenAI&lt;/span&gt;&lt;span class="p"&gt;())&lt;/span&gt;
&lt;span class="n"&gt;response&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;client&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;chat&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;completions&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;create&lt;/span&gt;&lt;span class="p"&gt;(...)&lt;/span&gt;
&lt;span class="c1"&gt;# Every call is now recorded with a tamper-evident audit trail
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;That's it. Every prompt, completion, tool call, and model decision is captured — with cryptographic integrity — stored on your own infrastructure, never leaving your control.&lt;/p&gt;

&lt;h2&gt;
  
  
  What It Actually Solves
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;When a customer disputes what your AI told them:&lt;/strong&gt; You have a signed, timestamped record of the exact conversation. Not a log file that could have been edited — cryptographic proof of what was said.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;When a regulator asks for your AI governance documentation:&lt;/strong&gt; AIR auto-generates compliance reports mapped to SOC 2, ISO 27001, and EU AI Act requirements. 22 controls, pre-mapped.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;When your agent goes off-script and you don't know why:&lt;/strong&gt; Deterministic replay lets you reproduce any AI decision exactly as it happened, in isolation, for debugging.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;When your team changes a model or prompt:&lt;/strong&gt; You have a before/after comparison with the same inputs, so you can prove the change didn't introduce new failure modes.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Ecosystem
&lt;/h2&gt;

&lt;p&gt;AIR isn't one repo — it's a complete accountability stack across 19 open-source repositories:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;&lt;a href="https://github.com/nostalgicskinco/air-blackbox-gateway" rel="noopener noreferrer"&gt;air-blackbox-gateway&lt;/a&gt;&lt;/strong&gt; — OpenAI-compatible reverse proxy that captures every LLM call&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;&lt;a href="https://github.com/nostalgicskinco/air-sdk-python" rel="noopener noreferrer"&gt;air-sdk-python&lt;/a&gt;&lt;/strong&gt; — Python SDK wrapping OpenAI, LangChain, and CrewAI&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;&lt;a href="https://github.com/nostalgicskinco/agent-policy-engine" rel="noopener noreferrer"&gt;agent-policy-engine&lt;/a&gt;&lt;/strong&gt; — Risk-tiered autonomy: policies, kill switches, trust scoring&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;&lt;a href="https://github.com/nostalgicskinco/eval-harness" rel="noopener noreferrer"&gt;eval-harness&lt;/a&gt;&lt;/strong&gt; — Replay episodes, score results, detect regressions&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;&lt;a href="https://github.com/nostalgicskinco/mcp-policy-gateway" rel="noopener noreferrer"&gt;mcp-policy-gateway&lt;/a&gt;&lt;/strong&gt; — Firewall for AI agent tool access via MCP&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The Python SDK is live on PyPI: &lt;code&gt;pip install air-blackbox-sdk&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;There's also an &lt;strong&gt;&lt;a href="https://nostalgicskinco.github.io/air-blackbox-gateway/" rel="noopener noreferrer"&gt;interactive demo&lt;/a&gt;&lt;/strong&gt; you can try right now in your browser — watch an agent run, inspect the audit chain, tamper with a record, and see the chain break.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Now
&lt;/h2&gt;

&lt;p&gt;The timing isn't accidental. 90% of enterprises use AI in daily operations. Only 18% have governance frameworks. The tools simply don't exist yet for most companies.&lt;/p&gt;

&lt;p&gt;But they will need to exist — and soon. August 2026 is when EU AI Act enforcement begins for high-risk systems. June 2026 for Colorado. January 2026 for Texas (already live).&lt;/p&gt;

&lt;p&gt;If you're building AI agents that affect real people, you need to be thinking about this now, not after your first customer dispute or regulatory inquiry.&lt;/p&gt;




&lt;p&gt;AIR is open source and free to use. The hard part is done — the code is real, the SDK is live, the demo works.&lt;/p&gt;

&lt;p&gt;If you're building AI agents in production, I'd love your feedback. Try the interactive demo, kick the tires on the SDK, file an issue if something doesn't work.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;GitHub:&lt;/strong&gt; &lt;a href="https://github.com/nostalgicskinco/air-blackbox-gateway" rel="noopener noreferrer"&gt;github.com/nostalgicskinco/air-blackbox-gateway&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Jason Shotwell builds e-commerce tooling and, apparently, AI infrastructure when his agents go rogue.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>opensource</category>
      <category>devops</category>
    </item>
  </channel>
</rss>
