<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Sam Richard</title>
    <description>The latest articles on Forem by Sam Richard (@sam_richard).</description>
    <link>https://forem.com/sam_richard</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2843548%2Ffc5b3355-d8a8-4d2a-9038-33e72080bbfd.png</url>
      <title>Forem: Sam Richard</title>
      <link>https://forem.com/sam_richard</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/sam_richard"/>
    <language>en</language>
    <item>
      <title>What are AI gateways in 2026, and do you actually need one now?</title>
      <dc:creator>Sam Richard</dc:creator>
      <pubDate>Tue, 14 Apr 2026 19:28:26 +0000</pubDate>
      <link>https://forem.com/sam_richard/what-are-ai-gateways-in-2026-and-do-you-actually-need-one-now-22a5</link>
      <guid>https://forem.com/sam_richard/what-are-ai-gateways-in-2026-and-do-you-actually-need-one-now-22a5</guid>
      <description>&lt;p&gt;Six months ago, we wrote about &lt;a href="https://dev.to/blog/ai-gateways-2025"&gt;AI gateways&lt;/a&gt; and whether you actually needed one. At the time, the pitch was straightforward: a middleware layer to manage API keys, handle failovers, and route prompts to the right model. Useful, but optional for most teams.&lt;/p&gt;

&lt;p&gt;That advice aged fast. The rise of &lt;strong&gt;agentic AI&lt;/strong&gt; (autonomous systems that plan, use tools, write code, and call other models on your behalf) has changed what AI infrastructure needs to handle. A single user request can now trigger dozens of LLM calls, tool invocations, and multi-step reasoning chains. The gateway isn't just routing prompts anymore. It's managing &lt;em&gt;sessions&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;Let's take a fresh look.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is an AI gateway (2026 edition)?
&lt;/h2&gt;

&lt;p&gt;An AI gateway is still a control tower for your AI traffic, a middleware layer between your applications and the AI services they rely on. That part hasn't changed.&lt;/p&gt;

&lt;p&gt;What &lt;em&gt;has&lt;/em&gt; changed is what "AI traffic" looks like. In 2025, it was mostly prompt-in, response-out. In 2026, it's agents calling Claude Opus for complex reasoning, then Haiku for fast classification, then hitting a &lt;a href="https://modelcontextprotocol.io/" rel="noopener noreferrer"&gt;Model Context Protocol (MCP)&lt;/a&gt; server to read from Slack, then writing to a database, then calling another model to verify the result—all from a single user request.&lt;/p&gt;

&lt;p&gt;AI gateways now play a role similar to what ngrok does for production API workloads. ngrok creates a secure, observable interface between your services and the public internet. AI gateways do the same, but for the increasingly complex web of model interactions, tool calls, and agent actions flowing through your stack.&lt;/p&gt;

&lt;p&gt;If &lt;strong&gt;ngrok&lt;/strong&gt; is the gateway to your web traffic, an &lt;strong&gt;AI gateway&lt;/strong&gt; is the gateway to your &lt;strong&gt;agent traffic&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why AI gateways went from "nice to have" to essential
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Agents changed the traffic pattern
&lt;/h3&gt;

&lt;p&gt;A simple chatbot makes one API call per user message. An AI agent might make 20–50 calls to complete a single task—mixing reasoning models, fast models for classification, tool-use calls, and code execution. Without a gateway, you have no visibility into what your agents are actually doing, what they're costing you, or whether they're behaving correctly.&lt;/p&gt;

&lt;p&gt;The old problem of "too many shovels (models), too little gold (control)" didn't go away. It got &lt;em&gt;worse&lt;/em&gt;. Now the shovels are wielding themselves.&lt;/p&gt;

&lt;h3&gt;
  
  
  MCP made tool integration universal
&lt;/h3&gt;

&lt;p&gt;MCP has emerged as the standard for connecting AI models to external tools and data sources. Your agents now talk to Slack, Notion, databases, browsers, and internal APIs through MCP servers. An AI gateway sitting at this boundary is the natural enforcement point for access control, rate limiting, and audit logging—the same role API gateways have played for REST traffic for over a decade.&lt;/p&gt;

&lt;h3&gt;
  
  
  Multi-model is now multi-everything
&lt;/h3&gt;

&lt;p&gt;In 2025, "multi-model" meant switching between OpenAI and Anthropic. In 2026, a single workflow might use Claude Opus for deep reasoning, Haiku for fast triage, a fine-tuned open-source model for domain-specific tasks, and a local model for sensitive data that can't leave your network. Intelligent routing across this matrix, factoring in cost, latency, capability, and data residency, is exactly what gateways are built for.&lt;/p&gt;

&lt;h2&gt;
  
  
  How do they actually work in 2026?
&lt;/h2&gt;

&lt;p&gt;The architecture has evolved from simple request proxying to &lt;strong&gt;session-aware orchestration&lt;/strong&gt;:&lt;/p&gt;

&lt;p&gt;As a category, we're converging on AI gateways that:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Intercept every LLM call, tool invocation, and agent action that passes through your stack&lt;/li&gt;
&lt;li&gt;Route to the right model based on task complexity, cost budget, latency requirements, and data sensitivity&lt;/li&gt;
&lt;li&gt;Track sessions across multi-step agent workflows, not just individual prompt/response pairs&lt;/li&gt;
&lt;li&gt;Enforce guardrails like content filtering, PII detection, and compliance rules at the gateway layer rather than in each application&lt;/li&gt;
&lt;li&gt;Give you full traces of agent behavior: what models were called, what tools were used, what data was accessed, and what it all cost&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://ngrok.com/docs/ai-gateway/overview" rel="noopener noreferrer"&gt;ngrok's AI gateway&lt;/a&gt; already handles several of these today: it intercepts LLM calls at the SDK level, routes across providers with automatic failover and cost-based selection, and manages API keys so your team doesn't have to. Guardrails like PII redaction, prompt injection detection, and compliance filtering are on the roadmap. If you've ever used ngrok's &lt;a href="https://ngrok.com/blog/endpoint-pools-load-balance-anything" rel="noopener noreferrer"&gt;Endpoint Pools&lt;/a&gt;, the pattern will feel familiar: a pool of endpoints behind a single intelligent entry point that distributes requests for reliability and performance.&lt;/p&gt;

&lt;h2&gt;
  
  
  Do you actually need one now?
&lt;/h2&gt;

&lt;p&gt;Our advice has shifted since 2025:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Scenario&lt;/th&gt;
&lt;th&gt;2025 advice&lt;/th&gt;
&lt;th&gt;2026 advice&lt;/th&gt;
&lt;th&gt;Why it changed&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Single model, simple chatbot&lt;/td&gt;
&lt;td&gt;Skip it&lt;/td&gt;
&lt;td&gt;Still probably skip it&lt;/td&gt;
&lt;td&gt;No agent behavior means your SDK still handles the basics&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Multiple models, production app&lt;/td&gt;
&lt;td&gt;Consider it&lt;/td&gt;
&lt;td&gt;Yes&lt;/td&gt;
&lt;td&gt;Multi-model routing now spans cost, latency, capability, and data residency&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Agentic workflows in production&lt;/td&gt;
&lt;td&gt;Barely existed&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;Essential&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;A single request can trigger 20–50 LLM calls, tool uses, and reasoning chains&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Regulated industry (healthcare, finance)&lt;/td&gt;
&lt;td&gt;Recommended&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;Non-negotiable&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Agents accessing tools and data via MCP need auditable access control&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Internal tools with MCP integrations&lt;/td&gt;
&lt;td&gt;N/A&lt;/td&gt;
&lt;td&gt;Strongly recommended&lt;/td&gt;
&lt;td&gt;MCP made tool integration universal, and gateways are the natural policy layer&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The threshold has dropped. If you're running &lt;em&gt;any&lt;/em&gt; agentic AI in production (and in 2026, most teams are), you need visibility and control over that traffic. An AI gateway gives you both.&lt;/p&gt;

&lt;p&gt;The only teams that can safely skip an AI gateway are those making straightforward, single-model API calls with no agent behavior. If your AI does more than answer questions, if it &lt;em&gt;takes actions&lt;/em&gt;, you want a gateway watching.&lt;/p&gt;

&lt;h2&gt;
  
  
  The future: agent-aware networking
&lt;/h2&gt;

&lt;p&gt;The prediction from our 2025 post is already coming true. AI gateways are evolving into &lt;strong&gt;agent-aware networking layers&lt;/strong&gt; that handle not just routing and security, but also semantic caching (why re-run an expensive reasoning chain for a query you've seen before?), cross-agent coordination, and workload balancing between providers the way CDNs distribute content globally.&lt;/p&gt;

&lt;p&gt;Here's where things sit on the modern AI infrastructure stack:&lt;/p&gt;



&lt;p&gt;The question is no longer &lt;em&gt;whether&lt;/em&gt; you need an AI gateway. It's whether your current infrastructure can handle the agent traffic that's already flowing through it.&lt;/p&gt;

&lt;h2&gt;
  
  
  Be part of what's next
&lt;/h2&gt;

&lt;p&gt;&lt;a href="http://ngrok.ai/" rel="noopener noreferrer"&gt;ngrok.ai&lt;/a&gt; is live, and we're building the next generation of AI-aware networking infrastructure. Follow along on &lt;a href="https://x.com/ngrokHQ" rel="noopener noreferrer"&gt;X&lt;/a&gt;, &lt;a href="http://linkedin.com/company/ngrok/" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;, &lt;a href="https://bsky.app/profile/ngrok.com" rel="noopener noreferrer"&gt;Bluesky&lt;/a&gt;, and &lt;a href="https://www.youtube.com/@ngrokHQ" rel="noopener noreferrer"&gt;YouTube&lt;/a&gt; for what's coming next.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>aiops</category>
    </item>
    <item>
      <title>How Shinobi Security gets access to apps with internal DNS names or redirect flows to run AI agents</title>
      <dc:creator>Sam Richard</dc:creator>
      <pubDate>Fri, 18 Jul 2025 18:38:23 +0000</pubDate>
      <link>https://forem.com/sam_richard/how-shinobi-security-gets-access-to-apps-with-internal-dns-names-or-redirect-flows-to-run-ai-agents-1k8o</link>
      <guid>https://forem.com/sam_richard/how-shinobi-security-gets-access-to-apps-with-internal-dns-names-or-redirect-flows-to-run-ai-agents-1k8o</guid>
      <description>&lt;p&gt;&lt;a href="https://shinobi.security/" rel="noopener noreferrer"&gt;Shinobi Security&lt;/a&gt; is like renting a ethical hacker—but this one's AI and never sleeps, while also being trained alongside the devs that built the app they're testing. &lt;/p&gt;

&lt;p&gt;Shinobi offers their customers teams of AI agents. These agents collaborate, escalate privileges, chain vulnerabilities, and think like real attackers, because their creators used to be ones.&lt;/p&gt;

&lt;p&gt;Developers treat these agents just like they would security teammates. As soon as developers create a new version of their apps, they prompt the agent with a bit of context about the app and how it is supposed to work, and the agents get to work hacking at the new app and attempting to find vulnerabilities. &lt;/p&gt;

&lt;p&gt;Shinobi flips the script of most security tools: instead of drowning you in dashboards and warnings, it proves your vulnerabilities with working exploits. You don’t get a random alert. You get told, “You will get hacked unless you fix this, and here’s how.”&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How does Shinobi get access to customers' apps?&lt;/strong&gt;&lt;br&gt;
Customers didn’t want even Shinobi's AI hackers testing on production apps, so they often requested that Shinobi agents work within the confines of staging or dev environments. Shinobi’s product works wonderfully once it’s set up in a customer’s environment, but customer environments can be tricky.&lt;/p&gt;

&lt;p&gt;Applications often required custom headers or authentication credentials just to honor requests from Shinobi's agent, which meant they would have to customize their solution for every nuance and oddity of those dev/staging environments. &lt;/p&gt;

&lt;p&gt;Shinobi Security selected &lt;a href="//ngrok.com"&gt;ngrok&lt;/a&gt; to create public endpoints for local machines within their customers' networks. ngrok's flexibility and ease of configuration were pivotal in their decision—Varun was able to get a PoC up and running in minutes. &lt;/p&gt;

&lt;p&gt;Implementing ngrok reduced the setup time to run penetration tests on their customers' apps to just 15 minutes.&lt;/p&gt;

&lt;p&gt;Even through ngrok simplified how Shinobi accessed customer networks and shrunk the time-to-test, they sometimes ran into other hurdles. &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;For example, some agents would need access to an app within a customer’s network that had internal DNS names or redirect flows that would break in an outbound tunnelled environment (e.g., 302 redirects to internal.corp.local). &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;To solve for this common issue (that devs usually tie a bunch of different tools together to fix) Shinobi wrote a unique policy unique &lt;a href="https://ngrok.com/docs/traffic-policy/" rel="noopener noreferrer"&gt;traffic policy&lt;/a&gt; to intercept and rewrite 302 redirect headers coming from customer apps, preserving UX and agent behavior.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;on_http_request:
  - actions:
      - type: add-headers
        config:
          headers:
            host: 127.0.0.1
      - type: forward-internal
        config:
          url: https://service-01.customer-abc.internal

on_http_response:
  - expressions:
      - "res.status_code == 302"
    actions:
      - type: set-vars
        config:
          vars:
            orig: "${url.parse(res.location).path}"
      - type: remove-headers
        config:
          headers:
            - Location
      - type: add-headers
        config:
          headers:
            Location: "$NGROK_DOMAIN/${vars.orig}"

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;As they expand their customer base, Shinobi also imagines writing other policies for ensuring their AI agents can properly authenticate themselves even in dev/staging environments.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>security</category>
      <category>networking</category>
      <category>devops</category>
    </item>
  </channel>
</rss>
