<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Catchpoint</title>
    <description>The latest articles on Forem by Catchpoint (@catchpoint).</description>
    <link>https://forem.com/catchpoint</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Forganization%2Fprofile_image%2F11080%2F8b0483cf-31ab-4146-8ca4-96d86da83c4d.png</url>
      <title>Forem: Catchpoint</title>
      <link>https://forem.com/catchpoint</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/catchpoint"/>
    <language>en</language>
    <item>
      <title>Semantic Caching: What We Measured, Why It Matters</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 27 Oct 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/semantic-caching-what-we-measured-why-it-matters-802</link>
      <guid>https://forem.com/catchpoint/semantic-caching-what-we-measured-why-it-matters-802</guid>
      <description>&lt;p&gt;Semantic caching promises to make AI systems faster and cheaper by reducing duplicate calls to large language models (LLMs). But what happens when it doesn’t work as expected?&lt;/p&gt;

&lt;p&gt;We built a test environment to find out. Through a caching system, we evaluated how semantically similar queries would behave. When the cache worked, response times were fast. When it didn’t, things got expensive. In fact, a single semantic cache miss increased latency by more than 2.5x. These failures didn’t show up in the API logs. But they cost real time and real money.&lt;/p&gt;

&lt;p&gt;This blog shares our findings, why semantic caching matters, what causes it to fail, and how to monitor it effectively.&lt;/p&gt;

&lt;p&gt;Traditional caching stores responses based on exact matches. If you’ve asked the same question before—verbatim—you’ll get a fast answer from cache. But if you phrase it differently, even slightly, it’s treated as a brand-new request.&lt;/p&gt;

&lt;p&gt;Semantic caching is different. Instead of matching by text, it matches by meaning.&lt;/p&gt;

&lt;p&gt;It uses embeddings—numerical representations of the intent behind a query—to determine whether a new request is close enough to a previous one. If two queries are semantically similar, the system can return the same result without reprocessing it through the LLM.&lt;/p&gt;

&lt;p&gt;This is especially useful in AI systems where users might ask the same thing in dozens of ways. With semantic caching, “Who’s the President of the US?” and “Who runs America?” can trigger the same cached response—saving time, compute, and cost.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why semantic caching matters in Agentic AI systems
&lt;/h2&gt;

&lt;p&gt;Agentic AI systems don’t just respond to commands—they plan, reason, and act across multiple steps. Each of those steps often involves an LLM call: retrieving documents, rephrasing responses, or deciding what to do next.&lt;/p&gt;

&lt;p&gt;The problem? LLM calls are expensive, especially when repeated across variations of the same question.&lt;/p&gt;

&lt;p&gt;Instead of reprocessing every variation of a query, it can reuse results from previous, similar requests—so long as the meaning is close enough.&lt;/p&gt;

&lt;p&gt;That’s where things get risky: in an agentic AI world, a silent failure in semantic caching doesn’t just mean a slower API call—it can derail entire multistep AI workflows. When semantic cache misses occur, queries go straight to the backend LLM—creating higher latency and skyrocketing costs. Worse yet, these failures are often invisible.  &lt;/p&gt;

&lt;p&gt;Your API returns a 200 OK, but behind the scenes, your cost and performance are suffering.&lt;/p&gt;

&lt;p&gt;Bottom line: Unlike traditional caching, semantic caches introduce new risks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Sudden model updates change embeddings and break matches.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Vector drift causes cache misses even for similar queries.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Users phrase things differently, leading to unexpected misses.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What we observed in the lab
&lt;/h2&gt;

&lt;p&gt;To experiment with how semantic caching affects user experience and infrastructure cost, we built a testing lab environment locally.  &lt;/p&gt;

&lt;p&gt;Here’s how it worked:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;We created a FastAPI application running locally, exposing an endpoint &lt;strong&gt;/search&lt;/strong&gt; via a public endpoint to make it reachable.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Every incoming search query followed this logic:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Cosmos DB Lookup&lt;/strong&gt; → Check whether we’ve seen a similar semantic query before, using the query text as the key.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;✅ &lt;strong&gt;Cache Hit&lt;/strong&gt; → Return the cached vector embedding directly, saving time and cost.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;❌ &lt;strong&gt;Cache Miss&lt;/strong&gt; → Call Gemini Pro to generate a new vector embedding for the query, then store it in Cosmos DB for future reuse.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;We sent the embedding to Azure AI Search, performing a vector similarity search to find the top matching documents.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Finally, we returned the search results plus two crucial custom HTTP headers:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;X-Semantic-Cache: hit or miss&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;X-Semantic-Score (e.g. 0.8543) indicating how close the new query is to previous ones.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Test results
&lt;/h2&gt;

&lt;p&gt;We configured a Catchpoint test to simulate user queries to our public endpoint URL. The test:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Sent randomized prompts (e.g. “NYC weather,” “New York forecast”) to trigger both cache hits and misses.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Captured and parsed the semantic headers using regex in Catchpoint to track:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cache efficiency (% of hits vs. misses)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Semantic similarity scores&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Visualized this data in dashboards to see latency differences between hits and misses.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This gave us real evidence of how semantic caching reduces API calls to the expensive LLM backend and improves response times—insights we could quantify directly in the Catchpoint &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) portal.  &lt;/p&gt;

&lt;p&gt;Below, are some helpful insights from our testing.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx49uwc0rlfdumaftisly.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx49uwc0rlfdumaftisly.png" alt="A graph with a line and a lineDescription automatically generated" width="800" height="344"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Cache Hits vs. Misses: Measurable Latency Gap&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;The trend line in the graph above shows that overall, the response time was about 50% to 250% higher when semantic cache returned a MISS compared to a HIT.&lt;/p&gt;

&lt;p&gt;Diving deeper, we observed that the first run of a prompt went to the backend (a cache miss), with higher latency and costs.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F32ao4nts5604pjhk505v.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F32ao4nts5604pjhk505v.png" alt="A screenshot of a computerDescription automatically generated" width="800" height="412"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;First Run: Cache Miss with High Latency&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;The second run of the same semantic query hit the cache, cutting response times by 40%.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsvmsw0jds9s2haxkzw6e.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsvmsw0jds9s2haxkzw6e.png" alt="A screenshot of a computerDescription automatically generated" width="800" height="652"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5w3wg4ano2he2e329ck1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5w3wg4ano2he2e329ck1.png" width="800" height="68"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Second Run: Served from Cache (hit). Same semantic score (prompt matched). Response time is 50% lower&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Monitoring strategies to make semantic caching reliable
&lt;/h2&gt;

&lt;p&gt;Semantic caching is no longer a background optimization—it’s a core pillar of agentic AI systems that reason and act in real-time. But to trust it, we need to measure how well it’s working.  &lt;/p&gt;

&lt;p&gt;Here are three ways to monitor and improve its reliability:&lt;/p&gt;

&lt;h1&gt;
  
  
  1. Test semantically similar queries
&lt;/h1&gt;




&lt;p&gt;Semantic caching lives or dies on how well it matches similar questions. Use synthetic monitoring to simulate different phrasings of the same intent:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;“Who’s the President of the US?”&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;“Who runs the US government?”&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;“Commander-in-Chief of America?”&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Then compare their outcomes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Did they result in cache hits or misses?&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;What were the semantic similarity scores?&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This gives you visibility into whether your caching system is recognizing intent consistently. For agentic AI, it’s not enough for one query to be fast. You need confidence that &lt;em&gt;all&lt;/em&gt; user expressions of the same intent are covered.&lt;/p&gt;

&lt;h1&gt;
  
  
  2. Track semantic similarity scores
&lt;/h1&gt;




&lt;p&gt;Semantic caches often return a similarity score (e.g. 0.85) to indicate how close the new query is to an existing cached answer. If your cache system returns a similarity score (e.g. 0.8343), you can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Monitor it over time&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Visualize trends&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Alert if scores drop below thresholds&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For instance, in our tests, both requests returned the same semantic score of 0.85224825.  &lt;/p&gt;

&lt;p&gt;But if the model changes or query phrasing drifts, scores could drop, leading to unexpected misses and rising costs.&lt;/p&gt;

&lt;p&gt;Monitoring these numbers ensures your semantic cache stays reliable—and that you’re not wasting money on backend calls unnecessarily.&lt;/p&gt;

&lt;h1&gt;
  
  
  3. Measure real-world latency differences
&lt;/h1&gt;




&lt;p&gt;One of the biggest promises of semantic caching is speed. Cache hits should be significantly faster than misses.&lt;/p&gt;

&lt;p&gt;Monitoring this can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Split metrics for cache hits vs. misses&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Show precise latency differences&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Alert when cache misses cause slowdowns&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;From our test results:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Cache miss response time: ~5 seconds&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cache hit response time: ~2 seconds&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That’s a 2.5x speedup. In the world of agentic AI, that gap is the difference between a seamless conversation—and a frustrating pause.&lt;/p&gt;

&lt;h2&gt;
  
  
  Final takeaway
&lt;/h2&gt;

&lt;p&gt;Semantic caching isn’t just a nice-to-have—it’s becoming core infrastructure for real-time AI systems. Cloud leaders like Fastly, AWS, and Azure are already baking it into their architectures. But it’s also uniquely fragile. Changes in language, embedding drift, or model updates can quietly degrade performance.&lt;/p&gt;

&lt;p&gt;By combining semantic caching with IPM*&lt;em&gt;,&lt;/em&gt;* teams can ensure that their systems are not only fast—but reliably so.&lt;/p&gt;

&lt;p&gt;If you're running AI agents at scale, silent cache failures aren't just inefficiencies. They're risks. Measure them, monitor them, and mitigate them.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn more:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.catchpoint.com/solutions/agentic-ai-resilience" rel="noopener noreferrer"&gt;&lt;strong&gt;Agentic AI Resilience Monitoring&lt;/strong&gt;&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.catchpoint.com/solutions/monitor-ai-assistant" rel="noopener noreferrer"&gt;&lt;strong&gt;AI Assistant Performance Monitoring&lt;/strong&gt;&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;Semantic caching promises to make AI systems faster and cheaper by reducing duplicate calls to large language models (LLMs). But what happens when it doesn’t work as expected?&lt;/p&gt;

&lt;p&gt;We built a test environment to find out. Through a caching system, we evaluated how semantically similar queries would behave. When the cache worked, response times were fast. When it didn’t, things got expensive. In fact, a single semantic cache miss increased latency by more than 2.5x. These failures didn’t show up in the API logs. But they cost real time and real money.&lt;/p&gt;

&lt;p&gt;This blog shares our findings, why semantic caching matters, what causes it to fail, and how to monitor it effectively.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is semantic caching?
&lt;/h2&gt;

&lt;p&gt;Traditional caching stores responses based on exact matches. If you’ve asked the same question before—verbatim—you’ll get a fast answer from cache. But if you phrase it differently, even slightly, it’s treated as a brand-new request.&lt;/p&gt;

&lt;p&gt;Semantic caching is different. Instead of matching by text, it matches by meaning.&lt;/p&gt;

&lt;p&gt;It uses embeddings—numerical representations of the intent behind a query—to determine whether a new request is close enough to a previous one. If two queries are semantically similar, the system can return the same result without reprocessing it through the LLM.&lt;/p&gt;

&lt;p&gt;This is especially useful in AI systems where users might ask the same thing in dozens of ways. With semantic caching, “Who’s the President of the US?” and “Who runs America?” can trigger the same cached response—saving time, compute, and cost.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why semantic caching matters in Agentic AI systems
&lt;/h2&gt;

&lt;p&gt;Agentic AI systems don’t just respond to commands—they plan, reason, and act across multiple steps. Each of those steps often involves an LLM call: retrieving documents, rephrasing responses, or deciding what to do next.&lt;/p&gt;

&lt;p&gt;The problem? LLM calls are expensive, especially when repeated across variations of the same question.&lt;/p&gt;

&lt;p&gt;Instead of reprocessing every variation of a query, it can reuse results from previous, similar requests—so long as the meaning is close enough.&lt;/p&gt;

&lt;p&gt;That’s where things get risky: in an agentic AI world, a silent failure in semantic caching doesn’t just mean a slower API call—it can derail entire multistep AI workflows. When semantic cache misses occur, queries go straight to the backend LLM—creating higher latency and skyrocketing costs. Worse yet, these failures are often invisible.  &lt;/p&gt;

&lt;p&gt;Your API returns a 200 OK, but behind the scenes, your cost and performance are suffering.&lt;/p&gt;

&lt;p&gt;Bottom line: Unlike traditional caching, semantic caches introduce new risks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Sudden model updates change embeddings and break matches.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Vector drift causes cache misses even for similar queries.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Users phrase things differently, leading to unexpected misses.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What we observed in the lab
&lt;/h2&gt;

&lt;p&gt;To experiment with how semantic caching affects user experience and infrastructure cost, we built a testing lab environment locally.  &lt;/p&gt;

&lt;p&gt;Here’s how it worked:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;We created a FastAPI application running locally, exposing an endpoint &lt;strong&gt;/search&lt;/strong&gt; via a public endpoint to make it reachable.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Every incoming search query followed this logic:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Cosmos DB Lookup&lt;/strong&gt; → Check whether we’ve seen a similar semantic query before, using the query text as the key.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;✅ &lt;strong&gt;Cache Hit&lt;/strong&gt; → Return the cached vector embedding directly, saving time and cost.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;❌ &lt;strong&gt;Cache Miss&lt;/strong&gt; → Call Gemini Pro to generate a new vector embedding for the query, then store it in Cosmos DB for future reuse.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;We sent the embedding to Azure AI Search, performing a vector similarity search to find the top matching documents.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Finally, we returned the search results plus two crucial custom HTTP headers:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;X-Semantic-Cache: hit or miss&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;X-Semantic-Score (e.g. 0.8543) indicating how close the new query is to previous ones.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Test results
&lt;/h2&gt;

&lt;p&gt;We configured a Catchpoint test to simulate user queries to our public endpoint URL. The test:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Sent randomized prompts (e.g. “NYC weather,” “New York forecast”) to trigger both cache hits and misses.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Captured and parsed the semantic headers using regex in Catchpoint to track:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cache efficiency (% of hits vs. misses)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Semantic similarity scores&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Visualized this data in dashboards to see latency differences between hits and misses.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This gave us real evidence of how semantic caching reduces API calls to the expensive LLM backend and improves response times—insights we could quantify directly in the Catchpoint &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) portal.  &lt;/p&gt;

&lt;p&gt;Below, are some helpful insights from our testing.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx49uwc0rlfdumaftisly.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx49uwc0rlfdumaftisly.png" alt="A graph with a line and a lineDescription automatically generated" width="800" height="344"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Cache Hits vs. Misses: Measurable Latency Gap&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;The trend line in the graph above shows that overall, the response time was about 50% to 250% higher when semantic cache returned a MISS compared to a HIT.&lt;/p&gt;

&lt;p&gt;Diving deeper, we observed that the first run of a prompt went to the backend (a cache miss), with higher latency and costs.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F32ao4nts5604pjhk505v.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F32ao4nts5604pjhk505v.png" alt="A screenshot of a computerDescription automatically generated" width="800" height="412"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;First Run: Cache Miss with High Latency&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;The second run of the same semantic query hit the cache, cutting response times by 40%.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsvmsw0jds9s2haxkzw6e.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsvmsw0jds9s2haxkzw6e.png" alt="A screenshot of a computerDescription automatically generated" width="800" height="652"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5w3wg4ano2he2e329ck1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5w3wg4ano2he2e329ck1.png" width="800" height="68"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Second Run: Served from Cache (hit). Same semantic score (prompt matched). Response time is 50% lower&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Monitoring strategies to make semantic caching reliable
&lt;/h2&gt;

&lt;p&gt;Semantic caching is no longer a background optimization—it’s a core pillar of agentic AI systems that reason and act in real-time. But to trust it, we need to measure how well it’s working.  &lt;/p&gt;

&lt;p&gt;Here are three ways to monitor and improve its reliability:&lt;/p&gt;

&lt;h1&gt;
  
  
  1. Test semantically similar queries
&lt;/h1&gt;




&lt;p&gt;Semantic caching lives or dies on how well it matches similar questions. Use synthetic monitoring to simulate different phrasings of the same intent:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;“Who’s the President of the US?”&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;“Who runs the US government?”&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;“Commander-in-Chief of America?”&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Then compare their outcomes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Did they result in cache hits or misses?&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;What were the semantic similarity scores?&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This gives you visibility into whether your caching system is recognizing intent consistently. For agentic AI, it’s not enough for one query to be fast. You need confidence that &lt;em&gt;all&lt;/em&gt; user expressions of the same intent are covered.&lt;/p&gt;

&lt;h1&gt;
  
  
  2. Track semantic similarity scores
&lt;/h1&gt;




&lt;p&gt;Semantic caches often return a similarity score (e.g. 0.85) to indicate how close the new query is to an existing cached answer. If your cache system returns a similarity score (e.g. 0.8343), you can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Monitor it over time&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Visualize trends&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Alert if scores drop below thresholds&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For instance, in our tests, both requests returned the same semantic score of 0.85224825.  &lt;/p&gt;

&lt;p&gt;But if the model changes or query phrasing drifts, scores could drop, leading to unexpected misses and rising costs.&lt;/p&gt;

&lt;p&gt;Monitoring these numbers ensures your semantic cache stays reliable—and that you’re not wasting money on backend calls unnecessarily.&lt;/p&gt;

&lt;h1&gt;
  
  
  3. Measure real-world latency differences
&lt;/h1&gt;




&lt;p&gt;One of the biggest promises of semantic caching is speed. Cache hits should be significantly faster than misses.&lt;/p&gt;

&lt;p&gt;Monitoring this can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Split metrics for cache hits vs. misses&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Show precise latency differences&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Alert when cache misses cause slowdowns&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;From our test results:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Cache miss response time: ~5 seconds&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cache hit response time: ~2 seconds&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That’s a 2.5x speedup. In the world of agentic AI, that gap is the difference between a seamless conversation—and a frustrating pause.&lt;/p&gt;

&lt;h2&gt;
  
  
  Final takeaway
&lt;/h2&gt;

&lt;p&gt;Semantic caching isn’t just a nice-to-have—it’s becoming core infrastructure for real-time AI systems. Cloud leaders like Fastly, AWS, and Azure are already baking it into their architectures. But it’s also uniquely fragile. Changes in language, embedding drift, or model updates can quietly degrade performance.&lt;/p&gt;

&lt;p&gt;By combining semantic caching with IPM*&lt;em&gt;,&lt;/em&gt;* teams can ensure that their systems are not only fast—but reliably so.&lt;/p&gt;

&lt;p&gt;If you're running AI agents at scale, silent cache failures aren't just inefficiencies. They're risks. Measure them, monitor them, and mitigate them.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn more:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.catchpoint.com/solutions/agentic-ai-resilience" rel="noopener noreferrer"&gt;&lt;strong&gt;Agentic AI Resilience Monitoring&lt;/strong&gt;&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.catchpoint.com/solutions/monitor-ai-assistant" rel="noopener noreferrer"&gt;&lt;strong&gt;AI Assistant Performance Monitoring&lt;/strong&gt;&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

</description>
      <category>ai</category>
    </item>
    <item>
      <title>Here’s the proof: What the fastest sites on the web have in common</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 20 Oct 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/heres-the-proof-what-the-fastest-sites-on-the-web-have-in-common-4081</link>
      <guid>https://forem.com/catchpoint/heres-the-proof-what-the-fastest-sites-on-the-web-have-in-common-4081</guid>
      <description>&lt;p&gt;60% of Gen Z won’t engage with a slow-loading website. In today’s digital economy, that’s a deal-breaker. Whether it’s a banking portal, a travel app, or an AI-powered SaaS platform, &lt;a href="https://www.catchpoint.com/statistics" rel="noopener noreferrer"&gt;users expect performance&lt;/a&gt;. Instant loading, global reliability, and smooth interactivity aren’t just nice to have—they define the winners.&lt;/p&gt;

&lt;p&gt;At Catchpoint, we’ve spent the past several years benchmarking the web performance of the world’s most recognizable brands—from HSBC and Google Flights to Salesforce and Nike. Our industry benchmark reports span sectors including &lt;a href="https://www.catchpoint.com/asset/banking-website-performance-benchmark-report-2025" rel="noopener noreferrer"&gt;banking&lt;/a&gt;, &lt;a href="https://www.catchpoint.com/learn/gen-ai-benchmark" rel="noopener noreferrer"&gt;GenAI&lt;/a&gt;, &lt;a href="https://www.catchpoint.com/asset/airline-website-performance-benchmark-report" rel="noopener noreferrer"&gt;airlines&lt;/a&gt;, &lt;a href="https://www.catchpoint.com/asset/hotel-website-performance-benchmark-report" rel="noopener noreferrer"&gt;hotels&lt;/a&gt;, &lt;a href="https://www.catchpoint.com/asset/travel-website-performance-benchmark-report" rel="noopener noreferrer"&gt;travel aggregators&lt;/a&gt;, and &lt;a href="https://www.catchpoint.com/asset/athletic-footwear-website-performance-benchmark-report" rel="noopener noreferrer"&gt;athletic footwear and apparel&lt;/a&gt;.  &lt;/p&gt;

&lt;p&gt;Each report is powered by our industry leading &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;IPM platform&lt;/a&gt; and &lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;Global Agent Network&lt;/a&gt; of over 3,000+ agents worldwide. This data-driven approach gives us a uniquely authoritative view into what separates the fastest sites on the web from the rest.&lt;/p&gt;

&lt;p&gt;This blog distils the key findings from our latest benchmark reports: six traits the fastest sites share, and how organizations across any industry can apply them to build faster, more resilient digital experiences.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#1. Lightning-fast infrastructure: DNS &amp;amp; TTFB&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;One common trait is a rock-solid network foundation – &lt;a href="https://www.catchpoint.com/asset/the-need-for-speed" rel="noopener noreferrer"&gt;quick DNS lookups&lt;/a&gt; and rapid server responses. The fastest sites minimize any lag between a user’s request and the first byte of data returned:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Optimized DNS:&lt;/strong&gt; Across industries, the fastest sites keep DNS resolution times extremely low—often under 50 ms. A quick DNS lookup shaves valuable milliseconds off the initial request and helps the page begin loading faster.&lt;/p&gt;

&lt;p&gt;For instance, in our 2024 benchmark of &lt;a href="https://www.catchpoint.com/asset/athletic-footwear-website-performance-benchmark-report" rel="noopener noreferrer"&gt;athletic footwear and apparel brands&lt;/a&gt;, top performers like Kappa and Under Armour resolved DNS in under 1 ms, while others such as Skechers and On remained comfortably below the 50 ms threshold. At the other end of the spectrum, slower-loading sites like Nike and VEJA saw DNS times of 169–187 ms—over 3× slower than the best performers.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5eppmkx96f5ilka9l2ad.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5eppmkx96f5ilka9l2ad.png" alt="A graph of blue barsAI-generated content may be incorrect., Picture" width="800" height="445"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This early delay often signals broader infrastructure inefficiencies. In nearly every sector we’ve analyzed, lower DNS times strongly correlate with faster overall load speeds.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Snappy TTFB:&lt;/strong&gt; Likewise, elite sites boast backend servers that respond in a heartbeat. The top two &lt;a href="https://www.catchpoint.com/asset/banking-website-performance-benchmark-report-2025" rel="noopener noreferrer"&gt;banking&lt;/a&gt; websites all returned server responses in under 200 ms, providing a swift handoff to the browser.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flcohrdrx1jgykqgz7wtm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flcohrdrx1jgykqgz7wtm.png" alt="A graph of blue squaresAI-generated content may be incorrect., Picture" width="800" height="321"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Banking Website Performance Benchmark report 2025&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;But in the athletic apparel space, only Puma and Under Armour hit that bar (101 ms and 107 ms, respectively).  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjbmjmaw4ujz24k1juuvg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjbmjmaw4ujz24k1juuvg.png" alt="A graph of blue bars with white textAI-generated content may be incorrect., Picture" width="800" height="451"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Many others struggled: Nike, VEJA, and Kappa all had TTFB exceeding 950 ms, with Nike peaking at 1.1 seconds—more than 5× slower than the recommended threshold.&lt;/p&gt;

&lt;p&gt;This kind of server-side delay often negates gains made elsewhere, like in front-end optimization or CDN use. Across all verticals we benchmarked, fast TTFB was a strong predictor of high overall performance.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#2. Globally distributed delivery&lt;/strong&gt;  
&lt;/h2&gt;

&lt;p&gt;The fastest websites don’t just perform well in one country – they invest in global delivery infrastructure, so users everywhere get quick load times. Our data shows that sites leveraging Content Delivery Networks (CDNs) and regional optimizations avoid the huge geographic performance gaps that plague others:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Localized CDNs&lt;/strong&gt;: Leading sites deploy globally distributed servers and CDN endpoints to bring content closer to users. In &lt;a href="https://www.catchpoint.com/asset/banking-website-performance-benchmark-report-2025" rel="noopener noreferrer"&gt;banking&lt;/a&gt;, for example, the highest performers (UBS, ING, HSBC, etc.) achieved consistent experiences worldwide by using localized infrastructure and optimized DNS routing. In contrast, sites without regional presence suffered severe slowdowns – in Africa and South America, some banks had page load times double those in North America. (One test showed 10.7 s waits in Africa vs ~5 s in the U.S.)&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiodsxfsexe4tpk4qisp3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiodsxfsexe4tpk4qisp3.png" alt="A graph of blue squaresAI-generated content may be incorrect., Picture" width="800" height="431"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;SaaS Website Performance Benchmark Report 2025&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Consistent speed everywhere:&lt;/strong&gt; The data confirms that being fast everywhere is part of being a fastest-in-class website. While lower-ranked sites still show wide regional performance gaps, top performers prove it’s possible to deliver fast, reliable experiences globally—though Africa remains a challenge for all.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv5if2pubk8rygruww0e2.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv5if2pubk8rygruww0e2.jpeg" alt="Regional disparities in document complete times across continents, showing Africa significantly lagging behind other regions, Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Regional disparities in document complete times across continents in the 2025 Banking Report, showing Africa significantly lagging behind other regions&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3b0xkjttqm5yhg84apie.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3b0xkjttqm5yhg84apie.jpeg" alt="Performance ratios reveal dramatic disparities: Africa and South America take more than double the time to complete documents compared to North America, Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Performance ratios reveal dramatic disparities: In the 2025 Banking Report, Africa and South America take more than double the time to complete documents compared to North America&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#3. Lean pages, fast Loads&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Another hallmark of the web’s fastest sites is lightweight, efficiently coded pages that reach full load in just a few seconds. They avoid the bloat and complexity that drag others down:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Blazing-fast loads:&lt;/strong&gt; The elite websites consistently fully load in ~2–3 seconds or less. In banking, only 25% of sites met the ideal of &amp;lt;3 s document complete, but every top-ranked bank did. In fact, the best performers in finance all loaded in under 2 s while maintaining ~99.9% uptime. Similarly, across other industries, top sites hit very low page load times – travel sites like Skyscanner achieved sub-0.7 s Largest Contentful Paint and mere ~1.2 s full loads, indicating extremely fast content delivery.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Minimal baggage:&lt;/strong&gt; In contrast, slower sites tend to be weighed down by heavy content and excess scripts. Many well-known brands ranked outside the top tier due to “heavy content and front-end complexity.” These bulky pages took over 5–9 seconds to load, or in the case of Marriot below, over 10 seconds.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fin16xgyegk1zj6sba3d3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fin16xgyegk1zj6sba3d3.png" alt="A screenshot of a graphAI-generated content may be incorrect., Picture" width="800" height="324"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Hotel Website Performance Benchmark Report&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;The takeaway: fastest sites keep their pages lean – optimizing images, compressing files, and capping third-party elements to avoid sluggish load times.&lt;/p&gt;

&lt;p&gt;By keeping page weight low and complexity in check, top performers ensure users aren’t left staring at blank screens. Fast complete load times not only improve user satisfaction but also reduce bounce rates – a win-win for experience and business metrics.  &lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#4. Stellar front-end optimization &amp;amp; core web vitals&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Beyond just loading quickly, the fastest websites also feel fast and stable to users. They achieve this through meticulous front-end optimization, evidenced by outstanding Core Web Vitals metrics like Largest Contentful Paint (LCP) and Cumulative Layout Shift (CLS):&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fast content rendering (low LCP):&lt;/strong&gt; Top sites prioritize delivering meaningful content immediately. In Catchpoint’s travel industry test, all of the 10 fastest sites had LCP well within recommended standards – many under ~1 second. For example, Skyscanner’s LCP came in around 613 ms, with even the 10th-best still near 1.1 s – comfortably beating the ~2.5 s guideline. This means users see the page’s main content almost instantly on these sites.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd2l4tfm1b5su2m9fzpev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd2l4tfm1b5su2m9fzpev.png" alt="A graph of a documentAI-generated content may be incorrect., Picture" width="800" height="396"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Travel Website Performance Benchmark Report&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Slower sites, on the other hand, often struggle with LCP above the 2.5 s mark, making them feel sluggish even if the total load isn’t horrific.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Stable layouts (no janky moves):&lt;/strong&gt; A striking data point – the top 10 travel websites all had a perfect CLS score of 0.00. Likewise, the best banks and retail sites exhibited virtually no unexpected layout shifts during load. In practice, this means no annoying page jumps (caused by late-loading ads or images) for the user.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffpfgs3onehmqxfz1b94n.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffpfgs3onehmqxfz1b94n.png" alt="A graph of blue bars with white textAI-generated content may be incorrect., Picture" width="800" height="371"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Airline Website Performance Benchmark Report&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Fast sites achieve this by reserving space for images/media and loading content in a controlled way. By contrast, lesser performers saw CLS values up to 0.8–1.5, causing noticeable “jank” and hurting user experience.&lt;/p&gt;

&lt;p&gt;It’s clear that front-end experience is the real differentiator among websites. Sites that not only load quickly but also render smoothly (low LCP) and stay visually stable (low CLS) ended up at the top of Catchpoint’s rankings. These Core Web Vitals were highly correlated with overall performance scores, underscoring that fastest sites focus heavily on user-centric front-end metrics.  &lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#5. High availability without sacrificing speed&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Leading websites understand that availability and speed must go hand in hand. It’s not enough to be up 24/7 if your site is slow or error-prone when loaded. The fastest sites manage to deliver near-perfect uptime and quick performance simultaneously:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Near-perfect uptime:&lt;/strong&gt; The majority of top performers maintain 99.9% or better availability. In the 2025 Banking benchmark, 75% of sites had at least 99.9% uptime – and all of the top finishers were essentially always-online. In fact, the #1 sites in multiple industries achieved ≈100% uptime during the testing period. This level of reliability ensures users rarely encounter errors or downtime.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F22933jp3ffcskn7iuyzz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F22933jp3ffcskn7iuyzz.png" alt="A blue rectangular object with black textAI-generated content may be incorrect., Picture" width="800" height="314"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;2025 Banking Website Performance Benchmark Report&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Reliability and Speed:&lt;/strong&gt; Crucially, the best sites don’t treat uptime as a safety net for poor performance. Availability isn’t the safety net it used to be.Several lower-ranked companies across industries had 99.9% uptime yet still fell behind due to slow pages or unstable fronts. Top sites avoid this trap by pairing high availability with fast, stable loads. They have resilient infrastructures to stay online, and they proactively monitor performance so that being “up” always means delivering a quality experience. Conversely, a few sites had serious outages (dipping to ~90% uptime) which obviously knocked them out of contention.&lt;/p&gt;

&lt;p&gt;The bottom line: reliability is foundational, but true excellence comes from reliability + speed. The fastest websites prioritize both. Users expect a site to be available and to respond instantly – leading sites deliver on that expectation consistently.  &lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#6. Culture of continuous optimization&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Speed isn’t a one-time project. The fastest websites treat performance as an ongoing discipline—one rooted in measurement, iteration, and readiness for peak demand.&lt;/p&gt;

&lt;p&gt;Instacart is a clear example. While their average TTFB during the &lt;a href="https://www.catchpoint.com/blog/why-super-bowl-2025-was-a-triumph-for-internet-resilience" rel="noopener noreferrer"&gt;Super Bowl benchmark&lt;/a&gt; period was a sluggish 1,051 ms, that changed on game day. On February 9th, when their Super Bowl ad aired, Instacart slashed TTFB significantly—demonstrating that with the right optimizations, even high-traffic moments can be performance wins.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4lhlnkd5kvwfasf97tof.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4lhlnkd5kvwfasf97tof.png" alt="A graph showing a line and a red circleAI-generated content may be incorrect., Picture" width="800" height="419"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Instacart’s TTFB improvement on the day of the Super Bowl&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;This kind of readiness doesn’t happen by accident. Across all industries, one of the strongest shared habits of high-performing sites is this culture of continuous optimization. When performance is treated as a living system—not just a launch checklist—organizations are better equipped to handle traffic spikes, user expectations, and the unexpected.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Final thoughts: speed is strategy&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;The patterns are clear. The fastest sites on the web invest in performance at every level: from back-end infrastructure to front-end polish, from global delivery to rigorous uptime, all under a philosophy of ongoing improvement. These sites prove that it’s possible to be both fast and reliable– delighting users with sub-3 second loads, silky-smooth pages, and dependable service. Meanwhile, competitors that neglect these areas struggle with slow load times, unstable pages, or regional outages, putting them at a serious disadvantage.&lt;/p&gt;

&lt;p&gt;The good news is that these best practices are well-understood. As our benchmark reports show, the top performers simply execute them better – and reap the rewards in user experience. By adopting the common habits of the fastest websites, including smart DNS/CDN usage, lightweight design, and Core Web Vitals focus, any website can dramatically improve its speed. After all, as the data reminds us, “it’s no longer enough to be available; you have to be fast everywhere”. The fastest sites have set the bar – and they invite others to catch up.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Explore the full benchmark series&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;See how your industry stacks up &lt;a href="https://www.catchpoint.com/resource-library?*=benchmark+report" rel="noopener noreferrer"&gt;View all benchmark reports&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;60% of Gen Z won’t engage with a slow-loading website. In today’s digital economy, that’s a deal-breaker. Whether it’s a banking portal, a travel app, or an AI-powered SaaS platform, &lt;a href="https://www.catchpoint.com/statistics" rel="noopener noreferrer"&gt;users expect performance&lt;/a&gt;. Instant loading, global reliability, and smooth interactivity aren’t just nice to have—they define the winners.&lt;/p&gt;

&lt;p&gt;At Catchpoint, we’ve spent the past several years benchmarking the web performance of the world’s most recognizable brands—from HSBC and Google Flights to Salesforce and Nike. Our industry benchmark reports span sectors including &lt;a href="https://www.catchpoint.com/asset/banking-website-performance-benchmark-report-2025" rel="noopener noreferrer"&gt;banking&lt;/a&gt;, &lt;a href="https://www.catchpoint.com/learn/gen-ai-benchmark" rel="noopener noreferrer"&gt;GenAI&lt;/a&gt;, &lt;a href="https://www.catchpoint.com/asset/airline-website-performance-benchmark-report" rel="noopener noreferrer"&gt;airlines&lt;/a&gt;, &lt;a href="https://www.catchpoint.com/asset/hotel-website-performance-benchmark-report" rel="noopener noreferrer"&gt;hotels&lt;/a&gt;, &lt;a href="https://www.catchpoint.com/asset/travel-website-performance-benchmark-report" rel="noopener noreferrer"&gt;travel aggregators&lt;/a&gt;, and &lt;a href="https://www.catchpoint.com/asset/athletic-footwear-website-performance-benchmark-report" rel="noopener noreferrer"&gt;athletic footwear and apparel&lt;/a&gt;.  &lt;/p&gt;

&lt;p&gt;Each report is powered by our industry leading &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;IPM platform&lt;/a&gt; and &lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;Global Agent Network&lt;/a&gt; of over 3,000+ agents worldwide. This data-driven approach gives us a uniquely authoritative view into what separates the fastest sites on the web from the rest.&lt;/p&gt;

&lt;p&gt;This blog distils the key findings from our latest benchmark reports: six traits the fastest sites share, and how organizations across any industry can apply them to build faster, more resilient digital experiences.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#1. Lightning-fast infrastructure: DNS &amp;amp; TTFB&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;One common trait is a rock-solid network foundation – &lt;a href="https://www.catchpoint.com/asset/the-need-for-speed" rel="noopener noreferrer"&gt;quick DNS lookups&lt;/a&gt; and rapid server responses. The fastest sites minimize any lag between a user’s request and the first byte of data returned:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Optimized DNS:&lt;/strong&gt; Across industries, the fastest sites keep DNS resolution times extremely low—often under 50 ms. A quick DNS lookup shaves valuable milliseconds off the initial request and helps the page begin loading faster.&lt;/p&gt;

&lt;p&gt;For instance, in our 2024 benchmark of &lt;a href="https://www.catchpoint.com/asset/athletic-footwear-website-performance-benchmark-report" rel="noopener noreferrer"&gt;athletic footwear and apparel brands&lt;/a&gt;, top performers like Kappa and Under Armour resolved DNS in under 1 ms, while others such as Skechers and On remained comfortably below the 50 ms threshold. At the other end of the spectrum, slower-loading sites like Nike and VEJA saw DNS times of 169–187 ms—over 3× slower than the best performers.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5eppmkx96f5ilka9l2ad.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5eppmkx96f5ilka9l2ad.png" alt="A graph of blue barsAI-generated content may be incorrect., Picture" width="800" height="445"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This early delay often signals broader infrastructure inefficiencies. In nearly every sector we’ve analyzed, lower DNS times strongly correlate with faster overall load speeds.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Snappy TTFB:&lt;/strong&gt; Likewise, elite sites boast backend servers that respond in a heartbeat. The top two &lt;a href="https://www.catchpoint.com/asset/banking-website-performance-benchmark-report-2025" rel="noopener noreferrer"&gt;banking&lt;/a&gt; websites all returned server responses in under 200 ms, providing a swift handoff to the browser.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flcohrdrx1jgykqgz7wtm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flcohrdrx1jgykqgz7wtm.png" alt="A graph of blue squaresAI-generated content may be incorrect., Picture" width="800" height="321"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Banking Website Performance Benchmark report 2025&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;But in the athletic apparel space, only Puma and Under Armour hit that bar (101 ms and 107 ms, respectively).  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjbmjmaw4ujz24k1juuvg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjbmjmaw4ujz24k1juuvg.png" alt="A graph of blue bars with white textAI-generated content may be incorrect., Picture" width="800" height="451"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Many others struggled: Nike, VEJA, and Kappa all had TTFB exceeding 950 ms, with Nike peaking at 1.1 seconds—more than 5× slower than the recommended threshold.&lt;/p&gt;

&lt;p&gt;This kind of server-side delay often negates gains made elsewhere, like in front-end optimization or CDN use. Across all verticals we benchmarked, fast TTFB was a strong predictor of high overall performance.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#2. Globally distributed delivery&lt;/strong&gt;  
&lt;/h2&gt;

&lt;p&gt;The fastest websites don’t just perform well in one country – they invest in global delivery infrastructure, so users everywhere get quick load times. Our data shows that sites leveraging Content Delivery Networks (CDNs) and regional optimizations avoid the huge geographic performance gaps that plague others:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Localized CDNs&lt;/strong&gt;: Leading sites deploy globally distributed servers and CDN endpoints to bring content closer to users. In &lt;a href="https://www.catchpoint.com/asset/banking-website-performance-benchmark-report-2025" rel="noopener noreferrer"&gt;banking&lt;/a&gt;, for example, the highest performers (UBS, ING, HSBC, etc.) achieved consistent experiences worldwide by using localized infrastructure and optimized DNS routing. In contrast, sites without regional presence suffered severe slowdowns – in Africa and South America, some banks had page load times double those in North America. (One test showed 10.7 s waits in Africa vs ~5 s in the U.S.)&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiodsxfsexe4tpk4qisp3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiodsxfsexe4tpk4qisp3.png" alt="A graph of blue squaresAI-generated content may be incorrect., Picture" width="800" height="431"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;SaaS Website Performance Benchmark Report 2025&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Consistent speed everywhere:&lt;/strong&gt; The data confirms that being fast everywhere is part of being a fastest-in-class website. While lower-ranked sites still show wide regional performance gaps, top performers prove it’s possible to deliver fast, reliable experiences globally—though Africa remains a challenge for all.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv5if2pubk8rygruww0e2.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv5if2pubk8rygruww0e2.jpeg" alt="Regional disparities in document complete times across continents, showing Africa significantly lagging behind other regions, Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Regional disparities in document complete times across continents in the 2025 Banking Report, showing Africa significantly lagging behind other regions&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3b0xkjttqm5yhg84apie.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3b0xkjttqm5yhg84apie.jpeg" alt="Performance ratios reveal dramatic disparities: Africa and South America take more than double the time to complete documents compared to North America, Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Performance ratios reveal dramatic disparities: In the 2025 Banking Report, Africa and South America take more than double the time to complete documents compared to North America&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#3. Lean pages, fast Loads&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Another hallmark of the web’s fastest sites is lightweight, efficiently coded pages that reach full load in just a few seconds. They avoid the bloat and complexity that drag others down:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Blazing-fast loads:&lt;/strong&gt; The elite websites consistently fully load in ~2–3 seconds or less. In banking, only 25% of sites met the ideal of &amp;lt;3 s document complete, but every top-ranked bank did. In fact, the best performers in finance all loaded in under 2 s while maintaining ~99.9% uptime. Similarly, across other industries, top sites hit very low page load times – travel sites like Skyscanner achieved sub-0.7 s Largest Contentful Paint and mere ~1.2 s full loads, indicating extremely fast content delivery.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Minimal baggage:&lt;/strong&gt; In contrast, slower sites tend to be weighed down by heavy content and excess scripts. Many well-known brands ranked outside the top tier due to “heavy content and front-end complexity.” These bulky pages took over 5–9 seconds to load, or in the case of Marriot below, over 10 seconds.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fin16xgyegk1zj6sba3d3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fin16xgyegk1zj6sba3d3.png" alt="A screenshot of a graphAI-generated content may be incorrect., Picture" width="800" height="324"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Hotel Website Performance Benchmark Report&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;The takeaway: fastest sites keep their pages lean – optimizing images, compressing files, and capping third-party elements to avoid sluggish load times.&lt;/p&gt;

&lt;p&gt;By keeping page weight low and complexity in check, top performers ensure users aren’t left staring at blank screens. Fast complete load times not only improve user satisfaction but also reduce bounce rates – a win-win for experience and business metrics.  &lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#4. Stellar front-end optimization &amp;amp; core web vitals&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Beyond just loading quickly, the fastest websites also feel fast and stable to users. They achieve this through meticulous front-end optimization, evidenced by outstanding Core Web Vitals metrics like Largest Contentful Paint (LCP) and Cumulative Layout Shift (CLS):&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fast content rendering (low LCP):&lt;/strong&gt; Top sites prioritize delivering meaningful content immediately. In Catchpoint’s travel industry test, all of the 10 fastest sites had LCP well within recommended standards – many under ~1 second. For example, Skyscanner’s LCP came in around 613 ms, with even the 10th-best still near 1.1 s – comfortably beating the ~2.5 s guideline. This means users see the page’s main content almost instantly on these sites.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd2l4tfm1b5su2m9fzpev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd2l4tfm1b5su2m9fzpev.png" alt="A graph of a documentAI-generated content may be incorrect., Picture" width="800" height="396"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Travel Website Performance Benchmark Report&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Slower sites, on the other hand, often struggle with LCP above the 2.5 s mark, making them feel sluggish even if the total load isn’t horrific.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Stable layouts (no janky moves):&lt;/strong&gt; A striking data point – the top 10 travel websites all had a perfect CLS score of 0.00. Likewise, the best banks and retail sites exhibited virtually no unexpected layout shifts during load. In practice, this means no annoying page jumps (caused by late-loading ads or images) for the user.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffpfgs3onehmqxfz1b94n.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffpfgs3onehmqxfz1b94n.png" alt="A graph of blue bars with white textAI-generated content may be incorrect., Picture" width="800" height="371"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Airline Website Performance Benchmark Report&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Fast sites achieve this by reserving space for images/media and loading content in a controlled way. By contrast, lesser performers saw CLS values up to 0.8–1.5, causing noticeable “jank” and hurting user experience.&lt;/p&gt;

&lt;p&gt;It’s clear that front-end experience is the real differentiator among websites. Sites that not only load quickly but also render smoothly (low LCP) and stay visually stable (low CLS) ended up at the top of Catchpoint’s rankings. These Core Web Vitals were highly correlated with overall performance scores, underscoring that fastest sites focus heavily on user-centric front-end metrics.  &lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#5. High availability without sacrificing speed&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Leading websites understand that availability and speed must go hand in hand. It’s not enough to be up 24/7 if your site is slow or error-prone when loaded. The fastest sites manage to deliver near-perfect uptime and quick performance simultaneously:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Near-perfect uptime:&lt;/strong&gt; The majority of top performers maintain 99.9% or better availability. In the 2025 Banking benchmark, 75% of sites had at least 99.9% uptime – and all of the top finishers were essentially always-online. In fact, the #1 sites in multiple industries achieved ≈100% uptime during the testing period. This level of reliability ensures users rarely encounter errors or downtime.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F22933jp3ffcskn7iuyzz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F22933jp3ffcskn7iuyzz.png" alt="A blue rectangular object with black textAI-generated content may be incorrect., Picture" width="800" height="314"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;2025 Banking Website Performance Benchmark Report&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Reliability and Speed:&lt;/strong&gt; Crucially, the best sites don’t treat uptime as a safety net for poor performance. Availability isn’t the safety net it used to be.Several lower-ranked companies across industries had 99.9% uptime yet still fell behind due to slow pages or unstable fronts. Top sites avoid this trap by pairing high availability with fast, stable loads. They have resilient infrastructures to stay online, and they proactively monitor performance so that being “up” always means delivering a quality experience. Conversely, a few sites had serious outages (dipping to ~90% uptime) which obviously knocked them out of contention.&lt;/p&gt;

&lt;p&gt;The bottom line: reliability is foundational, but true excellence comes from reliability + speed. The fastest websites prioritize both. Users expect a site to be available and to respond instantly – leading sites deliver on that expectation consistently.  &lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;#6. Culture of continuous optimization&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Speed isn’t a one-time project. The fastest websites treat performance as an ongoing discipline—one rooted in measurement, iteration, and readiness for peak demand.&lt;/p&gt;

&lt;p&gt;Instacart is a clear example. While their average TTFB during the &lt;a href="https://www.catchpoint.com/blog/why-super-bowl-2025-was-a-triumph-for-internet-resilience" rel="noopener noreferrer"&gt;Super Bowl benchmark&lt;/a&gt; period was a sluggish 1,051 ms, that changed on game day. On February 9th, when their Super Bowl ad aired, Instacart slashed TTFB significantly—demonstrating that with the right optimizations, even high-traffic moments can be performance wins.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4lhlnkd5kvwfasf97tof.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4lhlnkd5kvwfasf97tof.png" alt="A graph showing a line and a red circleAI-generated content may be incorrect., Picture" width="800" height="419"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Instacart’s TTFB improvement on the day of the Super Bowl&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;This kind of readiness doesn’t happen by accident. Across all industries, one of the strongest shared habits of high-performing sites is this culture of continuous optimization. When performance is treated as a living system—not just a launch checklist—organizations are better equipped to handle traffic spikes, user expectations, and the unexpected.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Final thoughts: speed is strategy&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;The patterns are clear. The fastest sites on the web invest in performance at every level: from back-end infrastructure to front-end polish, from global delivery to rigorous uptime, all under a philosophy of ongoing improvement. These sites prove that it’s possible to be both fast and reliable– delighting users with sub-3 second loads, silky-smooth pages, and dependable service. Meanwhile, competitors that neglect these areas struggle with slow load times, unstable pages, or regional outages, putting them at a serious disadvantage.&lt;/p&gt;

&lt;p&gt;The good news is that these best practices are well-understood. As our benchmark reports show, the top performers simply execute them better – and reap the rewards in user experience. By adopting the common habits of the fastest websites, including smart DNS/CDN usage, lightweight design, and Core Web Vitals focus, any website can dramatically improve its speed. After all, as the data reminds us, “it’s no longer enough to be available; you have to be fast everywhere”. The fastest sites have set the bar – and they invite others to catch up.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Explore the full benchmark series&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;See how your industry stacks up &lt;a href="https://www.catchpoint.com/resource-library?*=benchmark+report" rel="noopener noreferrer"&gt;View all benchmark reports&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

</description>
      <category>monitoring</category>
      <category>observabilty</category>
    </item>
    <item>
      <title>Observability isn’t about the tool. It’s about the truth</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 13 Oct 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/observability-isnt-about-the-tool-its-about-the-truth-c3f</link>
      <guid>https://forem.com/catchpoint/observability-isnt-about-the-tool-its-about-the-truth-c3f</guid>
      <description>&lt;p&gt;An enterprise client reports latency. Your dashboards say everything is fine. They blame you. You blame them. Nobody can prove it either way.  &lt;/p&gt;

&lt;p&gt;This is where most monitoring efforts hit a wall. Too often, the conversation gets stuck on dashboards and tools instead of the one thing that really matters: &lt;strong&gt;truth&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Observability isn’t about collecting metrics or building pretty dashboards. It’s about knowing the truth — the ability to quickly get to the root of a problem when your reputation and revenue are on the line.&lt;/p&gt;

&lt;p&gt;Not vanity metrics. Not checkbox features. Just fast, end-to-end, and undeniable truth.&lt;/p&gt;

&lt;h2&gt;
  
  
  What happens when two companies see the same issue differently?
&lt;/h2&gt;

&lt;p&gt;A leading financial services provider (let’s call them Company A) was suddenly under pressure. A key enterprise client—Company B—reported delays of 3 to 6 seconds when hitting APIs embedded in their customer-facing apps.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Company B: "Your APIs are slow. It’s impacting our customer experience."&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Company A (relying on &lt;strong&gt;Datadog&lt;/strong&gt; APM): "Everything looks fine on our side."&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;A stalemate. And a textbook case of observability failure.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why couldn’t Datadog find the issue?
&lt;/h2&gt;

&lt;p&gt;This isn’t a knock on Datadog. It’s an excellent Application Performance Monitoring (APM) tool—but it wasn’t built to see beyond your own infrastructure.  &lt;/p&gt;

&lt;p&gt;So even though Company A had robust APM and logging, they couldn’t see anything outside their own walls. They couldn’t install agents in Company B’s infrastructure, and they certainly couldn’t drop Real User Monitoring (RUM) scripts into someone else’s codebase.&lt;/p&gt;

&lt;p&gt;Here’s what each tool can (and can’t) do:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;APM (like Datadog)&lt;/strong&gt;: Great inside the app — once traffic arrives.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;RUM&lt;/strong&gt;: Excellent for frontend insights — but only if you own the app.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Logs&lt;/strong&gt;: Useful for what already happened — but not where packets got stuck in transit.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The common denominator with all three is that none of them can see what’s happening &lt;em&gt;between&lt;/em&gt; systems. Let’s get into why.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why do APIs create blind spots between companies?
&lt;/h2&gt;

&lt;p&gt;APIs are the interface between companies, the digital waiters of the software world. Just like you don’t walk into a restaurant kitchen to talk to the chef, companies don’t peek behind each other’s firewalls. They interact through APIs, exchanging structured requests and responses without ever seeing what’s really cooking on the other side.&lt;/p&gt;

&lt;p&gt;And that’s where blind spots creep in.  &lt;/p&gt;

&lt;p&gt;When two systems communicate through APIs, they lack visibility to each other’s inner workings. The moment a request leaves your infrastructure, it enters the black box of “someone else’s problem,” which include infrastructure, networks, and dependencies you don’t own and can’t instrument&lt;/p&gt;

&lt;p&gt;The root problem is that the Internet isn’t instrumentable. You can’t deploy agents or RUM scripts across the networks and infrastructure you don’t control. That’s why traditional observability tools stop at the edge. Beyond that lies the unknown.&lt;/p&gt;

&lt;p&gt;But delivering a great digital experience depends on multiple networks, protocols, agents, and sub-systems to work together in concert. These dependencies form what we call the &lt;a href="https://www.catchpoint.com/glossary/internet-stack" rel="noopener noreferrer"&gt;Internet Stack&lt;/a&gt;: DNS, CDN, BGP, ISP, last mile, backbone, and more.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy583eq0rwiaythemcag7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy583eq0rwiaythemcag7.png" width="800" height="399"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The Internet Stack&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;When performance breaks down somewhere in that chain, it doesn’t matter if it’s your fault or not—your customers still feel it. APIs, after all, were designed for efficiency, not visibility.&lt;/p&gt;

&lt;p&gt;This is where &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) becomes essential. IPM enables deep visibility into every layer of the Internet that can impact your service. Think of it as APM for the Internet Stack; purpose-built for the systems you don't own but still rely on.  &lt;/p&gt;

&lt;h2&gt;
  
  
  How do you get to the truth when APM falls short?  
&lt;/h2&gt;

&lt;p&gt;When traditional observability tools couldn’t explain the latency, IPM filled the gap. Instead of guessing, Company A used IPM to run &lt;strong&gt;synthetic API tests&lt;/strong&gt; across real-world networks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;From user ISPs: major U.S. carriers and fiber providers&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;From backbone and enterprise vantage points&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;From inside Company A’s own infrastructure&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Each test simulated actual API calls, complete with traceable request IDs and timestamps. And the results were undeniable.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbel82and5tcvvatejq73.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbel82and5tcvvatejq73.jpg" width="800" height="238"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This diagram maps the full path of an API call — from the client through Akamai, to internal proxy infrastructure and upstream systems. It clearly shows where latency accumulates:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;DNS, connect, and SSL times are negligible.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Akamai's edge processing is fast (~48ms).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Major delays occur during origin fetch (3,143ms) and proxy fetch (2,364ms)&lt;/strong&gt;—both inside the server infrastructure.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This confirms the problem isn’t with the client or CDN, but deep in the backend&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxumtbnkn5yuu21ipzciq.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxumtbnkn5yuu21ipzciq.jpg" width="800" height="369"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Latency breakdown across cities&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;This chart tracks average response and wait times across major U.S. cities. The key insight:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Latency patterns are &lt;strong&gt;remarkably consistent across geography&lt;/strong&gt;.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;A single spike appears across multiple regions, ruling out a location-specific issue.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This supports the insight that the bottleneck lives within the origin infrastructure, not in external networks.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdj3vzfxeqk7y9jy2aa43.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdj3vzfxeqk7y9jy2aa43.jpg" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;ISP breakdown&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Here, performance is analyzed by ISP (e.g., AT&amp;amp;T, Comcast, Verizon):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Despite some noise, the &lt;strong&gt;pattern is stable across providers&lt;/strong&gt;, with no single ISP showing consistently worse performance.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This helps eliminate ISP-side routing or congestion as a root cause.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The brief AT&amp;amp;T spike aligns with the same moment seen in city-level data.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;The result:&lt;/strong&gt; Consistent 3–6 second latency, internally and externally.&lt;/p&gt;

&lt;p&gt;With that intelligence, they could rule out the usual suspects:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;It wasn’t the ISP&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;It wasn’t the CDN&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;It wasn’t DNS&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;It wasn’t the proxy (Envoy)&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The process of elimination worked like a proper diagnostic: isolate each layer, eliminate what’s clean, and close in on the source. Parsing response headers like x-envoy-upstream-service-time confirmed the latency was occurring further upstream, deep within Company A’s own service environment. This pointed engineers in the right direction without them needing to sift through endless log lines. Trace IDs and timestamps were shared with internal teams to help pinpoint issues around application dependencies—eventually confirmed to be the root cause.&lt;/p&gt;

&lt;p&gt;This methodical approach, including initial discussion and setup, took just three hours and about 15 test runs. There was no guesswork. Just clarity.&lt;/p&gt;

&lt;p&gt;After internal validation, teams began work on the improvements, which are still ongoing but already measurable where it matters most.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsmbpi493pzv1151e5zx2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsmbpi493pzv1151e5zx2.png" width="800" height="346"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Backend latency has dropped significantly: both upstream service time and overall wait time have been cut nearly in half. These gains reflect steady optimization efforts that are clearly moving in the right direction.&lt;/p&gt;

&lt;h2&gt;
  
  
  What IPM delivers that APM can’t
&lt;/h2&gt;

&lt;p&gt;Let’s be clear: Datadog, New Relic, and Dynatrace are outstanding at what they do — inside your infrastructure. But they weren’t designed to monitor the Internet itself.&lt;/p&gt;

&lt;p&gt;Catchpoint IPM was. Here’s how:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;A vast Global Agent Network&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;3000+ agents&lt;/a&gt; across last-mile, backbone, cloud, enterprise, and on-prem environments&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Real-user network emulation, not cloud-only testbeds&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Full synthetic coverage&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  HTTP/S, APIs, Browser, DNS, SSL, BGP, MQTT, QUIC, Custom scripts&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Advanced diagnostics&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Packet loss, jitter, path tracing, hop analysis&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Region-specific degradation detection&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Frontend visibility&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;WebPageTest for in-depth frontend perf&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Browser + mobile RUM SDKs for teams who can instrument the frontend&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Seamless integration&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Feeds directly into Datadog, Splunk, New Relic, Dynatrace&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Enhances existing observability stacks without replacing them&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Provides end-to-end visibility across the Internet Stack via a &lt;a href="https://www.catchpoint.com/internet-stack-map" rel="noopener noreferrer"&gt;real-time dependency map&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Why teams cling to familiar tools even when they’re not fit for purpose
&lt;/h2&gt;

&lt;p&gt;Familiar tools are comfortable. They’re already deployed, widely understood, and politically safe. But too often, comfort wins out over capability—especially in large, mature organizations where tooling decisions are influenced by inertia, not fitness for purpose. But when seconds matter and customers are impacted, you need &lt;strong&gt;clarity&lt;/strong&gt;, not comfort.&lt;/p&gt;

&lt;h2&gt;
  
  
  Who takes the blame when APIs are slow?
&lt;/h2&gt;

&lt;p&gt;In this case, Company B blamed Company A. Company A blamed Company B. But neither had data to prove their case.&lt;/p&gt;

&lt;p&gt;Meanwhile, users just saw a slow experience.&lt;/p&gt;

&lt;p&gt;End users don’t know an API call is crossing company boundaries. They only see the brand they’re interacting with. If it’s slow, they assume that brand is to blame. That’s why solving performance issues quickly is about more than technical hygiene. It’s about protecting business relationships and customer trust.&lt;/p&gt;

&lt;h2&gt;
  
  
  Final thought: What’s the real job of observability?
&lt;/h2&gt;

&lt;p&gt;Observability isn’t about the coolest UI or the biggest vendor budget. It’s about getting to the truth, fast. And often, the truth lies &lt;strong&gt;outside&lt;/strong&gt; your four walls.&lt;/p&gt;

&lt;p&gt;In an AI-driven world, data powers decisions. But if your data is incomplete or your telemetry is limited to your own infrastructure, your AI is just guessing.&lt;/p&gt;

&lt;p&gt;Catchpoint IPM gives teams the ability to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Validate performance from the outside in&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Prove or disprove internal assumptions with independent data&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Pinpoint root causes in minutes, not days&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Because the point of observability isn’t the tool.&lt;/p&gt;

&lt;p&gt;It’s the truth.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Got a latency mystery your tools can’t solve?&lt;/strong&gt; &lt;a href="https://www.catchpoint.com/learn-more" rel="noopener noreferrer"&gt;Let’s talk.&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;An enterprise client reports latency. Your dashboards say everything is fine. They blame you. You blame them. Nobody can prove it either way.  &lt;/p&gt;

&lt;p&gt;This is where most monitoring efforts hit a wall. Too often, the conversation gets stuck on dashboards and tools instead of the one thing that really matters: &lt;strong&gt;truth&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Observability isn’t about collecting metrics or building pretty dashboards. It’s about knowing the truth — the ability to quickly get to the root of a problem when your reputation and revenue are on the line.&lt;/p&gt;

&lt;p&gt;Not vanity metrics. Not checkbox features. Just fast, end-to-end, and undeniable truth.&lt;/p&gt;

&lt;h2&gt;
  
  
  What happens when two companies see the same issue differently?
&lt;/h2&gt;

&lt;p&gt;A leading financial services provider (let’s call them Company A) was suddenly under pressure. A key enterprise client—Company B—reported delays of 3 to 6 seconds when hitting APIs embedded in their customer-facing apps.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Company B: "Your APIs are slow. It’s impacting our customer experience."&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Company A (relying on &lt;strong&gt;Datadog&lt;/strong&gt; APM): "Everything looks fine on our side."&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;A stalemate. And a textbook case of observability failure.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why couldn’t Datadog find the issue?
&lt;/h2&gt;

&lt;p&gt;This isn’t a knock on Datadog. It’s an excellent Application Performance Monitoring (APM) tool—but it wasn’t built to see beyond your own infrastructure.  &lt;/p&gt;

&lt;p&gt;So even though Company A had robust APM and logging, they couldn’t see anything outside their own walls. They couldn’t install agents in Company B’s infrastructure, and they certainly couldn’t drop Real User Monitoring (RUM) scripts into someone else’s codebase.&lt;/p&gt;

&lt;p&gt;Here’s what each tool can (and can’t) do:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;APM (like Datadog)&lt;/strong&gt;: Great inside the app — once traffic arrives.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;RUM&lt;/strong&gt;: Excellent for frontend insights — but only if you own the app.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Logs&lt;/strong&gt;: Useful for what already happened — but not where packets got stuck in transit.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The common denominator with all three is that none of them can see what’s happening &lt;em&gt;between&lt;/em&gt; systems. Let’s get into why.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why do APIs create blind spots between companies?
&lt;/h2&gt;

&lt;p&gt;APIs are the interface between companies, the digital waiters of the software world. Just like you don’t walk into a restaurant kitchen to talk to the chef, companies don’t peek behind each other’s firewalls. They interact through APIs, exchanging structured requests and responses without ever seeing what’s really cooking on the other side.&lt;/p&gt;

&lt;p&gt;And that’s where blind spots creep in.  &lt;/p&gt;

&lt;p&gt;When two systems communicate through APIs, they lack visibility to each other’s inner workings. The moment a request leaves your infrastructure, it enters the black box of “someone else’s problem,” which include infrastructure, networks, and dependencies you don’t own and can’t instrument&lt;/p&gt;

&lt;p&gt;The root problem is that the Internet isn’t instrumentable. You can’t deploy agents or RUM scripts across the networks and infrastructure you don’t control. That’s why traditional observability tools stop at the edge. Beyond that lies the unknown.&lt;/p&gt;

&lt;p&gt;But delivering a great digital experience depends on multiple networks, protocols, agents, and sub-systems to work together in concert. These dependencies form what we call the &lt;a href="https://www.catchpoint.com/glossary/internet-stack" rel="noopener noreferrer"&gt;Internet Stack&lt;/a&gt;: DNS, CDN, BGP, ISP, last mile, backbone, and more.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy583eq0rwiaythemcag7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy583eq0rwiaythemcag7.png" width="800" height="399"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The Internet Stack&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;When performance breaks down somewhere in that chain, it doesn’t matter if it’s your fault or not—your customers still feel it. APIs, after all, were designed for efficiency, not visibility.&lt;/p&gt;

&lt;p&gt;This is where &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) becomes essential. IPM enables deep visibility into every layer of the Internet that can impact your service. Think of it as APM for the Internet Stack; purpose-built for the systems you don't own but still rely on.  &lt;/p&gt;

&lt;h2&gt;
  
  
  How do you get to the truth when APM falls short?  
&lt;/h2&gt;

&lt;p&gt;When traditional observability tools couldn’t explain the latency, IPM filled the gap. Instead of guessing, Company A used IPM to run &lt;strong&gt;synthetic API tests&lt;/strong&gt; across real-world networks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;From user ISPs: major U.S. carriers and fiber providers&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;From backbone and enterprise vantage points&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;From inside Company A’s own infrastructure&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Each test simulated actual API calls, complete with traceable request IDs and timestamps. And the results were undeniable.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbel82and5tcvvatejq73.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbel82and5tcvvatejq73.jpg" width="800" height="238"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This diagram maps the full path of an API call — from the client through Akamai, to internal proxy infrastructure and upstream systems. It clearly shows where latency accumulates:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;DNS, connect, and SSL times are negligible.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Akamai's edge processing is fast (~48ms).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Major delays occur during origin fetch (3,143ms) and proxy fetch (2,364ms)&lt;/strong&gt;—both inside the server infrastructure.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This confirms the problem isn’t with the client or CDN, but deep in the backend&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxumtbnkn5yuu21ipzciq.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxumtbnkn5yuu21ipzciq.jpg" width="800" height="369"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Latency breakdown across cities&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;This chart tracks average response and wait times across major U.S. cities. The key insight:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Latency patterns are &lt;strong&gt;remarkably consistent across geography&lt;/strong&gt;.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;A single spike appears across multiple regions, ruling out a location-specific issue.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This supports the insight that the bottleneck lives within the origin infrastructure, not in external networks.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdj3vzfxeqk7y9jy2aa43.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdj3vzfxeqk7y9jy2aa43.jpg" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;ISP breakdown&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Here, performance is analyzed by ISP (e.g., AT&amp;amp;T, Comcast, Verizon):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Despite some noise, the &lt;strong&gt;pattern is stable across providers&lt;/strong&gt;, with no single ISP showing consistently worse performance.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This helps eliminate ISP-side routing or congestion as a root cause.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The brief AT&amp;amp;T spike aligns with the same moment seen in city-level data.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;The result:&lt;/strong&gt; Consistent 3–6 second latency, internally and externally.&lt;/p&gt;

&lt;p&gt;With that intelligence, they could rule out the usual suspects:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;It wasn’t the ISP&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;It wasn’t the CDN&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;It wasn’t DNS&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;It wasn’t the proxy (Envoy)&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The process of elimination worked like a proper diagnostic: isolate each layer, eliminate what’s clean, and close in on the source. Parsing response headers like x-envoy-upstream-service-time confirmed the latency was occurring further upstream, deep within Company A’s own service environment. This pointed engineers in the right direction without them needing to sift through endless log lines. Trace IDs and timestamps were shared with internal teams to help pinpoint issues around application dependencies—eventually confirmed to be the root cause.&lt;/p&gt;

&lt;p&gt;This methodical approach, including initial discussion and setup, took just three hours and about 15 test runs. There was no guesswork. Just clarity.&lt;/p&gt;

&lt;p&gt;After internal validation, teams began work on the improvements, which are still ongoing but already measurable where it matters most.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsmbpi493pzv1151e5zx2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsmbpi493pzv1151e5zx2.png" width="800" height="346"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Backend latency has dropped significantly: both upstream service time and overall wait time have been cut nearly in half. These gains reflect steady optimization efforts that are clearly moving in the right direction.&lt;/p&gt;

&lt;h2&gt;
  
  
  What IPM delivers that APM can’t
&lt;/h2&gt;

&lt;p&gt;Let’s be clear: Datadog, New Relic, and Dynatrace are outstanding at what they do — inside your infrastructure. But they weren’t designed to monitor the Internet itself.&lt;/p&gt;

&lt;p&gt;Catchpoint IPM was. Here’s how:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;A vast Global Agent Network&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;3000+ agents&lt;/a&gt; across last-mile, backbone, cloud, enterprise, and on-prem environments&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Real-user network emulation, not cloud-only testbeds&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Full synthetic coverage&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  HTTP/S, APIs, Browser, DNS, SSL, BGP, MQTT, QUIC, Custom scripts&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Advanced diagnostics&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Packet loss, jitter, path tracing, hop analysis&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Region-specific degradation detection&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Frontend visibility&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;WebPageTest for in-depth frontend perf&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Browser + mobile RUM SDKs for teams who can instrument the frontend&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Seamless integration&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Feeds directly into Datadog, Splunk, New Relic, Dynatrace&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Enhances existing observability stacks without replacing them&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Provides end-to-end visibility across the Internet Stack via a &lt;a href="https://www.catchpoint.com/internet-stack-map" rel="noopener noreferrer"&gt;real-time dependency map&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Why teams cling to familiar tools even when they’re not fit for purpose
&lt;/h2&gt;

&lt;p&gt;Familiar tools are comfortable. They’re already deployed, widely understood, and politically safe. But too often, comfort wins out over capability—especially in large, mature organizations where tooling decisions are influenced by inertia, not fitness for purpose. But when seconds matter and customers are impacted, you need &lt;strong&gt;clarity&lt;/strong&gt;, not comfort.&lt;/p&gt;

&lt;h2&gt;
  
  
  Who takes the blame when APIs are slow?
&lt;/h2&gt;

&lt;p&gt;In this case, Company B blamed Company A. Company A blamed Company B. But neither had data to prove their case.&lt;/p&gt;

&lt;p&gt;Meanwhile, users just saw a slow experience.&lt;/p&gt;

&lt;p&gt;End users don’t know an API call is crossing company boundaries. They only see the brand they’re interacting with. If it’s slow, they assume that brand is to blame. That’s why solving performance issues quickly is about more than technical hygiene. It’s about protecting business relationships and customer trust.&lt;/p&gt;

&lt;h2&gt;
  
  
  Final thought: What’s the real job of observability?
&lt;/h2&gt;

&lt;p&gt;Observability isn’t about the coolest UI or the biggest vendor budget. It’s about getting to the truth, fast. And often, the truth lies &lt;strong&gt;outside&lt;/strong&gt; your four walls.&lt;/p&gt;

&lt;p&gt;In an AI-driven world, data powers decisions. But if your data is incomplete or your telemetry is limited to your own infrastructure, your AI is just guessing.&lt;/p&gt;

&lt;p&gt;Catchpoint IPM gives teams the ability to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Validate performance from the outside in&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Prove or disprove internal assumptions with independent data&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Pinpoint root causes in minutes, not days&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Because the point of observability isn’t the tool.&lt;/p&gt;

&lt;p&gt;It’s the truth.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Got a latency mystery your tools can’t solve?&lt;/strong&gt; &lt;a href="https://www.catchpoint.com/learn-more" rel="noopener noreferrer"&gt;Let’s talk.&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

</description>
      <category>observability</category>
      <category>monitoring</category>
    </item>
    <item>
      <title>From the source to the edge: the six agent types you can’t ignore</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 06 Oct 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/from-the-source-to-the-edge-the-six-agent-types-you-cant-ignore-dd8</link>
      <guid>https://forem.com/catchpoint/from-the-source-to-the-edge-the-six-agent-types-you-cant-ignore-dd8</guid>
      <description>&lt;p&gt;Recently, Catchpoint expanded our Global Agent Network to over &lt;a href="https://www.catchpoint.com/blog/the-power-of-over-3000-intelligent-observability-agents" rel="noopener noreferrer"&gt;3,000 agents&lt;/a&gt;. In a crowded space, this is by far one of our key differentiators. At the time of writing, no one else boasts &lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;395 providers in 105 countries and 346 cities.&lt;/a&gt; As Director of ISP Strategy, I’m not here to pat myself on the back—my real question is: why? Why build such a massive, independent network — going through all the effort to place backbone agents in hard-to-reach regions like China, Russia, and several African countries?  &lt;/p&gt;

&lt;p&gt;The answer lies in how the Internet is built. It isn’t a single, monolithic network but a patchwork quilt of thousands of independent networks—ISPs, data-centers, backbones, wireless carriers, and more—all stitched together by peering and transit agreements worldwide. To monitor performance accurately, you need visibility into every layer of that quilt, or, what we like to call the &lt;a href="https://www.catchpoint.com/glossary/internet-stack" rel="noopener noreferrer"&gt;Internet Stack&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpmz6dtib086y2y7y1umu.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpmz6dtib086y2y7y1umu.png" alt="A blue and purple rectangular chart with iconsAI-generated content may be incorrect., Picture" width="800" height="375"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The Internet Stack&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;In this article, we’ll unpack the five types of Catchpoint synthetic agents—backbone, wireless, last-mile, cloud, and BGP—and show you exactly when and why each matters for keeping that quilt intact. First, let’s explore how the Internet actually connects end to end.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why multiple vantage points are key
&lt;/h2&gt;

&lt;p&gt;The “Internet” isn’t a single cloud—you can’t just tap into one place and see it all. It’s really tens of thousands of independent networks (ISPs, data centers, wireless carriers) stitched together by peering and transit deals. In a &lt;strong&gt;peering&lt;/strong&gt; arrangement, two networks exchange traffic for free. In a &lt;strong&gt;transit&lt;/strong&gt; relationship, one network pays another to carry its traffic. Peering keeps traffic local; transit carries it farther afield.&lt;/p&gt;

&lt;p&gt;Because each network makes its own choices about peering and transit, you need monitoring agents at many points to see what’s happening. A performance hiccup in one ISP’s peering location might not show up at a different ISP’s vantage point. That’s why Catchpoint places agents in dozens of key networks—so you won’t miss an issue that affects only a slice of the Internet.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxopl87dwp93nnveaw2ae.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxopl87dwp93nnveaw2ae.png" alt="A blue and pink colored networkDescription automatically generated with medium confidence, Picture" width="800" height="747"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;A map of how thousands of networks (Autonomous Systems) peer and buy transit around the world.&lt;/em&gt; &lt;a href="https://www.caida.org/projects/as-core/2020/#poster" rel="noopener noreferrer"&gt;&lt;em&gt;Source&lt;/em&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why tiers matter
&lt;/h2&gt;

&lt;p&gt;All those peering and transit agreements naturally sort networks into tiers:  &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Tier 1:&lt;/strong&gt; These are very large networks that peer with each other and don’t need to buy any transit at all to reach any corner of the Internet. They are considered the backbone of the Internet as they will typically carry long-distance Internet traffic. While the networks belonging to this group have changed since the beginning of the Internet, it’s remained relatively stable, including providers such as Lumen, AT&amp;amp;T, Cogent, Verizon, Orange, GTT, NTT, and Telxius. This category features very large traditional telecom providers that have been serving their domestic market for many years.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Tier 2:&lt;/strong&gt; Regional providers that both peer and buy transit. The scale of operations as well as the type of services they provide (IP transit, Ethernet or Dark Fiber wavelengths) will dictate the number of peering connections and transit providers. Most networks will fall into this category if they peer at one or more Internet exchange points and have two or more upstream providers.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Tier 3:&lt;/strong&gt; Smaller, local ISPs (often single-homed) that feed to an upstream provider. They show you what your end users see on a residential or localized network.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Putting agents in each tier matters because a Tier 1 network will have more visibility into global events (total or partial outages, backbone congestions, etc) as opposed to a regional Tier 2 or a local Tier 3 network. On the other hand, localized outages affecting a limited number of providers in a particular geographical area, won’t be easily observed unless having visibility from one of the affected networks.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why single-homed Tier 1/Tier 2 connectivity matters
&lt;/h2&gt;

&lt;p&gt;Now that we know how networks sort into tiers, let’s look at how those tiers influence the way we connect our agents.&lt;/p&gt;

&lt;p&gt;Many data centers, hosting and managed service providers typically use multiple upstream ISPs (Tier 1 and Tier 2) to create a single, aggregated connection to the Internet. This Internet connectivity, normally offered as a service, is often called blended bandwidth or multihoming.  &lt;/p&gt;

&lt;p&gt;Multihoming improves redundancy as traffic can be rerouted if one ISP goes down or has packet loss/congestion. It also improves performance as different ISPs may offer better latency to different geographies.&lt;/p&gt;

&lt;p&gt;But for &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) however, the use of multihomed agents instead of single-homed Tier1/Tier2 carriers introduces variability in your monitoring data, making it harder to identify and troubleshoot the issues affecting performance.  &lt;/p&gt;

&lt;p&gt;Here is why more than 96% of Catchpoint backbone agents use single-homed Tier 1/ Tier 2 connectivity instead of blended bandwidth:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Path consistency:&lt;/strong&gt; A consistent Tier 1 upstream path reduces variability, making anomalies and degradations easier to detect and attribute.  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Backbone visibility:&lt;/strong&gt; Tier1 visibility is essential to observe how the core Internet behaves in relation to routing anomalies, BGP hijacks or backbone congestion.  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Performance stability:&lt;/strong&gt; With blended connectivity, routes may change dynamically based on load-balancing or pricing strategies (e.g., BGP-based traffic engineering), affecting performance results.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Now that you understand how and why Catchpoint chooses single-homed Tier 1/Tier 2 connectivity, let’s look at each of our five synthetic agent types. We’ll explain where we place them, how we build them, and—most importantly—exactly what visibility each one gives you.&lt;/p&gt;

&lt;h2&gt;
  
  
  Backbone agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Backbone agents give you a “core-of-Internet” vantage point to catch global outages, BGP hijacks, and CDN-level issues no other agent can see.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;We place backbone agents in Tier 1 or Tier 2 ISPs worldwide, selecting carriers by:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Geography and market importance&lt;/strong&gt; (global connectivity hubs)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://asrank.caida.org/" rel="noopener noreferrer"&gt;&lt;strong&gt;CAIDA ASRank&lt;/strong&gt;&lt;/a&gt; &lt;strong&gt;&amp;amp;&lt;/strong&gt; &lt;a href="https://stats.labs.apnic.net/aspop" rel="noopener noreferrer"&gt;&lt;strong&gt;APNIC eyeball data&lt;/strong&gt;&lt;/a&gt; (to cover the most interconnected networks)&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Each backbone agent runs as a server cluster in a carrier-neutral data center with dedicated IP transit. Carrier neutrality ensures multiple international and domestic carriers via cross-connects, while colocating servers in one facility reduces colocation costs. In emerging markets (e.g., parts of Africa or China), where neutral data centers are scarce, we may host clusters in carrier-owned facilities—only as a last resort.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnad9i4ugzovg0ervecxy.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnad9i4ugzovg0ervecxy.png" width="800" height="651"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Connectivity diagram of backbone agents at a data center facility&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;Measuring performance and availability from backbone agents is critical for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Experience Level Objective (XLO) measurements&lt;/strong&gt;: Validate service performance when source and target share the same ISP.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;CDN performance/validation:&lt;/strong&gt; Ensure fast, reliable content delivery across the backbone.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Competitive benchmarking:&lt;/strong&gt; Compare your service to peers in the same Tier 1/2 networks.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Peering/ISP monitoring:&lt;/strong&gt; Detect routing changes, BGP anomalies, or unexpected transit behavior.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Geo-based DNS validation:&lt;/strong&gt; Confirm DNS resolution speed and correctness from the core network&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Public cloud agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Cloud agents give you visibility right inside public-cloud data centers—so you can catch platform-specific issues before they impact users.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;We run &lt;a href="https://www.catchpoint.com/global-observability-network#cloud" rel="noopener noreferrer"&gt;280+ cloud agents&lt;/a&gt; across every key availability region in AWS, Azure, Google, Oracle, Alibaba, Tencent, Akamai Compute, and OVH.  &lt;/p&gt;

&lt;p&gt;Measuring performance and availability to and from cloud agents is essential if you are hosting applications in the cloud or using any of their computing products. Cloud agents allow your SRE teams to pre-emptively detect performance degradations on public clouds that can affect how your users are experiencing your applications and services.&lt;/p&gt;

&lt;h2&gt;
  
  
  Wireless agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Wireless agents simulate real-world cellular conditions, giving you a true picture of how your applications perform on 3G/4G/5G networks.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;We place &lt;a href="https://www.catchpoint.com/global-observability-network#wireless" rel="noopener noreferrer"&gt;wireless agents&lt;/a&gt; using AWS Wavelength and independent carriers in the US, Canada, Japan, Germany, Korea, India, and the UK (e.g., Verizon, KDDI, BT, T-Mobile 5G, AT&amp;amp;T 5G).&lt;/p&gt;

&lt;p&gt;Running wireless tests alongside backbone tests lets you compare mobile experience to core-network performance—so you can spot issues like packet loss or DNS slowdowns that only affect cellular users.&lt;/p&gt;

&lt;h2&gt;
  
  
  Last-mile agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Last-mile agents live in real homes, giving you a true end-user view of broadband performance.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Our &lt;a href="https://www.catchpoint.com/global-observability-network#last-mile" rel="noopener noreferrer"&gt;last-mile agents&lt;/a&gt; run on small customer-premise devices that connect to a residential ISP. Use these agents to troubleshoot ISP-specific issues—like throttling, DNS failures, or regional outages—that only affect subscribers on a particular network.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Enterprise agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Enterprise agents give you visibility into your own network—from branch offices to data centers to edge locations.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Enterprise agents are deployed within your organization’s infrastructure. That includes office networks, private data centers, retail locations, or edge devices. These agents help you monitor internal applications, APIs, and services with the same level of granularity you get for external traffic. Combined with our Global Agent Network, enterprise agents complete the picture—giving you visibility from both outside-in and inside-out.  &lt;/p&gt;

&lt;h2&gt;
  
  
  BGP agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;BGP agents watch the real-time routing table, so you can catch hijacks, leaks, or unexpected path changes that threaten your service.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint maintains a route collector infrastructure that process real-time routing data from 1700+ BGP agents with the goal of monitoring BGP activity and detecting issues such as route hijacks and leaks.&lt;/p&gt;

&lt;p&gt;In addition to using RIPE RIS and RouteViews datasets, Catchpoint operates its own private collector infrastructure which includes agreements to receive data from 330+ BGP agents from 100 unique networks.&lt;/p&gt;

&lt;p&gt;If you share your own BGP sessions with our private collectors, you’ll gain even deeper insights in your portal—so you see exactly how routing anomalies affect your prefixes.&lt;/p&gt;

&lt;h2&gt;
  
  
  Wrapping it up
&lt;/h2&gt;

&lt;p&gt;When I asked, “Why build a network of over 3,000 agents in 105 countries and 346 cities?” The simple answer is that today’s Internet isn’t one giant cloud but a patchwork quilt of independent networks.  &lt;/p&gt;

&lt;p&gt;By spreading our agents across every layer of the Internet Stack, we can reveal problems at the very moment they start—whether it’s a routing change in a distant backbone, a subtle slowdown in a public cloud region, or a local ISP hiccup affecting a handful of homes.  &lt;/p&gt;

&lt;p&gt;This broad visibility isn’t about boasting coverage; it’s about ensuring that whenever something goes wrong, you know exactly where to look. In other words, the effort we put into building and maintaining such a diverse network isn’t just a technical feat. It’s the key to preventing those 3 a.m. wake-up calls for your IT team, avoiding frantically assembled war rooms, and keeping your users happy no matter where they connect.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Ready to unlock the full potentialof observability?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Learn more about Catchpoint'sintelligent agent network and how it can transform your monitoring strategy: &lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;https://www.catchpoint.com/global-observability-network&lt;/a&gt;  &lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;Recently, Catchpoint expanded our Global Agent Network to over &lt;a href="https://www.catchpoint.com/blog/the-power-of-over-3000-intelligent-observability-agents" rel="noopener noreferrer"&gt;3,000 agents&lt;/a&gt;. In a crowded space, this is by far one of our key differentiators. At the time of writing, no one else boasts &lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;395 providers in 105 countries and 346 cities.&lt;/a&gt; As Director of ISP Strategy, I’m not here to pat myself on the back—my real question is: why? Why build such a massive, independent network — going through all the effort to place backbone agents in hard-to-reach regions like China, Russia, and several African countries?  &lt;/p&gt;

&lt;p&gt;The answer lies in how the Internet is built. It isn’t a single, monolithic network but a patchwork quilt of thousands of independent networks—ISPs, data-centers, backbones, wireless carriers, and more—all stitched together by peering and transit agreements worldwide. To monitor performance accurately, you need visibility into every layer of that quilt, or, what we like to call the &lt;a href="https://www.catchpoint.com/glossary/internet-stack" rel="noopener noreferrer"&gt;Internet Stack&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpmz6dtib086y2y7y1umu.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpmz6dtib086y2y7y1umu.png" alt="A blue and purple rectangular chart with iconsAI-generated content may be incorrect., Picture" width="800" height="375"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The Internet Stack&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;In this article, we’ll unpack the five types of Catchpoint synthetic agents—backbone, wireless, last-mile, cloud, and BGP—and show you exactly when and why each matters for keeping that quilt intact. First, let’s explore how the Internet actually connects end to end.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why multiple vantage points are key
&lt;/h2&gt;

&lt;p&gt;The “Internet” isn’t a single cloud—you can’t just tap into one place and see it all. It’s really tens of thousands of independent networks (ISPs, data centers, wireless carriers) stitched together by peering and transit deals. In a &lt;strong&gt;peering&lt;/strong&gt; arrangement, two networks exchange traffic for free. In a &lt;strong&gt;transit&lt;/strong&gt; relationship, one network pays another to carry its traffic. Peering keeps traffic local; transit carries it farther afield.&lt;/p&gt;

&lt;p&gt;Because each network makes its own choices about peering and transit, you need monitoring agents at many points to see what’s happening. A performance hiccup in one ISP’s peering location might not show up at a different ISP’s vantage point. That’s why Catchpoint places agents in dozens of key networks—so you won’t miss an issue that affects only a slice of the Internet.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxopl87dwp93nnveaw2ae.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxopl87dwp93nnveaw2ae.png" alt="A blue and pink colored networkDescription automatically generated with medium confidence, Picture" width="800" height="747"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;A map of how thousands of networks (Autonomous Systems) peer and buy transit around the world.&lt;/em&gt; &lt;a href="https://www.caida.org/projects/as-core/2020/#poster" rel="noopener noreferrer"&gt;&lt;em&gt;Source&lt;/em&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why tiers matter
&lt;/h2&gt;

&lt;p&gt;All those peering and transit agreements naturally sort networks into tiers:  &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Tier 1:&lt;/strong&gt; These are very large networks that peer with each other and don’t need to buy any transit at all to reach any corner of the Internet. They are considered the backbone of the Internet as they will typically carry long-distance Internet traffic. While the networks belonging to this group have changed since the beginning of the Internet, it’s remained relatively stable, including providers such as Lumen, AT&amp;amp;T, Cogent, Verizon, Orange, GTT, NTT, and Telxius. This category features very large traditional telecom providers that have been serving their domestic market for many years.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Tier 2:&lt;/strong&gt; Regional providers that both peer and buy transit. The scale of operations as well as the type of services they provide (IP transit, Ethernet or Dark Fiber wavelengths) will dictate the number of peering connections and transit providers. Most networks will fall into this category if they peer at one or more Internet exchange points and have two or more upstream providers.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Tier 3:&lt;/strong&gt; Smaller, local ISPs (often single-homed) that feed to an upstream provider. They show you what your end users see on a residential or localized network.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Putting agents in each tier matters because a Tier 1 network will have more visibility into global events (total or partial outages, backbone congestions, etc) as opposed to a regional Tier 2 or a local Tier 3 network. On the other hand, localized outages affecting a limited number of providers in a particular geographical area, won’t be easily observed unless having visibility from one of the affected networks.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why single-homed Tier 1/Tier 2 connectivity matters
&lt;/h2&gt;

&lt;p&gt;Now that we know how networks sort into tiers, let’s look at how those tiers influence the way we connect our agents.&lt;/p&gt;

&lt;p&gt;Many data centers, hosting and managed service providers typically use multiple upstream ISPs (Tier 1 and Tier 2) to create a single, aggregated connection to the Internet. This Internet connectivity, normally offered as a service, is often called blended bandwidth or multihoming.  &lt;/p&gt;

&lt;p&gt;Multihoming improves redundancy as traffic can be rerouted if one ISP goes down or has packet loss/congestion. It also improves performance as different ISPs may offer better latency to different geographies.&lt;/p&gt;

&lt;p&gt;But for &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) however, the use of multihomed agents instead of single-homed Tier1/Tier2 carriers introduces variability in your monitoring data, making it harder to identify and troubleshoot the issues affecting performance.  &lt;/p&gt;

&lt;p&gt;Here is why more than 96% of Catchpoint backbone agents use single-homed Tier 1/ Tier 2 connectivity instead of blended bandwidth:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Path consistency:&lt;/strong&gt; A consistent Tier 1 upstream path reduces variability, making anomalies and degradations easier to detect and attribute.  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Backbone visibility:&lt;/strong&gt; Tier1 visibility is essential to observe how the core Internet behaves in relation to routing anomalies, BGP hijacks or backbone congestion.  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Performance stability:&lt;/strong&gt; With blended connectivity, routes may change dynamically based on load-balancing or pricing strategies (e.g., BGP-based traffic engineering), affecting performance results.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Now that you understand how and why Catchpoint chooses single-homed Tier 1/Tier 2 connectivity, let’s look at each of our five synthetic agent types. We’ll explain where we place them, how we build them, and—most importantly—exactly what visibility each one gives you.&lt;/p&gt;

&lt;h2&gt;
  
  
  Backbone agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Backbone agents give you a “core-of-Internet” vantage point to catch global outages, BGP hijacks, and CDN-level issues no other agent can see.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;We place backbone agents in Tier 1 or Tier 2 ISPs worldwide, selecting carriers by:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Geography and market importance&lt;/strong&gt; (global connectivity hubs)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://asrank.caida.org/" rel="noopener noreferrer"&gt;&lt;strong&gt;CAIDA ASRank&lt;/strong&gt;&lt;/a&gt; &lt;strong&gt;&amp;amp;&lt;/strong&gt; &lt;a href="https://stats.labs.apnic.net/aspop" rel="noopener noreferrer"&gt;&lt;strong&gt;APNIC eyeball data&lt;/strong&gt;&lt;/a&gt; (to cover the most interconnected networks)&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Each backbone agent runs as a server cluster in a carrier-neutral data center with dedicated IP transit. Carrier neutrality ensures multiple international and domestic carriers via cross-connects, while colocating servers in one facility reduces colocation costs. In emerging markets (e.g., parts of Africa or China), where neutral data centers are scarce, we may host clusters in carrier-owned facilities—only as a last resort.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnad9i4ugzovg0ervecxy.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnad9i4ugzovg0ervecxy.png" width="800" height="651"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Connectivity diagram of backbone agents at a data center facility&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;Measuring performance and availability from backbone agents is critical for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Experience Level Objective (XLO) measurements&lt;/strong&gt;: Validate service performance when source and target share the same ISP.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;CDN performance/validation:&lt;/strong&gt; Ensure fast, reliable content delivery across the backbone.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Competitive benchmarking:&lt;/strong&gt; Compare your service to peers in the same Tier 1/2 networks.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Peering/ISP monitoring:&lt;/strong&gt; Detect routing changes, BGP anomalies, or unexpected transit behavior.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Geo-based DNS validation:&lt;/strong&gt; Confirm DNS resolution speed and correctness from the core network&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Public cloud agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Cloud agents give you visibility right inside public-cloud data centers—so you can catch platform-specific issues before they impact users.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;We run &lt;a href="https://www.catchpoint.com/global-observability-network#cloud" rel="noopener noreferrer"&gt;280+ cloud agents&lt;/a&gt; across every key availability region in AWS, Azure, Google, Oracle, Alibaba, Tencent, Akamai Compute, and OVH.  &lt;/p&gt;

&lt;p&gt;Measuring performance and availability to and from cloud agents is essential if you are hosting applications in the cloud or using any of their computing products. Cloud agents allow your SRE teams to pre-emptively detect performance degradations on public clouds that can affect how your users are experiencing your applications and services.&lt;/p&gt;

&lt;h2&gt;
  
  
  Wireless agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Wireless agents simulate real-world cellular conditions, giving you a true picture of how your applications perform on 3G/4G/5G networks.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;We place &lt;a href="https://www.catchpoint.com/global-observability-network#wireless" rel="noopener noreferrer"&gt;wireless agents&lt;/a&gt; using AWS Wavelength and independent carriers in the US, Canada, Japan, Germany, Korea, India, and the UK (e.g., Verizon, KDDI, BT, T-Mobile 5G, AT&amp;amp;T 5G).&lt;/p&gt;

&lt;p&gt;Running wireless tests alongside backbone tests lets you compare mobile experience to core-network performance—so you can spot issues like packet loss or DNS slowdowns that only affect cellular users.&lt;/p&gt;

&lt;h2&gt;
  
  
  Last-mile agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Last-mile agents live in real homes, giving you a true end-user view of broadband performance.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Our &lt;a href="https://www.catchpoint.com/global-observability-network#last-mile" rel="noopener noreferrer"&gt;last-mile agents&lt;/a&gt; run on small customer-premise devices that connect to a residential ISP. Use these agents to troubleshoot ISP-specific issues—like throttling, DNS failures, or regional outages—that only affect subscribers on a particular network.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Enterprise agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;Enterprise agents give you visibility into your own network—from branch offices to data centers to edge locations.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Enterprise agents are deployed within your organization’s infrastructure. That includes office networks, private data centers, retail locations, or edge devices. These agents help you monitor internal applications, APIs, and services with the same level of granularity you get for external traffic. Combined with our Global Agent Network, enterprise agents complete the picture—giving you visibility from both outside-in and inside-out.  &lt;/p&gt;

&lt;h2&gt;
  
  
  BGP agents
&lt;/h2&gt;

&lt;p&gt;&lt;em&gt;BGP agents watch the real-time routing table, so you can catch hijacks, leaks, or unexpected path changes that threaten your service.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint maintains a route collector infrastructure that process real-time routing data from 1700+ BGP agents with the goal of monitoring BGP activity and detecting issues such as route hijacks and leaks.&lt;/p&gt;

&lt;p&gt;In addition to using RIPE RIS and RouteViews datasets, Catchpoint operates its own private collector infrastructure which includes agreements to receive data from 330+ BGP agents from 100 unique networks.&lt;/p&gt;

&lt;p&gt;If you share your own BGP sessions with our private collectors, you’ll gain even deeper insights in your portal—so you see exactly how routing anomalies affect your prefixes.&lt;/p&gt;

&lt;h2&gt;
  
  
  Wrapping it up
&lt;/h2&gt;

&lt;p&gt;When I asked, “Why build a network of over 3,000 agents in 105 countries and 346 cities?” The simple answer is that today’s Internet isn’t one giant cloud but a patchwork quilt of independent networks.  &lt;/p&gt;

&lt;p&gt;By spreading our agents across every layer of the Internet Stack, we can reveal problems at the very moment they start—whether it’s a routing change in a distant backbone, a subtle slowdown in a public cloud region, or a local ISP hiccup affecting a handful of homes.  &lt;/p&gt;

&lt;p&gt;This broad visibility isn’t about boasting coverage; it’s about ensuring that whenever something goes wrong, you know exactly where to look. In other words, the effort we put into building and maintaining such a diverse network isn’t just a technical feat. It’s the key to preventing those 3 a.m. wake-up calls for your IT team, avoiding frantically assembled war rooms, and keeping your users happy no matter where they connect.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Ready to unlock the full potentialof observability?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Learn more about Catchpoint'sintelligent agent network and how it can transform your monitoring strategy: &lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;https://www.catchpoint.com/global-observability-network&lt;/a&gt;  &lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

</description>
      <category>monitoring</category>
      <category>observability</category>
      <category>ipm</category>
      <category>rum</category>
    </item>
    <item>
      <title>Real-time detection of BGP blackholing and prefix hijacks</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 22 Sep 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/real-time-detection-of-bgp-blackholing-and-prefix-hijacks-5cea</link>
      <guid>https://forem.com/catchpoint/real-time-detection-of-bgp-blackholing-and-prefix-hijacks-5cea</guid>
      <description>&lt;p&gt;&lt;strong&gt;Border Gateway Protocol (BGP)&lt;/strong&gt; remains the backbone of inter-domain routing on the Internet, but its fundamental trust model leaves it vulnerable to misconfigurations, &lt;a href="https://www.catchpoint.com/blog/bgp-hijacking" rel="noopener noreferrer"&gt;hijacks&lt;/a&gt;, and blackholing. When these issues occur, they often go undetected by the impacted networks—until users report degraded performance or service outages.&lt;/p&gt;

&lt;p&gt;This post walks through a real-world incident in which a legitimate traffic spike led to an upstream provider mistakenly blackholing a critical IP address. The scenario illustrates how BGP blackholing can silently disrupt service and how external observability enables rapid diagnosis and resolution.&lt;/p&gt;

&lt;h2&gt;
  
  
  Understanding BGP blackholing
&lt;/h2&gt;

&lt;p&gt;BGP blackholing is a commonly used &lt;a href="https://www.catchpoint.com/blog/dns-ddos-mitigation" rel="noopener noreferrer"&gt;DDoS mitigation tactic&lt;/a&gt;. A network under attack announces a more specific route for the targeted IP or subnet, directing that traffic to a null interface to prevent it from reaching the intended service infrastructure. While effective in protecting resources during volumetric attacks, this approach can inadvertently block legitimate traffic when applied too aggressively.&lt;/p&gt;

&lt;p&gt;Let us try to understand this with help of an example:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe2djdsda68tdz5hc2wjw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe2djdsda68tdz5hc2wjw.png" alt="Picture 924464936, Picture" width="800" height="322"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;AS2 creates BGP blackhole. No traffic reaching intended server.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;In this case, the &lt;code&gt;/24&lt;/code&gt; prefix &lt;code&gt;1.0.0.0/24&lt;/code&gt; was owned and announced by one autonomous system (AS1). A specific point-of-presence (PoP) within this prefix was responsible for live-streaming a global event. The virtual IP for this PoP—&lt;code&gt;1.0.0.100&lt;/code&gt;—saw a surge in traffic from viewers worldwide.&lt;/p&gt;

&lt;p&gt;The traffic passed through an upstream provider (AS2), which monitored for DDoS patterns. Seeing the sudden spike, AS2’s automated mitigation system assumed the traffic was malicious. It responded by injecting a more specific &lt;code&gt;/32&lt;/code&gt; route for &lt;code&gt;1.0.0.100&lt;/code&gt; into the global routing table and directed it to a null interface.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The effect was immediate: traffic destined for the live-streaming service was dropped silently by AS2, resulting in widespread loss of availability for users across multiple regions.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Challenges in diagnosing upstream blackholing
&lt;/h2&gt;

&lt;p&gt;From AS1’s perspective, the service infrastructure remained operational, and no anomalies were observed in internal telemetry. However, users were unable to access the stream.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fubimzjc8n0ovdx90nuxh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fubimzjc8n0ovdx90nuxh.png" alt="A diagram of a computerAI-generated content may be incorrect., Picture" width="800" height="566"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Why traditional monitoring misses upstream blackholing&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Because the traffic was dropped before reaching AS1’s infrastructure, no logs or packet traces indicated a problem.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This is a common limitation when relying solely on internal monitoring. In upstream blackholing scenarios, routing changes happen outside of the origin network’s control, and the only observable symptom may be an unexplained drop in traffic or availability.&lt;/p&gt;

&lt;p&gt;The diagnostic challenge is further complicated by the specificity of the blackhole route. While the legitimate route for &lt;code&gt;1.0.0.0/24&lt;/code&gt; remained active, the injected &lt;code&gt;/32&lt;/code&gt; for &lt;code&gt;1.0.0.100&lt;/code&gt; took precedence due to BGP’s longest-prefix match rule, causing traffic to be rerouted and dropped at AS2.&lt;/p&gt;

&lt;h2&gt;
  
  
  Detecting origin AS mismatches and route hijacks
&lt;/h2&gt;

&lt;p&gt;The incident was identified through external route monitoring that detected an origin AS mismatch—the &lt;code&gt;/32&lt;/code&gt; prefix was being originated by AS2 instead of the expected AS1. This deviation triggered an alert, which prompted further analysis of the BGP path and propagation behavior.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F07i1b7elx84pbzpaixsc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F07i1b7elx84pbzpaixsc.png" alt="Picture 546420361, Picture" width="800" height="256"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Catchpoint platform BGP alert for ASN origin mismatch&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;An inspection of the AS path confirmed that certain regions were receiving the incorrect &lt;code&gt;/32&lt;/code&gt; advertisement and routing traffic through AS2, which blackholed the packets. The blackhole route had global reachability in select geographies, explaining the outage pattern observed by users.&lt;/p&gt;

&lt;p&gt;Mapping the propagation of the erroneous route helped identify the scope of the impact and enabled coordination with AS2 to withdraw the blackhole announcement. Once removed, traffic to &lt;code&gt;1.0.0.100&lt;/code&gt; resumed normal routing, and the live-streaming service was restored.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft8dfz070jvzlhaf2clwm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft8dfz070jvzlhaf2clwm.png" alt="Picture 1533534755, Picture" width="800" height="296"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Catchpoint platform showing BGP path, clearly identifying where traffic was split due to blackhole&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Broader implications
&lt;/h2&gt;

&lt;p&gt;This incident highlights the fragility of the global routing layer and the potential for automated systems to cause collateral damage, even when operating as designed. &lt;strong&gt;It also underscores the limitations of relying solely on internal data to understand end-to-end Internet performance.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0u043dgw7tf5splbewwc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0u043dgw7tf5splbewwc.png" alt="A map of the world with different colors of the world mapAI-generated content may be incorrect., Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Visibility into prefix propagation across the globe&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;External BGP monitoring allows operators to observe how their prefixes are being routed across the Internet and to detect anomalies such as:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Prefix hijacks by unintended or malicious ASes&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Upstream blackholing through more-specific announcements&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;AS path divergence and propagation anomalies&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Such visibility is critical for large-scale services that rely on third-party transit and upstream providers to reach global users.&lt;/p&gt;

&lt;h2&gt;
  
  
  Looking forward
&lt;/h2&gt;

&lt;p&gt;BGP remains a powerful but fragile protocol, and incidents like this illustrate the importance of proactive, third-party observability into Internet routing. As automated mitigation systems become more prevalent, it is increasingly important for network operators to verify not just whether their services are available, but whether their prefixes are being routed as intended.&lt;/p&gt;

&lt;p&gt;For a detailed exploration of BGP monitoring techniques and best practices&lt;a href="https://www.catchpoint.com/bgp-monitoring" rel="noopener noreferrer"&gt;&lt;strong&gt;, check out our in-depth guide&lt;/strong&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Border Gateway Protocol (BGP)&lt;/strong&gt; remains the backbone of inter-domain routing on the Internet, but its fundamental trust model leaves it vulnerable to misconfigurations, &lt;a href="https://www.catchpoint.com/blog/bgp-hijacking" rel="noopener noreferrer"&gt;hijacks&lt;/a&gt;, and blackholing. When these issues occur, they often go undetected by the impacted networks—until users report degraded performance or service outages.&lt;/p&gt;

&lt;p&gt;This post walks through a real-world incident in which a legitimate traffic spike led to an upstream provider mistakenly blackholing a critical IP address. The scenario illustrates how BGP blackholing can silently disrupt service and how external observability enables rapid diagnosis and resolution.&lt;/p&gt;

&lt;h2&gt;
  
  
  Understanding BGP blackholing
&lt;/h2&gt;

&lt;p&gt;BGP blackholing is a commonly used &lt;a href="https://www.catchpoint.com/blog/dns-ddos-mitigation" rel="noopener noreferrer"&gt;DDoS mitigation tactic&lt;/a&gt;. A network under attack announces a more specific route for the targeted IP or subnet, directing that traffic to a null interface to prevent it from reaching the intended service infrastructure. While effective in protecting resources during volumetric attacks, this approach can inadvertently block legitimate traffic when applied too aggressively.&lt;/p&gt;

&lt;p&gt;Let us try to understand this with help of an example:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe2djdsda68tdz5hc2wjw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe2djdsda68tdz5hc2wjw.png" alt="Picture 924464936, Picture" width="800" height="322"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;AS2 creates BGP blackhole. No traffic reaching intended server.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;In this case, the &lt;code&gt;/24&lt;/code&gt; prefix &lt;code&gt;1.0.0.0/24&lt;/code&gt; was owned and announced by one autonomous system (AS1). A specific point-of-presence (PoP) within this prefix was responsible for live-streaming a global event. The virtual IP for this PoP—&lt;code&gt;1.0.0.100&lt;/code&gt;—saw a surge in traffic from viewers worldwide.&lt;/p&gt;

&lt;p&gt;The traffic passed through an upstream provider (AS2), which monitored for DDoS patterns. Seeing the sudden spike, AS2’s automated mitigation system assumed the traffic was malicious. It responded by injecting a more specific &lt;code&gt;/32&lt;/code&gt; route for &lt;code&gt;1.0.0.100&lt;/code&gt; into the global routing table and directed it to a null interface.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The effect was immediate: traffic destined for the live-streaming service was dropped silently by AS2, resulting in widespread loss of availability for users across multiple regions.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Challenges in diagnosing upstream blackholing
&lt;/h2&gt;

&lt;p&gt;From AS1’s perspective, the service infrastructure remained operational, and no anomalies were observed in internal telemetry. However, users were unable to access the stream.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fubimzjc8n0ovdx90nuxh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fubimzjc8n0ovdx90nuxh.png" alt="A diagram of a computerAI-generated content may be incorrect., Picture" width="800" height="566"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Why traditional monitoring misses upstream blackholing&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Because the traffic was dropped before reaching AS1’s infrastructure, no logs or packet traces indicated a problem.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This is a common limitation when relying solely on internal monitoring. In upstream blackholing scenarios, routing changes happen outside of the origin network’s control, and the only observable symptom may be an unexplained drop in traffic or availability.&lt;/p&gt;

&lt;p&gt;The diagnostic challenge is further complicated by the specificity of the blackhole route. While the legitimate route for &lt;code&gt;1.0.0.0/24&lt;/code&gt; remained active, the injected &lt;code&gt;/32&lt;/code&gt; for &lt;code&gt;1.0.0.100&lt;/code&gt; took precedence due to BGP’s longest-prefix match rule, causing traffic to be rerouted and dropped at AS2.&lt;/p&gt;

&lt;h2&gt;
  
  
  Detecting origin AS mismatches and route hijacks
&lt;/h2&gt;

&lt;p&gt;The incident was identified through external route monitoring that detected an origin AS mismatch—the &lt;code&gt;/32&lt;/code&gt; prefix was being originated by AS2 instead of the expected AS1. This deviation triggered an alert, which prompted further analysis of the BGP path and propagation behavior.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F07i1b7elx84pbzpaixsc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F07i1b7elx84pbzpaixsc.png" alt="Picture 546420361, Picture" width="800" height="256"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Catchpoint platform BGP alert for ASN origin mismatch&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;An inspection of the AS path confirmed that certain regions were receiving the incorrect &lt;code&gt;/32&lt;/code&gt; advertisement and routing traffic through AS2, which blackholed the packets. The blackhole route had global reachability in select geographies, explaining the outage pattern observed by users.&lt;/p&gt;

&lt;p&gt;Mapping the propagation of the erroneous route helped identify the scope of the impact and enabled coordination with AS2 to withdraw the blackhole announcement. Once removed, traffic to &lt;code&gt;1.0.0.100&lt;/code&gt; resumed normal routing, and the live-streaming service was restored.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft8dfz070jvzlhaf2clwm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft8dfz070jvzlhaf2clwm.png" alt="Picture 1533534755, Picture" width="800" height="296"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Catchpoint platform showing BGP path, clearly identifying where traffic was split due to blackhole&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Broader implications
&lt;/h2&gt;

&lt;p&gt;This incident highlights the fragility of the global routing layer and the potential for automated systems to cause collateral damage, even when operating as designed. &lt;strong&gt;It also underscores the limitations of relying solely on internal data to understand end-to-end Internet performance.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0u043dgw7tf5splbewwc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0u043dgw7tf5splbewwc.png" alt="A map of the world with different colors of the world mapAI-generated content may be incorrect., Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Visibility into prefix propagation across the globe&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;External BGP monitoring allows operators to observe how their prefixes are being routed across the Internet and to detect anomalies such as:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Prefix hijacks by unintended or malicious ASes&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Upstream blackholing through more-specific announcements&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;AS path divergence and propagation anomalies&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Such visibility is critical for large-scale services that rely on third-party transit and upstream providers to reach global users.&lt;/p&gt;

&lt;h2&gt;
  
  
  Looking forward
&lt;/h2&gt;

&lt;p&gt;BGP remains a powerful but fragile protocol, and incidents like this illustrate the importance of proactive, third-party observability into Internet routing. As automated mitigation systems become more prevalent, it is increasingly important for network operators to verify not just whether their services are available, but whether their prefixes are being routed as intended.&lt;/p&gt;

&lt;p&gt;For a detailed exploration of BGP monitoring techniques and best practices&lt;a href="https://www.catchpoint.com/bgp-monitoring" rel="noopener noreferrer"&gt;&lt;strong&gt;, check out our in-depth guide&lt;/strong&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

</description>
      <category>bgp</category>
      <category>networking</category>
      <category>monitoring</category>
    </item>
    <item>
      <title>Google’s Agent-to-Agent (A2A) Protocol is here—Now Let’s Make it Observable</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 15 Sep 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/googles-agent-to-agent-a2a-protocol-is-here-now-lets-make-it-observable-5gh7</link>
      <guid>https://forem.com/catchpoint/googles-agent-to-agent-a2a-protocol-is-here-now-lets-make-it-observable-5gh7</guid>
      <description>&lt;p&gt;Can your AI tools really work together, or are they still stuck in silos? With Google’s new &lt;a href="https://developers.googleblog.com/en/a2a-a-new-era-of-agent-interoperability/" rel="noopener noreferrer"&gt;Agent-to-Agent (A2A) protocol&lt;/a&gt;, the days of isolated AI agents are numbered. This emerging standard lets specialized agents communicate, delegate, and collaborate—unlocking a new era of modular, scalable AI systems. Here’s how A2A could transform your workflows, and why making it observable is just as important as making it possible.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Why agent-to-agent is a breakthrough for collaborative AI
&lt;/h2&gt;

&lt;p&gt;To understand why A2A is such a breakthrough, it helps to look at how AI agents have evolved. Until now, most agents have relied on the Model Context Protocol (MCP), a mechanism that lets them enrich their responses by calling out to external tools, APIs, or functions in real time.&lt;/p&gt;

&lt;p&gt;MCP has been a game-changer, connecting agents to everything from knowledge bases and analytics dashboards to external services like GitHub and Jira, giving them far more context than what’s stored in their training data..&lt;/p&gt;

&lt;p&gt;However, MCP is still fundamentally a &lt;strong&gt;single-agent architecture&lt;/strong&gt;: the agent enhances itself by calling tools.&lt;/p&gt;

&lt;p&gt;Google’s &lt;strong&gt;A2A&lt;/strong&gt; protocol takes things a step further. It introduces a standard for how multiple AI agents can discover, understand, and collaborate with one another—delegating parts of a query to the agent most capable of resolving it.&lt;/p&gt;

&lt;p&gt;In a world where agents are being trained for niche domains (e.g., finance, healthcare, customer support, or DevOps), this multi-agent collaboration model could redefine how we build intelligent applications—modular, scalable, and highly specialized.&lt;/p&gt;

&lt;h2&gt;
  
  
  The industry has already gone multi—AI is next
&lt;/h2&gt;

&lt;p&gt;To appreciate why A2A is such a meaningful step, it helps to zoom out and see the broader trend across modern infrastructure:  &lt;/p&gt;

&lt;p&gt;Across DNS, CDN, cloud, and even AI, we've seen a shift from relying on a single provider to orchestrating &lt;strong&gt;multi-vendor ecosystems&lt;/strong&gt; that optimize for performance, cost, reliability, and use-case fit.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;DNS&lt;/strong&gt;: Where once a single DNS provider was the norm, many enterprises now use &lt;strong&gt;multi-DNS&lt;/strong&gt; strategies for faster resolution, better geographic coverage, and built-in failover.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;CDN&lt;/strong&gt;: The move from one CDN to &lt;strong&gt;multi-CDN architectures&lt;/strong&gt; enables companies to route traffic based on latency, region, or cost—while improving redundancy and performance at the edge.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Cloud&lt;/strong&gt;: With AWS, Azure, GCP, and others offering differentiated services, &lt;strong&gt;multi-cloud&lt;/strong&gt; is now a strategic choice. Teams pick the best-in-class services across vendors and reduce dependency on any single provider.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This "multi" strategy is not just about risk management—it's about specialization and optimization.&lt;/p&gt;

&lt;p&gt;Now, in the AI domain, we're witnessing the same pattern. While early adopters picked a single foundation model (e.g., GPT-4, Gemini, Claude), the next generation of intelligent systems will likely be &lt;strong&gt;multi-agent systems&lt;/strong&gt;. One agent might be optimized for data interpretation, another for decision-making, and another for domain-specific compliance.&lt;/p&gt;

&lt;h2&gt;
  
  
  Inside A2A: How agents discover and delegate in real time
&lt;/h2&gt;

&lt;p&gt;Google’s A2A protocol enables a framework where agents can collaborate dynamically. Think of this scenario:&lt;/p&gt;

&lt;p&gt;A user asks: &lt;em&gt;"What’s the weather in New York?"&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Agent 1 receives the query but lacks access to real-time weather data. However, it knows (via the A2A protocol) that Agent 2 is specialized in live weather updates. It queries Agent 2, gets the accurate data, and serves it back to the user—seamlessly.&lt;/p&gt;

&lt;p&gt;This interaction is powered by a few key concepts:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Host agent (client agent):&lt;/strong&gt; The initiating agent that receives the user query and delegates it if needed.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Remote agent:&lt;/strong&gt; An agent capable of fulfilling specialized tasks when invoked by another.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Agent card:&lt;/strong&gt; A JSON-based metadata descriptor published by agents to advertise their capabilities and endpoints—helping other agents discover and route tasks intelligently.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9su66nutbicb3no01vsq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9su66nutbicb3no01vsq.png" alt="A diagram of a customer serviceAI-generated content may be incorrect., Picture" width="800" height="611"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A2A facilitates communication between a "client" agent and a “remote” agent.&lt;/p&gt;

&lt;p&gt;I tried implementing a basic A2A interaction locally using the &lt;a href="https://github.com/google/A2A" rel="noopener noreferrer"&gt;open-source specification from Google&lt;/a&gt;. It’s remarkably modular and extensible—just like APIs revolutionized service-to-service communication, A2A may do the same for agent-to-agent orchestration.&lt;/p&gt;

&lt;p&gt;Here’s a snapshot from my local implementation:  &lt;/p&gt;

&lt;p&gt;The &lt;strong&gt;remote agent&lt;/strong&gt; listens on port 8001, ready to receive tasks. It advertises its capabilities via an Agent Card and executes incoming requests accordingly.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2d8tkn2a8ety1p7uk8it.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2d8tkn2a8ety1p7uk8it.png" alt="Picture 1, Picture" width="800" height="611"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The &lt;strong&gt;host agent&lt;/strong&gt; first discovers the remote agent, retrieves its capabilities, and sends a query prompt to the appropriate endpoint defined in the Agent Card. It then receives and returns the final response.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz59latknf1nw6pja45xb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz59latknf1nw6pja45xb.png" alt="Picture 1, Picture" width="800" height="440"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Achieving end-to-end visibility in multi-agent systems
&lt;/h2&gt;

&lt;p&gt;Multi-agent AI systems bring powerful new capabilities—but also new risks. In traditional architectures, observability stops at the edge of your stack. But in an A2A world, a single user request might pass through a chain of agents—each running on different systems, owned by different teams, and dependent on different APIs.&lt;/p&gt;

&lt;p&gt;Every agent interaction is essentially a service call. That means:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Added latency&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;More failure points&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Greater complexity when something goes wrong&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Take a chatbot for a ticket booking app. It may rely on internal microservices for availability and payments, but call out to a weather agent or flight-status agent using A2A. If one of those agents is slow or unresponsive, the whole experience degrades. And it’s hard to fix what you can’t see.&lt;/p&gt;

&lt;p&gt;This is where visibility matters. By mapping your service and agent dependencies—internal and external—you can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Pinpoint where slowdowns or errors occur&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Understand how agents interact across the chain&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Quickly isolate root causes when something fails&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Tools like &lt;a href="https://www.catchpoint.com/internet-stack-map" rel="noopener noreferrer"&gt;Catchpoint’s Internet Stack Map&lt;/a&gt; help teams visualize these flows. It leverages &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) to illustrate how requests flow through internal components and out to external agent APIs, making it clear where dependencies exist and where issues could arise.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F394jbgwh2ff2sw2fwvc5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F394jbgwh2ff2sw2fwvc5.png" alt="Picture 1, Picture" width="800" height="525"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint’s Internet Stack Map&lt;/p&gt;

&lt;p&gt;Just as we evolved from single-CDN to multi-CDN, or from monolithic apps to microservices, we are now entering an age of &lt;strong&gt;multi-agent intelligence&lt;/strong&gt;. And just like we learned to monitor those distributed system, we’ll now need to monitor multi-agent systems with the same rigor.&lt;/p&gt;

&lt;p&gt;Because the future isn’t just AI—it’s AI working together. Modular. distributed, collaborative. And IPM is what makes that visibility possible.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn more&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  See how Internet Stack Map can help you stay ahead of disruptions—&lt;a href="https://www.catchpoint.com/learn-more" rel="noopener noreferrer"&gt;schedule a demo today&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;Can your AI tools really work together, or are they still stuck in silos? With Google’s new &lt;a href="https://developers.googleblog.com/en/a2a-a-new-era-of-agent-interoperability/" rel="noopener noreferrer"&gt;Agent-to-Agent (A2A) protocol&lt;/a&gt;, the days of isolated AI agents are numbered. This emerging standard lets specialized agents communicate, delegate, and collaborate—unlocking a new era of modular, scalable AI systems. Here’s how A2A could transform your workflows, and why making it observable is just as important as making it possible.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Why agent-to-agent is a breakthrough for collaborative AI
&lt;/h2&gt;

&lt;p&gt;To understand why A2A is such a breakthrough, it helps to look at how AI agents have evolved. Until now, most agents have relied on the Model Context Protocol (MCP), a mechanism that lets them enrich their responses by calling out to external tools, APIs, or functions in real time.&lt;/p&gt;

&lt;p&gt;MCP has been a game-changer, connecting agents to everything from knowledge bases and analytics dashboards to external services like GitHub and Jira, giving them far more context than what’s stored in their training data..&lt;/p&gt;

&lt;p&gt;However, MCP is still fundamentally a &lt;strong&gt;single-agent architecture&lt;/strong&gt;: the agent enhances itself by calling tools.&lt;/p&gt;

&lt;p&gt;Google’s &lt;strong&gt;A2A&lt;/strong&gt; protocol takes things a step further. It introduces a standard for how multiple AI agents can discover, understand, and collaborate with one another—delegating parts of a query to the agent most capable of resolving it.&lt;/p&gt;

&lt;p&gt;In a world where agents are being trained for niche domains (e.g., finance, healthcare, customer support, or DevOps), this multi-agent collaboration model could redefine how we build intelligent applications—modular, scalable, and highly specialized.&lt;/p&gt;

&lt;h2&gt;
  
  
  The industry has already gone multi—AI is next
&lt;/h2&gt;

&lt;p&gt;To appreciate why A2A is such a meaningful step, it helps to zoom out and see the broader trend across modern infrastructure:  &lt;/p&gt;

&lt;p&gt;Across DNS, CDN, cloud, and even AI, we've seen a shift from relying on a single provider to orchestrating &lt;strong&gt;multi-vendor ecosystems&lt;/strong&gt; that optimize for performance, cost, reliability, and use-case fit.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;DNS&lt;/strong&gt;: Where once a single DNS provider was the norm, many enterprises now use &lt;strong&gt;multi-DNS&lt;/strong&gt; strategies for faster resolution, better geographic coverage, and built-in failover.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;CDN&lt;/strong&gt;: The move from one CDN to &lt;strong&gt;multi-CDN architectures&lt;/strong&gt; enables companies to route traffic based on latency, region, or cost—while improving redundancy and performance at the edge.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Cloud&lt;/strong&gt;: With AWS, Azure, GCP, and others offering differentiated services, &lt;strong&gt;multi-cloud&lt;/strong&gt; is now a strategic choice. Teams pick the best-in-class services across vendors and reduce dependency on any single provider.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This "multi" strategy is not just about risk management—it's about specialization and optimization.&lt;/p&gt;

&lt;p&gt;Now, in the AI domain, we're witnessing the same pattern. While early adopters picked a single foundation model (e.g., GPT-4, Gemini, Claude), the next generation of intelligent systems will likely be &lt;strong&gt;multi-agent systems&lt;/strong&gt;. One agent might be optimized for data interpretation, another for decision-making, and another for domain-specific compliance.&lt;/p&gt;

&lt;h2&gt;
  
  
  Inside A2A: How agents discover and delegate in real time
&lt;/h2&gt;

&lt;p&gt;Google’s A2A protocol enables a framework where agents can collaborate dynamically. Think of this scenario:&lt;/p&gt;

&lt;p&gt;A user asks: &lt;em&gt;"What’s the weather in New York?"&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Agent 1 receives the query but lacks access to real-time weather data. However, it knows (via the A2A protocol) that Agent 2 is specialized in live weather updates. It queries Agent 2, gets the accurate data, and serves it back to the user—seamlessly.&lt;/p&gt;

&lt;p&gt;This interaction is powered by a few key concepts:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Host agent (client agent):&lt;/strong&gt; The initiating agent that receives the user query and delegates it if needed.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Remote agent:&lt;/strong&gt; An agent capable of fulfilling specialized tasks when invoked by another.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Agent card:&lt;/strong&gt; A JSON-based metadata descriptor published by agents to advertise their capabilities and endpoints—helping other agents discover and route tasks intelligently.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9su66nutbicb3no01vsq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9su66nutbicb3no01vsq.png" alt="A diagram of a customer serviceAI-generated content may be incorrect., Picture" width="800" height="611"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A2A facilitates communication between a "client" agent and a “remote” agent.&lt;/p&gt;

&lt;p&gt;I tried implementing a basic A2A interaction locally using the &lt;a href="https://github.com/google/A2A" rel="noopener noreferrer"&gt;open-source specification from Google&lt;/a&gt;. It’s remarkably modular and extensible—just like APIs revolutionized service-to-service communication, A2A may do the same for agent-to-agent orchestration.&lt;/p&gt;

&lt;p&gt;Here’s a snapshot from my local implementation:  &lt;/p&gt;

&lt;p&gt;The &lt;strong&gt;remote agent&lt;/strong&gt; listens on port 8001, ready to receive tasks. It advertises its capabilities via an Agent Card and executes incoming requests accordingly.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2d8tkn2a8ety1p7uk8it.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2d8tkn2a8ety1p7uk8it.png" alt="Picture 1, Picture" width="800" height="611"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The &lt;strong&gt;host agent&lt;/strong&gt; first discovers the remote agent, retrieves its capabilities, and sends a query prompt to the appropriate endpoint defined in the Agent Card. It then receives and returns the final response.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz59latknf1nw6pja45xb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz59latknf1nw6pja45xb.png" alt="Picture 1, Picture" width="800" height="440"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Achieving end-to-end visibility in multi-agent systems
&lt;/h2&gt;

&lt;p&gt;Multi-agent AI systems bring powerful new capabilities—but also new risks. In traditional architectures, observability stops at the edge of your stack. But in an A2A world, a single user request might pass through a chain of agents—each running on different systems, owned by different teams, and dependent on different APIs.&lt;/p&gt;

&lt;p&gt;Every agent interaction is essentially a service call. That means:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Added latency&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;More failure points&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Greater complexity when something goes wrong&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Take a chatbot for a ticket booking app. It may rely on internal microservices for availability and payments, but call out to a weather agent or flight-status agent using A2A. If one of those agents is slow or unresponsive, the whole experience degrades. And it’s hard to fix what you can’t see.&lt;/p&gt;

&lt;p&gt;This is where visibility matters. By mapping your service and agent dependencies—internal and external—you can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Pinpoint where slowdowns or errors occur&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Understand how agents interact across the chain&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Quickly isolate root causes when something fails&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Tools like &lt;a href="https://www.catchpoint.com/internet-stack-map" rel="noopener noreferrer"&gt;Catchpoint’s Internet Stack Map&lt;/a&gt; help teams visualize these flows. It leverages &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) to illustrate how requests flow through internal components and out to external agent APIs, making it clear where dependencies exist and where issues could arise.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F394jbgwh2ff2sw2fwvc5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F394jbgwh2ff2sw2fwvc5.png" alt="Picture 1, Picture" width="800" height="525"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint’s Internet Stack Map&lt;/p&gt;

&lt;p&gt;Just as we evolved from single-CDN to multi-CDN, or from monolithic apps to microservices, we are now entering an age of &lt;strong&gt;multi-agent intelligence&lt;/strong&gt;. And just like we learned to monitor those distributed system, we’ll now need to monitor multi-agent systems with the same rigor.&lt;/p&gt;

&lt;p&gt;Because the future isn’t just AI—it’s AI working together. Modular. distributed, collaborative. And IPM is what makes that visibility possible.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn more&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  See how Internet Stack Map can help you stay ahead of disruptions—&lt;a href="https://www.catchpoint.com/learn-more" rel="noopener noreferrer"&gt;schedule a demo today&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>aiops</category>
      <category>monitoring</category>
      <category>a2a</category>
    </item>
    <item>
      <title>Monitoring in the Age of Complexity: 5 Assumptions CIOs Need to Rethink</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 08 Sep 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/monitoring-in-the-age-of-complexity-5-assumptions-cios-need-to-rethink-1gh4</link>
      <guid>https://forem.com/catchpoint/monitoring-in-the-age-of-complexity-5-assumptions-cios-need-to-rethink-1gh4</guid>
      <description>&lt;p&gt;In 2025, the average enterprise juggles over 150 SaaS applications, hybrid cloud infrastructures, and a workforce that expects seamless digital experiences—yet most CIOs still rely on monitoring strategies built for the data center era. The result? A $1.5 trillion annual hit to global GDP from downtime and performance lags, according to recent industry estimates. The problem isn’t the tools—it’s the thinking behind them.&lt;/p&gt;

&lt;p&gt;Monitoring isn’t just about keeping the lights on anymore. It’s a strategic lever for resilience, customer trust, and competitive edge. But outdated assumptions about what ‘good monitoring’ looks like are holding organizations back. Here are five myths CIOs and VPs must confront to lead in an era where complexity is the only constant.&lt;/p&gt;

&lt;h2&gt;
  
  
  Myth #1: Monitoring Is just an IT operations problem
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fym79wjikb6vg0xkkjhah.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fym79wjikb6vg0xkkjhah.jpeg" alt="A computer screens with graphs and diagramsAI-generated content may be incorrect., Picture" width="800" height="448"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The reality:&lt;/strong&gt; Monitoring is a business-critical function that impacts revenue and customer experience.&lt;/p&gt;

&lt;p&gt;When &lt;a href="https://www.forrester.com/press-newsroom/forrester-2024-us-customer-experience-index/" rel="noopener noreferrer"&gt;73% of customers&lt;/a&gt; say they’ll abandon a brand after two bad digital experiences, monitoring becomes a C-suite priority. It’s not about server uptime—it’s about revenue protection and brand equity. The narrative needs to shift from reactive alerts to proactive business alignment. Monitoring should be seen as a strategic function that directly impacts customer satisfaction and business outcomes.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Align IT metrics with business outcomes:&lt;/strong&gt; Use metrics like customer churn, conversion rates, or revenue impact instead of focusing solely on technical KPIs.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Adopt observability practices:&lt;/strong&gt; Integrate observability tools that provide real-time insights into how IT performance affects business outcomes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Promote cross-functional collaboration:&lt;/strong&gt; Ensure IT teams work closely with business units to prioritize monitoring efforts that directly impact customer satisfaction.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://shapeitrecruitment.co.uk/gartners-top-strategic-tech-trends-for-2023/" rel="noopener noreferrer"&gt;Gartner&lt;/a&gt; predicts that by 2026, 70% of organizations successfully applying observability will achieve shorter latency for decision-making, enabling competitive advantage for IT and business processes. This highlights the importance of aligning monitoring with business outcomes, not just IT metrics.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Myth #2: More data equals better visibility&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The reality:&lt;/strong&gt; More data often creates noise; actionable insights come from focusing on the right data.&lt;/p&gt;

&lt;p&gt;Modern systems generate terabytes of telemetry daily, but effective monitoring isn’t about collecting everything—it’s about identifying patterns and correlations that matter. AI-powered tools can help prioritize signal over noise, enabling faster root cause analysis.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Focus on key metrics:&lt;/strong&gt; Identify the most critical KPIs for your business and monitor those closely.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Leverage AI for noise reduction:&lt;/strong&gt; Use AI-driven tools to filter irrelevant data and surface actionable insights.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Implement distributed tracing:&lt;/strong&gt; Understand how different services interact with &lt;a href="https://www.catchpoint.com/tracing" rel="noopener noreferrer"&gt;Tracing&lt;/a&gt; to pinpoint bottlenecks or failures more effectively.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Organizations that implement AI-powered monitoring tools should see a reduction in mean time to resolution (MTTR). This is because AI can help identify patterns and anomalies in large datasets, making it easier to pinpoint the root cause of issues.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Myth #3: Internal metrics tell the full story&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faawkujeh78p8heb0rsij.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faawkujeh78p8heb0rsij.jpeg" alt="Picture 4, Picture" width="800" height="532"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Reality:&lt;/strong&gt; Most performance issues originate outside your firewall—true visibility requires end-to-end observability.&lt;/p&gt;

&lt;p&gt;Your cloud provider’s 99.99% uptime SLA doesn’t account for the &lt;a href="https://www.catchpoint.com/blog/cloud-monitorings-blind-spot-the-user-perspective" rel="noopener noreferrer"&gt;last mile&lt;/a&gt;—where 80% of performance issues originate. True observability looks beyond the firewall to the user’s reality. It’s essential to monitor the end-to-end user experience, including external factors that could affect performance. This holistic view ensures that organizations can address issues that impact the user experience, not just internal metrics.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Expand monitoring scope:&lt;/strong&gt; Include metrics like page load times, API response times, and third-party service performance.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Integrate business metrics with XLOs:&lt;/strong&gt; Move beyond technical KPIs to monitor customer-centric metrics such as abandonment rates, user satisfaction scores, and conversion rates. These Experience Level Objectives (XLOs) bridge IT performance with business outcomes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Use Internet Performance Monitoring (IPM):&lt;/strong&gt; Simulate user interactions from different geographies with &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;IPM&lt;/a&gt; to proactively identify potential issues.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The shift towards experience-centric monitoring will enable organizations to make more informed decisions and prioritize investments based on their impact on the bottom line.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Myth #4: AI will fix monitoring automatically&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvvscrmk30qxjwzipftwq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvvscrmk30qxjwzipftwq.png" alt="A computer screen showing a computer and trash cansAI-generated content may be incorrect., Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The reality:&lt;/strong&gt; AI is only as good as the data it analyzes—clean, contextual data is essential for success.&lt;/p&gt;

&lt;p&gt;AI can enhance monitoring by identifying patterns and predicting failures, but poor data quality undermines its effectiveness. Feed it garbage, and you’ll get polished garbage out. &lt;a href="https://www.gartner.com/en/newsroom/press-releases/2025-02-26-lack-of-ai-ready-data-puts-ai-projects-at-risk" rel="noopener noreferrer"&gt;Gartner&lt;/a&gt; warns that by 2026, 60% of AI-driven IT projects will fail without proper data readiness.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Invest in data quality:&lt;/strong&gt; Establish governance frameworks to ensure clean and consistent data inputs for AI models.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Adopt shift-left Observability:&lt;/strong&gt; Integrate monitoring into development cycles to identify issues earlier in the lifecycle.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Tailor AI solutions by context:&lt;/strong&gt; Customize AI-driven monitoring strategies based on the criticality of each application or service.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Organizations must adopt a "shift-left" approach to monitoring, where monitoring is integrated into the development lifecycle from the beginning. This allows organizations to identify and address potential issues early on, reducing the risk of costly downtime and performance problems.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Myth #5: Downtime is the only metric that matters&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvo6or3vq3dhzi1yfzqhl.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvo6or3vq3dhzi1yfzqhl.jpeg" alt="Picture 2, Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The reality:&lt;/strong&gt; Slow is the new down—Performance degradation can erode trust long before outages occur.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.catchpoint.com/statistics" rel="noopener noreferrer"&gt;53% of mobile users drop off&lt;/a&gt; if a page takes more than 3 seconds to load. Performance degradation silently erodes trust before outages even hit. Monitoring must evolve from tracking availability alone to measuring user experience metrics like page load times and transaction speeds.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Monitor user experience metrics:&lt;/strong&gt; Track latency, load times, and transaction completion rates alongside traditional uptime metrics. These XLOs ensure monitoring aligns with user satisfaction and business outcomes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Use predictive analytics:&lt;/strong&gt; Leverage historical data trends to anticipate potential slowdowns before they impact users, enabling proactive intervention.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Implement proactive remediation plans:&lt;/strong&gt; Automate responses for common performance issues, such as traffic spikes or resource bottlenecks, to minimize user impact and ensure seamless experiences.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Leveraging AI-powered predictive analytics to improve IT operations will enable organizations to move from a reactive to a proactive approach to monitoring, reducing downtime and improving overall system reliability&lt;/p&gt;

&lt;h2&gt;
  
  
  Rethinking monitoring for the age of complexity
&lt;/h2&gt;

&lt;p&gt;As enterprises face increasing complexity, monitoring has evolved from a back-office function to a strategic enabler of resilience, customer trust, and competitive differentiation. CIOs who cling to outdated assumptions risk falling behind—not just competitors, but their own customers’ expectations. The myths addressed in this article highlight the need for a paradigm shift in how organizations approach monitoring.&lt;/p&gt;

&lt;p&gt;Modern monitoring isn’t just about uptime or data collection; it’s about aligning IT performance with business outcomes, prioritizing user experience, and leveraging predictive analytics to stay ahead of issues. By embracing these principles, CIOs can transform monitoring into a competitive advantage.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Key takeaways for CIOs&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Monitoring is strategic:&lt;/strong&gt; Elevate monitoring from an IT operations function to a C-suite priority tied directly to revenue and customer satisfaction.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Focus on actionable insights:&lt;/strong&gt; Collect the right data—not just more data—and use AI-driven tools to surface meaningful patterns.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Expand visibility:&lt;/strong&gt; Go beyond internal metrics to monitor end-to-end user experiences and external factors affecting performance.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Prioritize data quality:&lt;/strong&gt; Invest in clean, contextual data to unlock the full potential of AI-driven monitoring.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Measure user experience:&lt;/strong&gt; Adopt XLOs to track metrics that reflect customer satisfaction alongside technical KPIs.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Ask yourself: Are you monitoring what truly matters—or just what’s easy? The answer will define your organization’s ability to thrive in an era where seamless digital experiences are the foundation of success.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Dig deeper:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Check out &lt;a href="https://vmblog.com/archive/2023/02/22/goodbye-lan-the-internet-is-the-network.aspx" rel="noopener noreferrer"&gt;&lt;em&gt;Goodbye LAN – The Internet is the Network&lt;/em&gt;&lt;/a&gt; on VMblog to learn how the Internet became the new enterprise network.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.catchpoint.com/blog/mastering-ipm-monitor-what-matters-from-where-it-matters" rel="noopener noreferrer"&gt;&lt;em&gt;Mastering IPM: Monitor what matters from where it matters&lt;/em&gt;&lt;/a&gt; – learn how Internet Performance Monitoring helps you stay resilient by focusing on the right metrics, from the right vantage points.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;In 2025, the average enterprise juggles over 150 SaaS applications, hybrid cloud infrastructures, and a workforce that expects seamless digital experiences—yet most CIOs still rely on monitoring strategies built for the data center era. The result? A $1.5 trillion annual hit to global GDP from downtime and performance lags, according to recent industry estimates. The problem isn’t the tools—it’s the thinking behind them.&lt;/p&gt;

&lt;p&gt;Monitoring isn’t just about keeping the lights on anymore. It’s a strategic lever for resilience, customer trust, and competitive edge. But outdated assumptions about what ‘good monitoring’ looks like are holding organizations back. Here are five myths CIOs and VPs must confront to lead in an era where complexity is the only constant.&lt;/p&gt;

&lt;h2&gt;
  
  
  Myth #1: Monitoring Is just an IT operations problem
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fym79wjikb6vg0xkkjhah.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fym79wjikb6vg0xkkjhah.jpeg" alt="A computer screens with graphs and diagramsAI-generated content may be incorrect., Picture" width="800" height="448"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The reality:&lt;/strong&gt; Monitoring is a business-critical function that impacts revenue and customer experience.&lt;/p&gt;

&lt;p&gt;When &lt;a href="https://www.forrester.com/press-newsroom/forrester-2024-us-customer-experience-index/" rel="noopener noreferrer"&gt;73% of customers&lt;/a&gt; say they’ll abandon a brand after two bad digital experiences, monitoring becomes a C-suite priority. It’s not about server uptime—it’s about revenue protection and brand equity. The narrative needs to shift from reactive alerts to proactive business alignment. Monitoring should be seen as a strategic function that directly impacts customer satisfaction and business outcomes.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Align IT metrics with business outcomes:&lt;/strong&gt; Use metrics like customer churn, conversion rates, or revenue impact instead of focusing solely on technical KPIs.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Adopt observability practices:&lt;/strong&gt; Integrate observability tools that provide real-time insights into how IT performance affects business outcomes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Promote cross-functional collaboration:&lt;/strong&gt; Ensure IT teams work closely with business units to prioritize monitoring efforts that directly impact customer satisfaction.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://shapeitrecruitment.co.uk/gartners-top-strategic-tech-trends-for-2023/" rel="noopener noreferrer"&gt;Gartner&lt;/a&gt; predicts that by 2026, 70% of organizations successfully applying observability will achieve shorter latency for decision-making, enabling competitive advantage for IT and business processes. This highlights the importance of aligning monitoring with business outcomes, not just IT metrics.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Myth #2: More data equals better visibility&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The reality:&lt;/strong&gt; More data often creates noise; actionable insights come from focusing on the right data.&lt;/p&gt;

&lt;p&gt;Modern systems generate terabytes of telemetry daily, but effective monitoring isn’t about collecting everything—it’s about identifying patterns and correlations that matter. AI-powered tools can help prioritize signal over noise, enabling faster root cause analysis.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Focus on key metrics:&lt;/strong&gt; Identify the most critical KPIs for your business and monitor those closely.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Leverage AI for noise reduction:&lt;/strong&gt; Use AI-driven tools to filter irrelevant data and surface actionable insights.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Implement distributed tracing:&lt;/strong&gt; Understand how different services interact with &lt;a href="https://www.catchpoint.com/tracing" rel="noopener noreferrer"&gt;Tracing&lt;/a&gt; to pinpoint bottlenecks or failures more effectively.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Organizations that implement AI-powered monitoring tools should see a reduction in mean time to resolution (MTTR). This is because AI can help identify patterns and anomalies in large datasets, making it easier to pinpoint the root cause of issues.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Myth #3: Internal metrics tell the full story&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faawkujeh78p8heb0rsij.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faawkujeh78p8heb0rsij.jpeg" alt="Picture 4, Picture" width="800" height="532"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Reality:&lt;/strong&gt; Most performance issues originate outside your firewall—true visibility requires end-to-end observability.&lt;/p&gt;

&lt;p&gt;Your cloud provider’s 99.99% uptime SLA doesn’t account for the &lt;a href="https://www.catchpoint.com/blog/cloud-monitorings-blind-spot-the-user-perspective" rel="noopener noreferrer"&gt;last mile&lt;/a&gt;—where 80% of performance issues originate. True observability looks beyond the firewall to the user’s reality. It’s essential to monitor the end-to-end user experience, including external factors that could affect performance. This holistic view ensures that organizations can address issues that impact the user experience, not just internal metrics.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Expand monitoring scope:&lt;/strong&gt; Include metrics like page load times, API response times, and third-party service performance.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Integrate business metrics with XLOs:&lt;/strong&gt; Move beyond technical KPIs to monitor customer-centric metrics such as abandonment rates, user satisfaction scores, and conversion rates. These Experience Level Objectives (XLOs) bridge IT performance with business outcomes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Use Internet Performance Monitoring (IPM):&lt;/strong&gt; Simulate user interactions from different geographies with &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;IPM&lt;/a&gt; to proactively identify potential issues.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The shift towards experience-centric monitoring will enable organizations to make more informed decisions and prioritize investments based on their impact on the bottom line.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Myth #4: AI will fix monitoring automatically&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvvscrmk30qxjwzipftwq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvvscrmk30qxjwzipftwq.png" alt="A computer screen showing a computer and trash cansAI-generated content may be incorrect., Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The reality:&lt;/strong&gt; AI is only as good as the data it analyzes—clean, contextual data is essential for success.&lt;/p&gt;

&lt;p&gt;AI can enhance monitoring by identifying patterns and predicting failures, but poor data quality undermines its effectiveness. Feed it garbage, and you’ll get polished garbage out. &lt;a href="https://www.gartner.com/en/newsroom/press-releases/2025-02-26-lack-of-ai-ready-data-puts-ai-projects-at-risk" rel="noopener noreferrer"&gt;Gartner&lt;/a&gt; warns that by 2026, 60% of AI-driven IT projects will fail without proper data readiness.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Invest in data quality:&lt;/strong&gt; Establish governance frameworks to ensure clean and consistent data inputs for AI models.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Adopt shift-left Observability:&lt;/strong&gt; Integrate monitoring into development cycles to identify issues earlier in the lifecycle.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Tailor AI solutions by context:&lt;/strong&gt; Customize AI-driven monitoring strategies based on the criticality of each application or service.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Organizations must adopt a "shift-left" approach to monitoring, where monitoring is integrated into the development lifecycle from the beginning. This allows organizations to identify and address potential issues early on, reducing the risk of costly downtime and performance problems.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Myth #5: Downtime is the only metric that matters&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvo6or3vq3dhzi1yfzqhl.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvo6or3vq3dhzi1yfzqhl.jpeg" alt="Picture 2, Picture" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The reality:&lt;/strong&gt; Slow is the new down—Performance degradation can erode trust long before outages occur.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.catchpoint.com/statistics" rel="noopener noreferrer"&gt;53% of mobile users drop off&lt;/a&gt; if a page takes more than 3 seconds to load. Performance degradation silently erodes trust before outages even hit. Monitoring must evolve from tracking availability alone to measuring user experience metrics like page load times and transaction speeds.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What you can do:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Monitor user experience metrics:&lt;/strong&gt; Track latency, load times, and transaction completion rates alongside traditional uptime metrics. These XLOs ensure monitoring aligns with user satisfaction and business outcomes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Use predictive analytics:&lt;/strong&gt; Leverage historical data trends to anticipate potential slowdowns before they impact users, enabling proactive intervention.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Implement proactive remediation plans:&lt;/strong&gt; Automate responses for common performance issues, such as traffic spikes or resource bottlenecks, to minimize user impact and ensure seamless experiences.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Leveraging AI-powered predictive analytics to improve IT operations will enable organizations to move from a reactive to a proactive approach to monitoring, reducing downtime and improving overall system reliability&lt;/p&gt;

&lt;h2&gt;
  
  
  Rethinking monitoring for the age of complexity
&lt;/h2&gt;

&lt;p&gt;As enterprises face increasing complexity, monitoring has evolved from a back-office function to a strategic enabler of resilience, customer trust, and competitive differentiation. CIOs who cling to outdated assumptions risk falling behind—not just competitors, but their own customers’ expectations. The myths addressed in this article highlight the need for a paradigm shift in how organizations approach monitoring.&lt;/p&gt;

&lt;p&gt;Modern monitoring isn’t just about uptime or data collection; it’s about aligning IT performance with business outcomes, prioritizing user experience, and leveraging predictive analytics to stay ahead of issues. By embracing these principles, CIOs can transform monitoring into a competitive advantage.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Key takeaways for CIOs&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Monitoring is strategic:&lt;/strong&gt; Elevate monitoring from an IT operations function to a C-suite priority tied directly to revenue and customer satisfaction.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Focus on actionable insights:&lt;/strong&gt; Collect the right data—not just more data—and use AI-driven tools to surface meaningful patterns.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Expand visibility:&lt;/strong&gt; Go beyond internal metrics to monitor end-to-end user experiences and external factors affecting performance.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Prioritize data quality:&lt;/strong&gt; Invest in clean, contextual data to unlock the full potential of AI-driven monitoring.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Measure user experience:&lt;/strong&gt; Adopt XLOs to track metrics that reflect customer satisfaction alongside technical KPIs.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Ask yourself: Are you monitoring what truly matters—or just what’s easy? The answer will define your organization’s ability to thrive in an era where seamless digital experiences are the foundation of success.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Dig deeper:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Check out &lt;a href="https://vmblog.com/archive/2023/02/22/goodbye-lan-the-internet-is-the-network.aspx" rel="noopener noreferrer"&gt;&lt;em&gt;Goodbye LAN – The Internet is the Network&lt;/em&gt;&lt;/a&gt; on VMblog to learn how the Internet became the new enterprise network.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.catchpoint.com/blog/mastering-ipm-monitor-what-matters-from-where-it-matters" rel="noopener noreferrer"&gt;&lt;em&gt;Mastering IPM: Monitor what matters from where it matters&lt;/em&gt;&lt;/a&gt; – learn how Internet Performance Monitoring helps you stay resilient by focusing on the right metrics, from the right vantage points.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

</description>
      <category>monitoring</category>
      <category>cio</category>
      <category>observability</category>
    </item>
    <item>
      <title>Critical Requirements for Modern API Monitoring</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 01 Sep 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/critical-requirements-for-modern-api-monitoring-58o7</link>
      <guid>https://forem.com/catchpoint/critical-requirements-for-modern-api-monitoring-58o7</guid>
      <description>&lt;p&gt;Enterprises lose millions annually due to API outages and performance degradation. Modern observability strategies are crucial to mitigate these risks.&lt;/p&gt;

&lt;p&gt;Today, almost every system is dependent on APIs. Data integration, authentication, payment processing, and many other functions rely on multiple reliable and performant APIs. Banks around the world, for example, have adopted the &lt;a href="https://stripe.com/resources/more/open-banking-apis-explained-what-they-are-and-how-they-work" rel="noopener noreferrer"&gt;Open Banking API&lt;/a&gt; for payments, credit scoring, lending origination, fraud detection, and lots more.&lt;/p&gt;

&lt;h2&gt;
  
  
  APIs are everywhere—and critical to everything
&lt;/h2&gt;

&lt;p&gt;APIs are the internal workers of the internet.  Connecting to a single website, using a business application like an ATM, or a mobile app likely means dozens if not hundreds or even thousands of API calls.  Each one has the possibility to impact the overall service: if it’s slow, the service might be slow.  If it returns an error, the service may fail.  Understanding the interaction between your services and the APIs they consume is critical to making your services resilient.&lt;/p&gt;

&lt;p&gt;There are different ways to monitor APIs.  The minimum that every system should use is proactively monitoring, measuring, and testing your critical APIs – both your own as well as third party APIs. &lt;a href="https://www.catchpoint.com/application-experience/api-monitoring" rel="noopener noreferrer"&gt;API monitoring systems&lt;/a&gt; have been around for some time. From the basic ping to ensure an API is reachable to more advanced multi-step, scripted, proactive monitoring that looks at response time, functional validation, etc. More advanced API monitoring will incorporate Chaos Engineering methodologies, like blocking or simulating an error on particular APIs and observing the impact on the overall system.&lt;/p&gt;

&lt;h2&gt;
  
  
  API resilience isn’t optional—here’s why
&lt;/h2&gt;

&lt;p&gt;In a world where most applications and systems that are interconnected via APIs are geographically distributed, touching different clouds and traversing multiple points across the internet, simple proactive monitoring is no longer enough.  A traditional approach to API monitoring will not only be insufficient, but it may also miss many important incidents and would not be helpful in identifying root cause.&lt;/p&gt;

&lt;p&gt;The goal is to have &lt;em&gt;resilient APIs&lt;/em&gt;. The &lt;a href="https://www.catchpoint.com/blog/reachability-availability-performance-reliability" rel="noopener noreferrer"&gt;formula for resilience&lt;/a&gt; is&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Reachability:&lt;/strong&gt; Can I get to the API from where I am? (Or in the case of APIs, where its consumers are).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Availability:&lt;/strong&gt; Is the API functional – does it do what it is supposed to do?&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Performance:&lt;/strong&gt; Does the API respond within the expected time?&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Reliability:&lt;/strong&gt; Can I trust the API will be working consistently?&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Then, take this formula and combine it for every API used in your system!&lt;/p&gt;

&lt;p&gt;System Resilience = minimum resilience across all APIs in use&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2rmmp6rb4g7diyxw6b0r.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2rmmp6rb4g7diyxw6b0r.png" alt="Picture 1, Picture" width="606" height="124"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Let’s illustrate the point:&lt;/p&gt;

&lt;p&gt;&lt;em&gt;A system that has 100 APIs with five nines of availability must have five nines of availability from each API!  If designed in a resilient way, it’s possible to have individual API dependencies fail without impacting the overall system, but it has to be carefully designed and verified to be resilient.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2iy61svp3iulmmhqtyn7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2iy61svp3iulmmhqtyn7.png" alt="Picture 1, Picture" width="530" height="490"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Figure 1: One misbehaving dependency can make your whole system fail&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;With this goal in mind, let’s understand the requirements for API monitoring.  &lt;/p&gt;

&lt;h2&gt;
  
  
  What a basic API Monitoring strategy should include
&lt;/h2&gt;

&lt;p&gt;These are the foundational capabilities every API monitoring strategy should include. They help teams detect issues, ensure availability, and validate performance at a basic operational level.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Response Time:&lt;/strong&gt; Measures the time taken for an API to respond to a request, helping identify latency issues.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Error Rate:&lt;/strong&gt; Tracks the percentage of failed requests to detect anomalies or bugs.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Throughput: Monitors the number of API requests processed over a specific period to ensure scalability.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Uptime and Availability:&lt;/strong&gt; Ensures APIs are consistently reachable and operational.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Logging:&lt;/strong&gt; Collect detailed logs of API events, including timestamps, event types (e.g., errors, warnings), and messages, to aid in troubleshooting and post-incident analysis.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Alerts:&lt;/strong&gt; Set up alerts based on predefined thresholds or anomalies (e.g., response time exceeding 200ms or error rates surpassing 5%).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Functional testing: verify that API endpoints return expected results  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;CI/CD integration:&lt;/strong&gt; the ability to integrate monitoring into pipelines (and tools like &lt;a href="https://www.jenkins.io/" rel="noopener noreferrer"&gt;Jenkins&lt;/a&gt; or &lt;a href="https://www.terraform.io/" rel="noopener noreferrer"&gt;Terraform&lt;/a&gt;) for automated creation and update of tests also known as “monitoring as code”.  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Proactive monitoring:&lt;/strong&gt; Use of synthetic mechanisms to continuously observe API performance to detect issues as they occur.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Scripting:&lt;/strong&gt; Support for scripting standards such as Playwright to enable testing specific customer &amp;amp; API flows.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Historic data:&lt;/strong&gt; A minimum of 13-month data retention to enable comparison of performance with the same period a year ago&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;High-cardinality data analysis:&lt;/strong&gt; Analyze detailed data points such as unique user IDs or session-specific information for granular insights into performance trends or anomalies.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Chaos Engineering:&lt;/strong&gt; Introduce errors into the system on purpose during a period of low use, or in a non-production environment in order to verify resilience.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Modern, internet-centric API monitoring requirements
&lt;/h2&gt;

&lt;p&gt;Today’s systems, however, demand more than basic uptime checks or response metrics. Modern API monitoring must account for real-world complexity—geography, infrastructure, user experience, and external dependencies. The capabilities below go beyond the basics to provide deep, actionable insight.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Monitor from where it matters&lt;/strong&gt; – most monitoring tools have agents in cloud servers, which often have very different connectivity, resources, and bandwidth than real-world systems, and are blind to geographical differences in routing, ISP congestion, etc. It is critical to &lt;a href="https://www.catchpoint.com/blog/the-power-of-over-3000-intelligent-observability-agents" rel="noopener noreferrer"&gt;monitor from all the locations from where a system will consume an API&lt;/a&gt;, using an agent that has similar characteristics. It is also important to consider where your intermediate services are. For example, you may test your full application from the customer-facing API from last-mile agents. Then test intermediate microservices from the cloud provider where they're located, and test your back-end API from a backbone agent in the city &amp;amp; ISP where your datacenter is located or use an enterprise agent inside the datacenter.  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Visibility into the Internet Stack&lt;/strong&gt; – While it is useful to know &lt;em&gt;when&lt;/em&gt; an API is unresponsive or slow, it is more powerful to understand why. Modern API monitoring provides insight into anything in the &lt;a href="https://www.catchpoint.com/glossary/internet-stack" rel="noopener noreferrer"&gt;Internet Stack&lt;/a&gt; that impacts an API including DNS resolution, SSL, routing, etc. – as well as visibility into the impact on latency and performance introduced by systems such as internal networks, SASE implementations, or gateways.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0pugweeon1qp91yoe36b.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0pugweeon1qp91yoe36b.png" alt="A blue and purple rectangular chart with iconsAI-generated content may be incorrect., Picture" width="800" height="358"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The Internet Stack&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Authentication&lt;/strong&gt; – No modern monitoring system should have hard-coded credentials into a secure API, therefore monitoring systems must support secrets management, OAuth, tokens and modern authentication mechanisms.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Synthetic Code Tracing&lt;/strong&gt; - As an API is being tested, collect and understand code execution traces to identify server-side issues including application, connectivity, and database problems.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Open Telemetry Support&lt;/strong&gt; – modern observability implementations must support OTel as the standard mechanism to share and integrate data from multiple systems and to provide flexibility.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Focus on User experience&lt;/strong&gt; – An API is only a component of a broader overall system performance. A payment API is likely a component of an online purchase transaction. You want to ensure the entire transaction system performs – from the end-user perspective. Ideally, an operations team would be able to see a &lt;a href="https://www.catchpoint.com/internet-stack-map" rel="noopener noreferrer"&gt;visual representation of every dependency in the user transaction&lt;/a&gt;, from end-user across the internet, network, systems, APIs – all the way into code tracing.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Broad support for protocols&lt;/strong&gt; – While many APIs use REST over HTTP, it may be important for your monitoring system to be able to test from both IPv4 and IPv6 agents, as well as support modern protocols like http/3 and QUIC,  MQTT for IoT applications, NTP for time synchronization, or even custom or proprietary protocols that your applications use.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Traditional vs modern API monitoring at a glance
&lt;/h2&gt;

&lt;p&gt;The following table summarizes the key differences between legacy API monitoring approaches and modern, Internet Performance Monitoring strategies that support resilience and user experience.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Feature: Scope

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Server-centric metrics&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): End-to-end user experience + infrastructure&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Protocol Support

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Limited to HTTP/S, REST&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): HTTP/3, QUIC, MQTT, custom protocols&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Data Granularity

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: High-cardinality data available but often limited to service boundaries&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): High-cardinality traces with cross-system correlation (user IDs, sessions)&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Root Cause Analysis

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Limited to app/server layers&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): Full internet stack (DNS, SSL, routing, etc.)&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Testing Perspective

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Cloud datacenters&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): Last mile, backbone, cloud, wireless, and enterprise intelligent agents&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Performance Context

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: API performance in the context of code&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): API performance in context of user experience&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Alerting Methodology

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Alert thresholds based on error rates&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): Experience scores and XLOs&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Visualization

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Code-centric dashboards&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): Visual representation of everything impacting a system&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;/ul&gt;

&lt;h2&gt;
  
  
  Rethink what API monitoring should do
&lt;/h2&gt;

&lt;p&gt;It is somewhat surprising that the cloud is only 15 years old. As technology and system architecture has evolved, our monitoring has to evolve including API monitoring.  &lt;/p&gt;

&lt;p&gt;What is today considered “owned” or “on-premises” infrastructure is most likely in a colocation datacenter, relying on a DNS and SSL provider, connected through at least two ISPs, depending on a cloud-based authentication system, connected through a cloud-based security provider and maybe a few other APIs.&lt;/p&gt;

&lt;p&gt;We hear all the time “My APM system shows my systems are green, but my users keep complaining”. A monitoring system that only monitors your “on premises” API is not going to be able to spot, diagnose, or provide useful root-cause information to prevent or solve incidents quickly.&lt;/p&gt;

&lt;p&gt;To ensure API resilience, enterprises must invest in modern monitoring solutions that provide end-to-end visibility, proactive alerting, and automated remediation capabilities.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Ready to modernize your API monitoring?&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Discover how &lt;a href="https://www.catchpoint.com/application-experience/api-monitoring" rel="noopener noreferrer"&gt;Catchpoint’s API Monitoring&lt;/a&gt; delivers the end-to-end visibility and resilience your users expect.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;Enterprises lose millions annually due to API outages and performance degradation. Modern observability strategies are crucial to mitigate these risks.&lt;/p&gt;

&lt;p&gt;Today, almost every system is dependent on APIs. Data integration, authentication, payment processing, and many other functions rely on multiple reliable and performant APIs. Banks around the world, for example, have adopted the &lt;a href="https://stripe.com/resources/more/open-banking-apis-explained-what-they-are-and-how-they-work" rel="noopener noreferrer"&gt;Open Banking API&lt;/a&gt; for payments, credit scoring, lending origination, fraud detection, and lots more.&lt;/p&gt;

&lt;h2&gt;
  
  
  APIs are everywhere—and critical to everything
&lt;/h2&gt;

&lt;p&gt;APIs are the internal workers of the internet.  Connecting to a single website, using a business application like an ATM, or a mobile app likely means dozens if not hundreds or even thousands of API calls.  Each one has the possibility to impact the overall service: if it’s slow, the service might be slow.  If it returns an error, the service may fail.  Understanding the interaction between your services and the APIs they consume is critical to making your services resilient.&lt;/p&gt;

&lt;p&gt;There are different ways to monitor APIs.  The minimum that every system should use is proactively monitoring, measuring, and testing your critical APIs – both your own as well as third party APIs. &lt;a href="https://www.catchpoint.com/application-experience/api-monitoring" rel="noopener noreferrer"&gt;API monitoring systems&lt;/a&gt; have been around for some time. From the basic ping to ensure an API is reachable to more advanced multi-step, scripted, proactive monitoring that looks at response time, functional validation, etc. More advanced API monitoring will incorporate Chaos Engineering methodologies, like blocking or simulating an error on particular APIs and observing the impact on the overall system.&lt;/p&gt;

&lt;h2&gt;
  
  
  API resilience isn’t optional—here’s why
&lt;/h2&gt;

&lt;p&gt;In a world where most applications and systems that are interconnected via APIs are geographically distributed, touching different clouds and traversing multiple points across the internet, simple proactive monitoring is no longer enough.  A traditional approach to API monitoring will not only be insufficient, but it may also miss many important incidents and would not be helpful in identifying root cause.&lt;/p&gt;

&lt;p&gt;The goal is to have &lt;em&gt;resilient APIs&lt;/em&gt;. The &lt;a href="https://www.catchpoint.com/blog/reachability-availability-performance-reliability" rel="noopener noreferrer"&gt;formula for resilience&lt;/a&gt; is&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Reachability:&lt;/strong&gt; Can I get to the API from where I am? (Or in the case of APIs, where its consumers are).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Availability:&lt;/strong&gt; Is the API functional – does it do what it is supposed to do?&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Performance:&lt;/strong&gt; Does the API respond within the expected time?&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Reliability:&lt;/strong&gt; Can I trust the API will be working consistently?&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Then, take this formula and combine it for every API used in your system!&lt;/p&gt;

&lt;p&gt;System Resilience = minimum resilience across all APIs in use&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2rmmp6rb4g7diyxw6b0r.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2rmmp6rb4g7diyxw6b0r.png" alt="Picture 1, Picture" width="606" height="124"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Let’s illustrate the point:&lt;/p&gt;

&lt;p&gt;&lt;em&gt;A system that has 100 APIs with five nines of availability must have five nines of availability from each API!  If designed in a resilient way, it’s possible to have individual API dependencies fail without impacting the overall system, but it has to be carefully designed and verified to be resilient.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2iy61svp3iulmmhqtyn7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2iy61svp3iulmmhqtyn7.png" alt="Picture 1, Picture" width="530" height="490"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Figure 1: One misbehaving dependency can make your whole system fail&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;With this goal in mind, let’s understand the requirements for API monitoring.  &lt;/p&gt;

&lt;h2&gt;
  
  
  What a basic API Monitoring strategy should include
&lt;/h2&gt;

&lt;p&gt;These are the foundational capabilities every API monitoring strategy should include. They help teams detect issues, ensure availability, and validate performance at a basic operational level.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Response Time:&lt;/strong&gt; Measures the time taken for an API to respond to a request, helping identify latency issues.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Error Rate:&lt;/strong&gt; Tracks the percentage of failed requests to detect anomalies or bugs.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Throughput: Monitors the number of API requests processed over a specific period to ensure scalability.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Uptime and Availability:&lt;/strong&gt; Ensures APIs are consistently reachable and operational.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Logging:&lt;/strong&gt; Collect detailed logs of API events, including timestamps, event types (e.g., errors, warnings), and messages, to aid in troubleshooting and post-incident analysis.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Alerts:&lt;/strong&gt; Set up alerts based on predefined thresholds or anomalies (e.g., response time exceeding 200ms or error rates surpassing 5%).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Functional testing: verify that API endpoints return expected results  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;CI/CD integration:&lt;/strong&gt; the ability to integrate monitoring into pipelines (and tools like &lt;a href="https://www.jenkins.io/" rel="noopener noreferrer"&gt;Jenkins&lt;/a&gt; or &lt;a href="https://www.terraform.io/" rel="noopener noreferrer"&gt;Terraform&lt;/a&gt;) for automated creation and update of tests also known as “monitoring as code”.  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Proactive monitoring:&lt;/strong&gt; Use of synthetic mechanisms to continuously observe API performance to detect issues as they occur.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Scripting:&lt;/strong&gt; Support for scripting standards such as Playwright to enable testing specific customer &amp;amp; API flows.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Historic data:&lt;/strong&gt; A minimum of 13-month data retention to enable comparison of performance with the same period a year ago&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;High-cardinality data analysis:&lt;/strong&gt; Analyze detailed data points such as unique user IDs or session-specific information for granular insights into performance trends or anomalies.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Chaos Engineering:&lt;/strong&gt; Introduce errors into the system on purpose during a period of low use, or in a non-production environment in order to verify resilience.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Modern, internet-centric API monitoring requirements
&lt;/h2&gt;

&lt;p&gt;Today’s systems, however, demand more than basic uptime checks or response metrics. Modern API monitoring must account for real-world complexity—geography, infrastructure, user experience, and external dependencies. The capabilities below go beyond the basics to provide deep, actionable insight.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Monitor from where it matters&lt;/strong&gt; – most monitoring tools have agents in cloud servers, which often have very different connectivity, resources, and bandwidth than real-world systems, and are blind to geographical differences in routing, ISP congestion, etc. It is critical to &lt;a href="https://www.catchpoint.com/blog/the-power-of-over-3000-intelligent-observability-agents" rel="noopener noreferrer"&gt;monitor from all the locations from where a system will consume an API&lt;/a&gt;, using an agent that has similar characteristics. It is also important to consider where your intermediate services are. For example, you may test your full application from the customer-facing API from last-mile agents. Then test intermediate microservices from the cloud provider where they're located, and test your back-end API from a backbone agent in the city &amp;amp; ISP where your datacenter is located or use an enterprise agent inside the datacenter.  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Visibility into the Internet Stack&lt;/strong&gt; – While it is useful to know &lt;em&gt;when&lt;/em&gt; an API is unresponsive or slow, it is more powerful to understand why. Modern API monitoring provides insight into anything in the &lt;a href="https://www.catchpoint.com/glossary/internet-stack" rel="noopener noreferrer"&gt;Internet Stack&lt;/a&gt; that impacts an API including DNS resolution, SSL, routing, etc. – as well as visibility into the impact on latency and performance introduced by systems such as internal networks, SASE implementations, or gateways.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0pugweeon1qp91yoe36b.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0pugweeon1qp91yoe36b.png" alt="A blue and purple rectangular chart with iconsAI-generated content may be incorrect., Picture" width="800" height="358"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The Internet Stack&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Authentication&lt;/strong&gt; – No modern monitoring system should have hard-coded credentials into a secure API, therefore monitoring systems must support secrets management, OAuth, tokens and modern authentication mechanisms.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Synthetic Code Tracing&lt;/strong&gt; - As an API is being tested, collect and understand code execution traces to identify server-side issues including application, connectivity, and database problems.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Open Telemetry Support&lt;/strong&gt; – modern observability implementations must support OTel as the standard mechanism to share and integrate data from multiple systems and to provide flexibility.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Focus on User experience&lt;/strong&gt; – An API is only a component of a broader overall system performance. A payment API is likely a component of an online purchase transaction. You want to ensure the entire transaction system performs – from the end-user perspective. Ideally, an operations team would be able to see a &lt;a href="https://www.catchpoint.com/internet-stack-map" rel="noopener noreferrer"&gt;visual representation of every dependency in the user transaction&lt;/a&gt;, from end-user across the internet, network, systems, APIs – all the way into code tracing.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Broad support for protocols&lt;/strong&gt; – While many APIs use REST over HTTP, it may be important for your monitoring system to be able to test from both IPv4 and IPv6 agents, as well as support modern protocols like http/3 and QUIC,  MQTT for IoT applications, NTP for time synchronization, or even custom or proprietary protocols that your applications use.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Traditional vs modern API monitoring at a glance
&lt;/h2&gt;

&lt;p&gt;The following table summarizes the key differences between legacy API monitoring approaches and modern, Internet Performance Monitoring strategies that support resilience and user experience.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Feature: Scope

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Server-centric metrics&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): End-to-end user experience + infrastructure&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Protocol Support

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Limited to HTTP/S, REST&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): HTTP/3, QUIC, MQTT, custom protocols&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Data Granularity

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: High-cardinality data available but often limited to service boundaries&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): High-cardinality traces with cross-system correlation (user IDs, sessions)&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Root Cause Analysis

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Limited to app/server layers&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): Full internet stack (DNS, SSL, routing, etc.)&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Testing Perspective

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Cloud datacenters&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): Last mile, backbone, cloud, wireless, and enterprise intelligent agents&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Performance Context

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: API performance in the context of code&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): API performance in context of user experience&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Alerting Methodology

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Alert thresholds based on error rates&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): Experience scores and XLOs&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Feature: Visualization

&lt;ul&gt;
&lt;li&gt;Traditional API Monitoring: Code-centric dashboards&lt;/li&gt;
&lt;li&gt;Modern API Monitoring (IPM): Visual representation of everything impacting a system&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;/ul&gt;

&lt;h2&gt;
  
  
  Rethink what API monitoring should do
&lt;/h2&gt;

&lt;p&gt;It is somewhat surprising that the cloud is only 15 years old. As technology and system architecture has evolved, our monitoring has to evolve including API monitoring.  &lt;/p&gt;

&lt;p&gt;What is today considered “owned” or “on-premises” infrastructure is most likely in a colocation datacenter, relying on a DNS and SSL provider, connected through at least two ISPs, depending on a cloud-based authentication system, connected through a cloud-based security provider and maybe a few other APIs.&lt;/p&gt;

&lt;p&gt;We hear all the time “My APM system shows my systems are green, but my users keep complaining”. A monitoring system that only monitors your “on premises” API is not going to be able to spot, diagnose, or provide useful root-cause information to prevent or solve incidents quickly.&lt;/p&gt;

&lt;p&gt;To ensure API resilience, enterprises must invest in modern monitoring solutions that provide end-to-end visibility, proactive alerting, and automated remediation capabilities.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Ready to modernize your API monitoring?&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Discover how &lt;a href="https://www.catchpoint.com/application-experience/api-monitoring" rel="noopener noreferrer"&gt;Catchpoint’s API Monitoring&lt;/a&gt; delivers the end-to-end visibility and resilience your users expect.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

&lt;h2&gt;
  
  
  You might also like
&lt;/h2&gt;

&lt;h3&gt;
  
  
  LLMs don’t stand still: How to monitor and trust the models powering your AI
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.catchpoint.com/blog/llms-dont-stand-still-how-to-monitor-and-trust-the-models-powering-your-ai" rel="noopener noreferrer"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  APM vs observability: why your definitions are broken
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.catchpoint.com/blog/apm-vs-observability-why-your-definitions-are-broken" rel="noopener noreferrer"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Semantic Caching: What We Measured, Why It Matters
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.catchpoint.com/blog/semantic-caching-what-we-measured-why-it-matters" rel="noopener noreferrer"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>api</category>
      <category>monitoring</category>
      <category>observability</category>
    </item>
    <item>
      <title>Why Intelligent Traffic Steering is Critical for Performance and Cost Optimization</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 25 Aug 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/why-intelligent-traffic-steering-is-critical-for-performance-and-cost-optimization-1om9</link>
      <guid>https://forem.com/catchpoint/why-intelligent-traffic-steering-is-critical-for-performance-and-cost-optimization-1om9</guid>
      <description>&lt;p&gt;In today’s world of globally distributed applications, user experience is everything. Whether your platform runs across multiple cloud providers or uses a Multi CDN with numerous points of presence (PoPs), efficiently routing user traffic can make or break performance. That's where intelligent traffic steering becomes not just a nice-to-have, but a must-have.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F70dieewamr3bf4u3tmhx.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F70dieewamr3bf4u3tmhx.png" alt="This diagram illustrates the complex web of connections between a modern core banking system and its extended digital ecosystem. It includes cloud-based services, APIs, third-party integrations, CDN layers, internal networks, and diverse user access points. Components are connected via ISPs and DNS lookups, highlighting potential performance bottlenecks and visibility gaps across digital services. " width="800" height="453"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;End-to-end ecosystem of a core banking system&lt;/p&gt;

&lt;p&gt;At our recent joint webinar with IBM NS1, we explored how Catchpoint's real-time &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) data integrates with NS1's powerful traffic steering capabilities to solve a critical problem: ensuring traffic is routed not just to the nearest server, but to the best-performing one.&lt;/p&gt;

&lt;h2&gt;
  
  
  The challenge: static routing in a dynamic world
&lt;/h2&gt;

&lt;p&gt;Traditionally, DNS-based routing strategies like round-robin or proximity-based decisions have been used to direct user traffic. However, these methods don't account for real-time performance metrics. A nearby server might be under heavy load or facing regional ISP issues, yet static routing would still direct users there, resulting in higher latency and degraded experience.&lt;/p&gt;

&lt;p&gt;This issue is magnified in multi-cloud and multi-CDN environments, where routing inefficiencies can also drive up cloud and infrastructure costs. Without real-time visibility, you’re essentially flying blind.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd8juctmxnofjfutsspop.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd8juctmxnofjfutsspop.png" alt="The maps on the left compare AWS and Oracle response times across different global regions, highlighting areas where performance varies significantly. Green indicates faster response times, while red signals slower performance. The lower maps offer a city-level breakdown. The chart on the right shows traffic distribution before dynamic routing, with traffic nearly evenly split: 52% to AWS and 48% to Oracle." width="800" height="380"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Regional performance differences and traffic distribution between AWS and Oracle cloud environments&lt;/p&gt;

&lt;p&gt;The image above provides a great example of this challenge in a multi-cloud environment using AWS and Oracle. The global response time maps show noticeable regional performance differences between the two providers. Yet, because traffic is being routed using a static round-robin method (seen in the 52% AWS / 48% Oracle split), users in underperforming regions are still being directed to slower endpoints. For instance, the Oracle path shows significant latency in North America and Oceania, while AWS performs better there. Still, the lack of performance-aware steering results in an inefficient and inconsistent user experience.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Solution: Adaptive traffic steering with Catchpoint and IBM NS1
&lt;/h2&gt;

&lt;p&gt;By combining Catchpoint’s extensive &lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;Global Agent Network&lt;/a&gt;—spanning 2,880+ intelligent agents and millions of connected devices—with IBM NS1’s real-time DNS traffic steering capabilities, organizations can shift from reactive to proactive, performance-driven routing.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcizkpzur7ctujx5ox7yn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcizkpzur7ctujx5ox7yn.png" alt="This graphic illustrates the scale and distribution of Catchpoint’s monitoring infrastructure, showcasing the world’s largest independent agent network. With 2,986 intelligent agents deployed across 106 countries and 348 cities." width="800" height="436"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint’s Global Agent Network&lt;/p&gt;

&lt;p&gt;Check out the video below for an overview of how Catchpoint and IBM NS1 work together to enable real-time, performance-aware traffic steering.  &lt;/p&gt;

&lt;p&gt;Catchpoint IPM continuously collects metrics such as DNS resolution time, connect time, SSL, wait time, and total response time using synthetic tests from backbone, last mile, cloud, and wireless networks. This real-time telemetry is then fed into NS1's filter chain logic.&lt;/p&gt;

&lt;p&gt;NS1 applies this data through its intelligent filter chains, which support decisions based on geolocation, ASN, availability, and actual performance. The diagram below illustrates how the end user’s DNS request is routed through NS1, where IPM-powered performance metrics guide the decision-making engine. This ensures traffic is not just sent to the closest server, but the best-performing server at that moment.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpv4pzb4hpkrj26rzbcjn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpv4pzb4hpkrj26rzbcjn.png" alt="This diagram shows the flow of a DNS request and how real-time performance metrics power intelligent traffic steering with NS1 and Catchpoint." width="800" height="361"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;How real time metrics power intelligent traffic steering with NS1 and Catchpoint&lt;/p&gt;

&lt;p&gt;This feedback loop enables true adaptive routing, drastically reducing latency and improving reliability without waiting for end-user impact to trigger change.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Feeding Catchpoint Data into IBM NS1 Connect
&lt;/h2&gt;

&lt;p&gt;Catchpoint enables real-time performance telemetry to be pushed directly into NS1 IBM Connect using Data Webhooks. This integration empowers traffic steering logic to be based on real, actionable insights rather than static assumptions.&lt;/p&gt;

&lt;p&gt;Metrics such as DNS time, connect time, SSL handshake duration, wait time (time to first byte), TTFB, and overall response time are just the start. Users can also feed in custom data sources like CDN-specific server timing metrics or application-specific performance KPIs. This flexibility ensures that the traffic steering logic is tailor-fit to business and technical requirements whether optimizing for speed, reliability, or cost.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fh83xrm7np5ir4s5svvqw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fh83xrm7np5ir4s5svvqw.png" alt="This dashboard shows how real user metrics from Catchpoint are fed into IBM NS1 Connect to compare performance across different cloud providers—in this case, AWS and Oracle. The graph visualizes real-time response times over a 6-hour window, highlighting latency variations that inform traffic steering decisions. " width="800" height="429"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Using Catchpoint IPM data in IBM NS1 Connect to compare multi-cloud performance&lt;/p&gt;

&lt;p&gt;All this data is consumed by NS1 Pulsar's filter chain engine, which uses it to dynamically apply routing rules. These filters can be chained together based on geography, ASN, latency, availability, or any of the performance metrics provided by Catchpoint IPM.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0wiy76ubfavdhg1ukafm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0wiy76ubfavdhg1ukafm.png" alt="NS1 Pulsar filter chain using Catchpoint data to route traffic, with dashboards showing traffic distribution across cloud providers and ASNs." width="800" height="320"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;NS1 Pulsar filter chain powered by Catchpoint IPM data&lt;/p&gt;

&lt;p&gt;As a result, routing decisions aren't just smart they're precisely aligned with your most up-to-date performance landscape.&lt;/p&gt;

&lt;h2&gt;
  
  
  Real-world impact: Better performance and lower costs
&lt;/h2&gt;

&lt;p&gt;In our joint implementation example, traffic was initially split evenly between AWS and Oracle cloud environments. After enabling dynamic routing with IPM-fed data, 77% of the traffic shifted to the better-performing cloud region, significantly reducing wait times and improving end-user experience.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffdmipr4aj4ocnz1pyj3e.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffdmipr4aj4ocnz1pyj3e.png" alt="This side-by-side comparison shows traffic distribution across AWS and Oracle before and after implementing performance-based traffic steering. Before steering, traffic was nearly evenly split (52% AWS, 48% Oracle). After steering, 77% of traffic shifted to AWS, resulting in more efficient regional routing and improved user experience." width="800" height="365"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Impact of traffic steering on cloud traffic distribution&lt;/p&gt;

&lt;p&gt;The results speak for themselves:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;86.57% reduction in wait time&lt;/strong&gt; on AWS&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;85.30% reduction in wait time&lt;/strong&gt; on Oracle&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyhu2pgwdmxz9odm1xmsm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyhu2pgwdmxz9odm1xmsm.png" alt="This graph compares the reduction in wait time (GM Wait in ms) for AWS and Oracle environments after implementing intelligent traffic steering. AWS saw an 86.57% reduction, while Oracle experienced an 85.30% reduction, demonstrating the effectiveness of routing decisions informed by real-time performance data. " width="800" height="404"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Wait time improvements after enabling performance-based traffic steering&lt;/p&gt;

&lt;p&gt;Wait time improvements after enabling performance-based traffic steering&lt;/p&gt;

&lt;p&gt;These improvements not only enhance digital experience but also minimize the frustration and churn that come with sluggish performance.&lt;/p&gt;

&lt;p&gt;Beyond latency improvements, traffic steering also drives operational efficiency. By understanding which cloud or PoP performs best in specific regions, businesses can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Scale down underperforming infrastructure&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Optimize cloud costs by avoiding overprovisioning&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Improve regional performance by steering traffic intelligently&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is the power of proactive, performance-aware traffic steering.&lt;/p&gt;

&lt;h2&gt;
  
  
  Proactive, not reactive
&lt;/h2&gt;

&lt;p&gt;Most traditional monitoring tools rely on real user monitoring (RUM) data, which only becomes useful after users have already experienced performance issues. Catchpoint IPM flips this model by using &lt;a href="https://www.catchpoint.com/synthetic-monitoring" rel="noopener noreferrer"&gt;synthetic testing&lt;/a&gt; for proactive insights. This means traffic can be rerouted before users are impacted, creating a continuous feedback loop that enhances reliability and experience.&lt;/p&gt;

&lt;h2&gt;
  
  
  Next up: the setup guide
&lt;/h2&gt;

&lt;p&gt;In an upcoming post, we’ll walk through the step-by-step implementation of intelligent traffic steering—from configuring Catchpoint tests to feeding metrics into NS1 and setting up intelligent filter chains. Watch this space.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn More&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Check out the &lt;a href="https://www.ibm.com/docs/en/ns1-connect" rel="noopener noreferrer"&gt;IBM NS1 Connect Documentation&lt;/a&gt; to get started with traffic steering and DNS configuration_._  &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;In today’s world of globally distributed applications, user experience is everything. Whether your platform runs across multiple cloud providers or uses a Multi CDN with numerous points of presence (PoPs), efficiently routing user traffic can make or break performance. That's where intelligent traffic steering becomes not just a nice-to-have, but a must-have.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F70dieewamr3bf4u3tmhx.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F70dieewamr3bf4u3tmhx.png" alt="This diagram illustrates the complex web of connections between a modern core banking system and its extended digital ecosystem. It includes cloud-based services, APIs, third-party integrations, CDN layers, internal networks, and diverse user access points. Components are connected via ISPs and DNS lookups, highlighting potential performance bottlenecks and visibility gaps across digital services. " width="800" height="453"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;End-to-end ecosystem of a core banking system&lt;/p&gt;

&lt;p&gt;At our recent joint webinar with IBM NS1, we explored how Catchpoint's real-time &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Internet Performance Monitoring&lt;/a&gt; (IPM) data integrates with NS1's powerful traffic steering capabilities to solve a critical problem: ensuring traffic is routed not just to the nearest server, but to the best-performing one.&lt;/p&gt;

&lt;h2&gt;
  
  
  The challenge: static routing in a dynamic world
&lt;/h2&gt;

&lt;p&gt;Traditionally, DNS-based routing strategies like round-robin or proximity-based decisions have been used to direct user traffic. However, these methods don't account for real-time performance metrics. A nearby server might be under heavy load or facing regional ISP issues, yet static routing would still direct users there, resulting in higher latency and degraded experience.&lt;/p&gt;

&lt;p&gt;This issue is magnified in multi-cloud and multi-CDN environments, where routing inefficiencies can also drive up cloud and infrastructure costs. Without real-time visibility, you’re essentially flying blind.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd8juctmxnofjfutsspop.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd8juctmxnofjfutsspop.png" alt="The maps on the left compare AWS and Oracle response times across different global regions, highlighting areas where performance varies significantly. Green indicates faster response times, while red signals slower performance. The lower maps offer a city-level breakdown. The chart on the right shows traffic distribution before dynamic routing, with traffic nearly evenly split: 52% to AWS and 48% to Oracle." width="800" height="380"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Regional performance differences and traffic distribution between AWS and Oracle cloud environments&lt;/p&gt;

&lt;p&gt;The image above provides a great example of this challenge in a multi-cloud environment using AWS and Oracle. The global response time maps show noticeable regional performance differences between the two providers. Yet, because traffic is being routed using a static round-robin method (seen in the 52% AWS / 48% Oracle split), users in underperforming regions are still being directed to slower endpoints. For instance, the Oracle path shows significant latency in North America and Oceania, while AWS performs better there. Still, the lack of performance-aware steering results in an inefficient and inconsistent user experience.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Solution: Adaptive traffic steering with Catchpoint and IBM NS1
&lt;/h2&gt;

&lt;p&gt;By combining Catchpoint’s extensive &lt;a href="https://www.catchpoint.com/global-observability-network" rel="noopener noreferrer"&gt;Global Agent Network&lt;/a&gt;—spanning 2,880+ intelligent agents and millions of connected devices—with IBM NS1’s real-time DNS traffic steering capabilities, organizations can shift from reactive to proactive, performance-driven routing.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcizkpzur7ctujx5ox7yn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcizkpzur7ctujx5ox7yn.png" alt="This graphic illustrates the scale and distribution of Catchpoint’s monitoring infrastructure, showcasing the world’s largest independent agent network. With 2,986 intelligent agents deployed across 106 countries and 348 cities." width="800" height="436"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint’s Global Agent Network&lt;/p&gt;

&lt;p&gt;Check out the video below for an overview of how Catchpoint and IBM NS1 work together to enable real-time, performance-aware traffic steering.  &lt;/p&gt;

&lt;p&gt;Catchpoint IPM continuously collects metrics such as DNS resolution time, connect time, SSL, wait time, and total response time using synthetic tests from backbone, last mile, cloud, and wireless networks. This real-time telemetry is then fed into NS1's filter chain logic.&lt;/p&gt;

&lt;p&gt;NS1 applies this data through its intelligent filter chains, which support decisions based on geolocation, ASN, availability, and actual performance. The diagram below illustrates how the end user’s DNS request is routed through NS1, where IPM-powered performance metrics guide the decision-making engine. This ensures traffic is not just sent to the closest server, but the best-performing server at that moment.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpv4pzb4hpkrj26rzbcjn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpv4pzb4hpkrj26rzbcjn.png" alt="This diagram shows the flow of a DNS request and how real-time performance metrics power intelligent traffic steering with NS1 and Catchpoint." width="800" height="361"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;How real time metrics power intelligent traffic steering with NS1 and Catchpoint&lt;/p&gt;

&lt;p&gt;This feedback loop enables true adaptive routing, drastically reducing latency and improving reliability without waiting for end-user impact to trigger change.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Feeding Catchpoint Data into IBM NS1 Connect
&lt;/h2&gt;

&lt;p&gt;Catchpoint enables real-time performance telemetry to be pushed directly into NS1 IBM Connect using Data Webhooks. This integration empowers traffic steering logic to be based on real, actionable insights rather than static assumptions.&lt;/p&gt;

&lt;p&gt;Metrics such as DNS time, connect time, SSL handshake duration, wait time (time to first byte), TTFB, and overall response time are just the start. Users can also feed in custom data sources like CDN-specific server timing metrics or application-specific performance KPIs. This flexibility ensures that the traffic steering logic is tailor-fit to business and technical requirements whether optimizing for speed, reliability, or cost.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fh83xrm7np5ir4s5svvqw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fh83xrm7np5ir4s5svvqw.png" alt="This dashboard shows how real user metrics from Catchpoint are fed into IBM NS1 Connect to compare performance across different cloud providers—in this case, AWS and Oracle. The graph visualizes real-time response times over a 6-hour window, highlighting latency variations that inform traffic steering decisions. " width="800" height="429"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Using Catchpoint IPM data in IBM NS1 Connect to compare multi-cloud performance&lt;/p&gt;

&lt;p&gt;All this data is consumed by NS1 Pulsar's filter chain engine, which uses it to dynamically apply routing rules. These filters can be chained together based on geography, ASN, latency, availability, or any of the performance metrics provided by Catchpoint IPM.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0wiy76ubfavdhg1ukafm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0wiy76ubfavdhg1ukafm.png" alt="NS1 Pulsar filter chain using Catchpoint data to route traffic, with dashboards showing traffic distribution across cloud providers and ASNs." width="800" height="320"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;NS1 Pulsar filter chain powered by Catchpoint IPM data&lt;/p&gt;

&lt;p&gt;As a result, routing decisions aren't just smart they're precisely aligned with your most up-to-date performance landscape.&lt;/p&gt;

&lt;h2&gt;
  
  
  Real-world impact: Better performance and lower costs
&lt;/h2&gt;

&lt;p&gt;In our joint implementation example, traffic was initially split evenly between AWS and Oracle cloud environments. After enabling dynamic routing with IPM-fed data, 77% of the traffic shifted to the better-performing cloud region, significantly reducing wait times and improving end-user experience.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffdmipr4aj4ocnz1pyj3e.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffdmipr4aj4ocnz1pyj3e.png" alt="This side-by-side comparison shows traffic distribution across AWS and Oracle before and after implementing performance-based traffic steering. Before steering, traffic was nearly evenly split (52% AWS, 48% Oracle). After steering, 77% of traffic shifted to AWS, resulting in more efficient regional routing and improved user experience." width="800" height="365"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Impact of traffic steering on cloud traffic distribution&lt;/p&gt;

&lt;p&gt;The results speak for themselves:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;86.57% reduction in wait time&lt;/strong&gt; on AWS&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;85.30% reduction in wait time&lt;/strong&gt; on Oracle&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyhu2pgwdmxz9odm1xmsm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyhu2pgwdmxz9odm1xmsm.png" alt="This graph compares the reduction in wait time (GM Wait in ms) for AWS and Oracle environments after implementing intelligent traffic steering. AWS saw an 86.57% reduction, while Oracle experienced an 85.30% reduction, demonstrating the effectiveness of routing decisions informed by real-time performance data. " width="800" height="404"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Wait time improvements after enabling performance-based traffic steering&lt;/p&gt;

&lt;p&gt;Wait time improvements after enabling performance-based traffic steering&lt;/p&gt;

&lt;p&gt;These improvements not only enhance digital experience but also minimize the frustration and churn that come with sluggish performance.&lt;/p&gt;

&lt;p&gt;Beyond latency improvements, traffic steering also drives operational efficiency. By understanding which cloud or PoP performs best in specific regions, businesses can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Scale down underperforming infrastructure&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Optimize cloud costs by avoiding overprovisioning&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Improve regional performance by steering traffic intelligently&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is the power of proactive, performance-aware traffic steering.&lt;/p&gt;

&lt;h2&gt;
  
  
  Proactive, not reactive
&lt;/h2&gt;

&lt;p&gt;Most traditional monitoring tools rely on real user monitoring (RUM) data, which only becomes useful after users have already experienced performance issues. Catchpoint IPM flips this model by using &lt;a href="https://www.catchpoint.com/synthetic-monitoring" rel="noopener noreferrer"&gt;synthetic testing&lt;/a&gt; for proactive insights. This means traffic can be rerouted before users are impacted, creating a continuous feedback loop that enhances reliability and experience.&lt;/p&gt;

&lt;h2&gt;
  
  
  Next up: the setup guide
&lt;/h2&gt;

&lt;p&gt;In an upcoming post, we’ll walk through the step-by-step implementation of intelligent traffic steering—from configuring Catchpoint tests to feeding metrics into NS1 and setting up intelligent filter chains. Watch this space.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn More&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Check out the &lt;a href="https://www.ibm.com/docs/en/ns1-connect" rel="noopener noreferrer"&gt;IBM NS1 Connect Documentation&lt;/a&gt; to get started with traffic steering and DNS configuration_._  &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

</description>
      <category>monitoring</category>
      <category>observability</category>
    </item>
    <item>
      <title>The $1 Million Lesson: Building a Culture of Quality Through SLAs</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 18 Aug 2025 04:00:00 +0000</pubDate>
      <link>https://forem.com/catchpoint/the-1-million-lesson-building-a-culture-of-quality-through-slas-55i8</link>
      <guid>https://forem.com/catchpoint/the-1-million-lesson-building-a-culture-of-quality-through-slas-55i8</guid>
      <description>&lt;p&gt;In the early days of DoubleClick, back when SaaS was still known as Application Service Provider (ASP), I was tasked with setting up the QoS (Quality of Service) Team. Our primary mission was to establish a monitoring system, but we quickly found ourselves managing Service Level Agreements (SLAs)—a task that became critical after we paid out over &lt;strong&gt;$1 million in penalties&lt;/strong&gt; for SLA violations to a single customer. The reason? Someone had signed a contract promising &lt;strong&gt;100% uptime&lt;/strong&gt;, an impossible commitment.  &lt;/p&gt;

&lt;p&gt;This is the story of how we took control of our SLAs, stopped the financial bleeding, and built a culture of quality around service metrics. Whether you’re managing SLAs today or just curious about how they work, this post will provide valuable insights into the challenges we faced, the solutions we implemented, and the lessons we learned along the way.&lt;/p&gt;

&lt;h2&gt;
  
  
  What are SLAs?
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi27ag99x6l839alc1i5e.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi27ag99x6l839alc1i5e.jpeg" alt="A clipboard with a pen and paper clipsAI-generated content may be incorrect." width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;An &lt;strong&gt;SLA (Service Level Agreement)&lt;/strong&gt; is a contractual agreement between a vendor and a customer that outlines the expected level of service. Under this legal umbrella, you’ll find &lt;strong&gt;Service Level Objectives (SLOs)&lt;/strong&gt;, which define specific metrics like uptime, speed, or transactions per second.&lt;/p&gt;

&lt;p&gt;At DoubleClick, we defined SLAs with the following principles in mind:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Attainable:&lt;/strong&gt; The goals should be realistic.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Repeatable:&lt;/strong&gt; The metrics should be consistently measurable.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Measurable:&lt;/strong&gt; The performance should be quantifiable.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Meaningful:&lt;/strong&gt; The metrics should matter to the business.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Mutually Acceptable:&lt;/strong&gt; Both parties should agree on the terms.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;SLAs benefit both the customer and the vendor. For customers, they provide &lt;strong&gt;objective grading criteria&lt;/strong&gt; and protection from poor service. For vendors, they set clear expectations and incentivize quality improvements&lt;/p&gt;

&lt;h2&gt;
  
  
  Ground zero, discovery
&lt;/h2&gt;

&lt;p&gt;When we first tackled the SLA problem, we were in crisis mode. The first step was to compile a list of all contracts, extract the SLAs and SLOs, and document the associated penalties. We stored this information in a database and began educating stakeholders—business leaders, legal teams, and executives—about the importance of SLAs.&lt;/p&gt;

&lt;p&gt;From the beginning, we focused on &lt;strong&gt;end-user experience-based SLAs&lt;/strong&gt;. This meant measuring performance from the user’s perspective, not just from the server’s perspective.&lt;/p&gt;

&lt;h2&gt;
  
  
  A universal challenge
&lt;/h2&gt;

&lt;p&gt;Over the years, I’ve seen many companies face similar issues. Not all SRE and Dev teams fully grasp the SLAs their organization has with customers—they often focus heavily on internal SLOs while overlooking how those metrics tie directly to contractual commitments. For instance, after facing significant penalties, companies like &lt;a href="https://www.techtarget.com/searchunifiedcommunications/news/252470248/Slack-waters-down-cloud-SLA-after-82-million-payout" rel="noopener noreferrer"&gt;Slack&lt;/a&gt; revised their SLA terms to better align internal goals with customer promises.&lt;/p&gt;

&lt;h2&gt;
  
  
  SLA Application Performance
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6gw8kr33l4lp4lp5r73q.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6gw8kr33l4lp4lp5r73q.jpeg" alt="A person typing on a computerAI-generated content may be incorrect." width="800" height="262"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Establishing an SLA is more than just putting a few sentences in the contract. The reason we paid $1 million is that there was no SLA Management System in place. We started then by building a Service Level Management (SLM) practice that relied on 4 pillars: Administration, Monitoring, Reporting, and Compliance (AMRC).&lt;/p&gt;

&lt;h2&gt;
  
  
  The SLM process
&lt;/h2&gt;

&lt;p&gt;We sat down with business partners, customers, legal, and finance teams to create a process that would prevent costly mistakes in the future. This process, which we called the SLA lifecycle, was reviewed quarterly to ensure it remained effective and aligned with our business goals.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Risk simulations with data science:&lt;/strong&gt; One of the most critical steps in our SLM process was using our in-house data scientists to run simulations. These simulations analyzed historical data from our monitoring tools to assess the risk of breaching SLAs. The goal was to set realistic SLAs that wouldn’t be breached every day, while still meeting customer expectations.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;“What-if” scenarios:&lt;/strong&gt; We also ran multiple “what-if” scenarios to understand the relationship between availability and revenue. These scenarios helped us evaluate the impact of downtime at different hours of the day and days of the week. For example, we could see how a 10-minute outage during peak traffic hours would affect revenue compared to the same outage during off-peak times.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;The SLA desk:&lt;/strong&gt; To streamline the process, we created an online tool in 2001—essentially an “SLA desk”—that allowed our sales team to request SLA portfolios for customers. These requests were reviewed and approved by our QoS team, ensuring that every SLA was realistic, measurable, and aligned with our capabilities.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Aligning external and internal SLAs
&lt;/h2&gt;

&lt;p&gt;One of the biggest challenges we faced was the mismatch between external SLAs (what we promised customers) and internal SLAs (what we measured internally). For example, customers would ask for ad-serving uptime, while our tech team measured server availability.&lt;/p&gt;

&lt;p&gt;To solve this, we aligned our external and internal SLOs and made the internal objectives (the targets) very very high. This was a huge victory because it allowed us to rely on one set of metrics to understand our SLA risk position and drive operational excellence. Our tech group (Ops, Engineering, etc.) also became more sensitive to the notion of a business SLA and started to care a lot about not breaching them.&lt;/p&gt;

&lt;h2&gt;
  
  
  Monitoring – the key to SLA success
&lt;/h2&gt;

&lt;p&gt;For availability and performance, we relied on three synthetic products. Internally, we ran Sitescope in 17 data centers and used two external synthetic products. We wanted to have as many data points as possible from as many tools as possible. The stakes were just too high not to invest in multiple tools. This entire SLM project was not cheap to implement and run on an annual basis, but I also knew the cost of not doing it right the hard way.&lt;/p&gt;

&lt;p&gt;For monitoring, it became clear we needed to test as often as possible from as many vantage points as possible:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;If you only check your SLO endpoints once an hour, you must wait 59 minutes between checks. That gap can lead to false downtime alerts.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;You also need many data points to ensure statistical significance. Smaller datasets lower precision and power, while larger one’s help manage false positives and false negatives.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Enter Differential Performance Measurement (DPM)
&lt;/h2&gt;

&lt;p&gt;One of our biggest challenges was finding an effective way to measure the ad delivery speed and capture it in our SLAs. Clients would look at their site performance and notice spikes and they would attribute it to our system, meanwhile our performance telemetry would not show any problems. We couldn’t correlate the two charts; therefore we couldn’t come to an agreement whether it was our problem or someone else’s problem.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4orxf2j834sc8uaodyqi.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4orxf2j834sc8uaodyqi.jpeg" alt="image" width="800" height="274"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fidakqn7btek4wn09ugi6.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fidakqn7btek4wn09ugi6.jpeg" alt="image" width="800" height="268"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;To address this, we developed a methodology called &lt;a href="https://contracts.onecle.com/ask/doubleclick.svc.2004.02.27.shtml" rel="noopener noreferrer"&gt;Differential Performance Measurement&lt;/a&gt; (DPM). Our goal was to measure Doubleclick’s performance and availability with precision, and to understand how it affected our customers’ pages. We also wanted to be accountable for what we controlled, so we could avoid blame and finger-pointing.&lt;/p&gt;

&lt;p&gt;The methodology added context to the measurements. DPM introduced clarity and comparison, removing absolute performance numbers from the SLAs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Recipe for Differential Performance Measurement (example with an advert.):&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; Take two pages—one without ads and one with a single ad call.&lt;/li&gt;
&lt;/ol&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Page A = No ads&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Page B = One ad&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Make sure the pages do not contain any other third-party references (CDNs, etc.).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Make sure the page sizes (in KB) are the same.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;“Bake” – Measure response times for both pages and you get the following metrics:&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Differential Response&lt;/strong&gt; (DR) will be (Response Time of page B) minus (Response Time of page A)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Differential Response Percentage&lt;/strong&gt; (DRP) = DR / A. (e.g. If Page A is 2 seconds, and Page B is 2.1 seconds, DR is 0.1 second, and DRP is 0.1/2=0.05 or 5%)&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This approach helped eliminate noise caused by:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Internet-related issues beyond our control (e.g., fiber cuts).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Monitoring agent inconsistencies (raising the need to monitor our monitoring tools).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Other third-party dependencies.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;To visualize the impact of Differential Performance Measurement (DPM), the chart below compares response times for two scenarios&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwmlp6kwbyot84pvcu9sw.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwmlp6kwbyot84pvcu9sw.jpeg" alt="image" width="800" height="409"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Scenario 1:&lt;/strong&gt; The ad-serving company experienced performance issues, which negatively impacted the customer’s site. The vendor breached the SLA threshold between Time 4 and Time 8.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Scenario 2:&lt;/strong&gt; The website itself encountered performance problems, unrelated to the ad-serving company.&lt;/p&gt;

&lt;h2&gt;
  
  
  Reporting – Transparency and Accountability
&lt;/h2&gt;

&lt;p&gt;After the $1 million penalty, SLA management became a top priority, with visibility extending all the way to the CEO. We reported monthly on compliance and breaches, using tools like DigitalFuel to detect issues in real-time.&lt;/p&gt;

&lt;p&gt;By the end of 2001, we were tracking over 100 Operational Level Agreements (OLAs), and a Culture of Quality had emerged at DoubleClick. Everyone—from engineers to executives—was aligned around business service metrics, and no one wanted to breach an SLA.&lt;/p&gt;

&lt;h2&gt;
  
  
  Lessons learned and the road ahead
&lt;/h2&gt;

&lt;p&gt;Implementing a comprehensive SLM process at DoubleClick allowed us to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Manage hundreds of contracts&lt;/strong&gt; with up to five SLOs each.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Offer scalable SLAs&lt;/strong&gt; that could adapt to new products.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Reduce financial risks&lt;/strong&gt; by avoiding costly penalties.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Maintain our reputation&lt;/strong&gt; by providing accurate and meaningful SLAs.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Detect breaches in real-time&lt;/strong&gt;, allowing us to take proactive measures.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;One of the biggest advantages was knowing in advance when an SLA was at risk. For example, we could predict that adding four minutes of downtime would breach 12 contracts and result in $X in penalties. This insight helped our Ops team act—pausing releases or preventing any changes that could impact uptime.&lt;/p&gt;

&lt;p&gt;Some people dismiss SLAs, and in many cases, that skepticism is justified. Bad SLAs—those with unrealistic guarantees, no real penalties, or vague measurement criteria—undermine trust. I often see SLAs promising 0% packet loss, but when you ask how it’s measured, you quickly realize it’s meaningless. These kinds of SLAs give the entire concept a bad reputation.&lt;/p&gt;

&lt;p&gt;However, when done right, SLAs are essential. They align customers and vendors, reduce friction, and eliminate blame games. That said, customers need to demand useful SLAs—not just ones that sound good on paper. The goal isn’t to drive vendors out of business but to hold them accountable. If they fail to deliver, they should feel the impact.&lt;/p&gt;

&lt;h2&gt;
  
  
  The evolution of SLAs
&lt;/h2&gt;

&lt;p&gt;Back in 2001, we knew SLA management was critical, but could we have predicted how integral it would become in today’s cloud-driven world? SLAs have evolved from simple uptime guarantees to complex agreements that cover everything from latency to data residency. XLOs (Experience Level Objectives) are a thing—metrics that focus on the customer’s experience, not just the server’s performance. This shift in focus—from internal metrics to customer outcomes—is the future of performance management.&lt;/p&gt;

&lt;p&gt;Stay tuned for Part 2, where we’ll explore how businesses can align their internal metrics with what truly matters: the customer’s experience&lt;/p&gt;

&lt;h2&gt;
  
  
  Learn more
&lt;/h2&gt;

&lt;p&gt;New to SLAs, SLOs, and SLIs? &lt;a href="https://www.catchpoint.com/blog/mastering-ipm-protecting-revenue-through-sla-monitoring" rel="noopener noreferrer"&gt;Read this post&lt;/a&gt; to learn the fundamentals, best practices, and how they impact service reliability.&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;In the early days of DoubleClick, back when SaaS was still known as Application Service Provider (ASP), I was tasked with setting up the QoS (Quality of Service) Team. Our primary mission was to establish a monitoring system, but we quickly found ourselves managing Service Level Agreements (SLAs)—a task that became critical after we paid out over &lt;strong&gt;$1 million in penalties&lt;/strong&gt; for SLA violations to a single customer. The reason? Someone had signed a contract promising &lt;strong&gt;100% uptime&lt;/strong&gt;, an impossible commitment.  &lt;/p&gt;

&lt;p&gt;This is the story of how we took control of our SLAs, stopped the financial bleeding, and built a culture of quality around service metrics. Whether you’re managing SLAs today or just curious about how they work, this post will provide valuable insights into the challenges we faced, the solutions we implemented, and the lessons we learned along the way.&lt;/p&gt;

&lt;h2&gt;
  
  
  What are SLAs?
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi27ag99x6l839alc1i5e.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi27ag99x6l839alc1i5e.jpeg" alt="A clipboard with a pen and paper clipsAI-generated content may be incorrect." width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;An &lt;strong&gt;SLA (Service Level Agreement)&lt;/strong&gt; is a contractual agreement between a vendor and a customer that outlines the expected level of service. Under this legal umbrella, you’ll find &lt;strong&gt;Service Level Objectives (SLOs)&lt;/strong&gt;, which define specific metrics like uptime, speed, or transactions per second.&lt;/p&gt;

&lt;p&gt;At DoubleClick, we defined SLAs with the following principles in mind:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Attainable:&lt;/strong&gt; The goals should be realistic.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Repeatable:&lt;/strong&gt; The metrics should be consistently measurable.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Measurable:&lt;/strong&gt; The performance should be quantifiable.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Meaningful:&lt;/strong&gt; The metrics should matter to the business.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Mutually Acceptable:&lt;/strong&gt; Both parties should agree on the terms.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;SLAs benefit both the customer and the vendor. For customers, they provide &lt;strong&gt;objective grading criteria&lt;/strong&gt; and protection from poor service. For vendors, they set clear expectations and incentivize quality improvements&lt;/p&gt;

&lt;h2&gt;
  
  
  Ground zero, discovery
&lt;/h2&gt;

&lt;p&gt;When we first tackled the SLA problem, we were in crisis mode. The first step was to compile a list of all contracts, extract the SLAs and SLOs, and document the associated penalties. We stored this information in a database and began educating stakeholders—business leaders, legal teams, and executives—about the importance of SLAs.&lt;/p&gt;

&lt;p&gt;From the beginning, we focused on &lt;strong&gt;end-user experience-based SLAs&lt;/strong&gt;. This meant measuring performance from the user’s perspective, not just from the server’s perspective.&lt;/p&gt;

&lt;h2&gt;
  
  
  A universal challenge
&lt;/h2&gt;

&lt;p&gt;Over the years, I’ve seen many companies face similar issues. Not all SRE and Dev teams fully grasp the SLAs their organization has with customers—they often focus heavily on internal SLOs while overlooking how those metrics tie directly to contractual commitments. For instance, after facing significant penalties, companies like &lt;a href="https://www.techtarget.com/searchunifiedcommunications/news/252470248/Slack-waters-down-cloud-SLA-after-82-million-payout" rel="noopener noreferrer"&gt;Slack&lt;/a&gt; revised their SLA terms to better align internal goals with customer promises.&lt;/p&gt;

&lt;h2&gt;
  
  
  SLA Application Performance
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6gw8kr33l4lp4lp5r73q.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6gw8kr33l4lp4lp5r73q.jpeg" alt="A person typing on a computerAI-generated content may be incorrect." width="800" height="262"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Establishing an SLA is more than just putting a few sentences in the contract. The reason we paid $1 million is that there was no SLA Management System in place. We started then by building a Service Level Management (SLM) practice that relied on 4 pillars: Administration, Monitoring, Reporting, and Compliance (AMRC).&lt;/p&gt;

&lt;h2&gt;
  
  
  The SLM process
&lt;/h2&gt;

&lt;p&gt;We sat down with business partners, customers, legal, and finance teams to create a process that would prevent costly mistakes in the future. This process, which we called the SLA lifecycle, was reviewed quarterly to ensure it remained effective and aligned with our business goals.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Risk simulations with data science:&lt;/strong&gt; One of the most critical steps in our SLM process was using our in-house data scientists to run simulations. These simulations analyzed historical data from our monitoring tools to assess the risk of breaching SLAs. The goal was to set realistic SLAs that wouldn’t be breached every day, while still meeting customer expectations.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;“What-if” scenarios:&lt;/strong&gt; We also ran multiple “what-if” scenarios to understand the relationship between availability and revenue. These scenarios helped us evaluate the impact of downtime at different hours of the day and days of the week. For example, we could see how a 10-minute outage during peak traffic hours would affect revenue compared to the same outage during off-peak times.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;The SLA desk:&lt;/strong&gt; To streamline the process, we created an online tool in 2001—essentially an “SLA desk”—that allowed our sales team to request SLA portfolios for customers. These requests were reviewed and approved by our QoS team, ensuring that every SLA was realistic, measurable, and aligned with our capabilities.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Aligning external and internal SLAs
&lt;/h2&gt;

&lt;p&gt;One of the biggest challenges we faced was the mismatch between external SLAs (what we promised customers) and internal SLAs (what we measured internally). For example, customers would ask for ad-serving uptime, while our tech team measured server availability.&lt;/p&gt;

&lt;p&gt;To solve this, we aligned our external and internal SLOs and made the internal objectives (the targets) very very high. This was a huge victory because it allowed us to rely on one set of metrics to understand our SLA risk position and drive operational excellence. Our tech group (Ops, Engineering, etc.) also became more sensitive to the notion of a business SLA and started to care a lot about not breaching them.&lt;/p&gt;

&lt;h2&gt;
  
  
  Monitoring – the key to SLA success
&lt;/h2&gt;

&lt;p&gt;For availability and performance, we relied on three synthetic products. Internally, we ran Sitescope in 17 data centers and used two external synthetic products. We wanted to have as many data points as possible from as many tools as possible. The stakes were just too high not to invest in multiple tools. This entire SLM project was not cheap to implement and run on an annual basis, but I also knew the cost of not doing it right the hard way.&lt;/p&gt;

&lt;p&gt;For monitoring, it became clear we needed to test as often as possible from as many vantage points as possible:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;If you only check your SLO endpoints once an hour, you must wait 59 minutes between checks. That gap can lead to false downtime alerts.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;You also need many data points to ensure statistical significance. Smaller datasets lower precision and power, while larger one’s help manage false positives and false negatives.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Enter Differential Performance Measurement (DPM)
&lt;/h2&gt;

&lt;p&gt;One of our biggest challenges was finding an effective way to measure the ad delivery speed and capture it in our SLAs. Clients would look at their site performance and notice spikes and they would attribute it to our system, meanwhile our performance telemetry would not show any problems. We couldn’t correlate the two charts; therefore we couldn’t come to an agreement whether it was our problem or someone else’s problem.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4orxf2j834sc8uaodyqi.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4orxf2j834sc8uaodyqi.jpeg" alt="image" width="800" height="274"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fidakqn7btek4wn09ugi6.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fidakqn7btek4wn09ugi6.jpeg" alt="image" width="800" height="268"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;To address this, we developed a methodology called &lt;a href="https://contracts.onecle.com/ask/doubleclick.svc.2004.02.27.shtml" rel="noopener noreferrer"&gt;Differential Performance Measurement&lt;/a&gt; (DPM). Our goal was to measure Doubleclick’s performance and availability with precision, and to understand how it affected our customers’ pages. We also wanted to be accountable for what we controlled, so we could avoid blame and finger-pointing.&lt;/p&gt;

&lt;p&gt;The methodology added context to the measurements. DPM introduced clarity and comparison, removing absolute performance numbers from the SLAs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Recipe for Differential Performance Measurement (example with an advert.):&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; Take two pages—one without ads and one with a single ad call.&lt;/li&gt;
&lt;/ol&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Page A = No ads&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Page B = One ad&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Make sure the pages do not contain any other third-party references (CDNs, etc.).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Make sure the page sizes (in KB) are the same.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;“Bake” – Measure response times for both pages and you get the following metrics:&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Differential Response&lt;/strong&gt; (DR) will be (Response Time of page B) minus (Response Time of page A)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Differential Response Percentage&lt;/strong&gt; (DRP) = DR / A. (e.g. If Page A is 2 seconds, and Page B is 2.1 seconds, DR is 0.1 second, and DRP is 0.1/2=0.05 or 5%)&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This approach helped eliminate noise caused by:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Internet-related issues beyond our control (e.g., fiber cuts).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Monitoring agent inconsistencies (raising the need to monitor our monitoring tools).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Other third-party dependencies.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;To visualize the impact of Differential Performance Measurement (DPM), the chart below compares response times for two scenarios&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwmlp6kwbyot84pvcu9sw.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwmlp6kwbyot84pvcu9sw.jpeg" alt="image" width="800" height="409"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Scenario 1:&lt;/strong&gt; The ad-serving company experienced performance issues, which negatively impacted the customer’s site. The vendor breached the SLA threshold between Time 4 and Time 8.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Scenario 2:&lt;/strong&gt; The website itself encountered performance problems, unrelated to the ad-serving company.&lt;/p&gt;

&lt;h2&gt;
  
  
  Reporting – Transparency and Accountability
&lt;/h2&gt;

&lt;p&gt;After the $1 million penalty, SLA management became a top priority, with visibility extending all the way to the CEO. We reported monthly on compliance and breaches, using tools like DigitalFuel to detect issues in real-time.&lt;/p&gt;

&lt;p&gt;By the end of 2001, we were tracking over 100 Operational Level Agreements (OLAs), and a Culture of Quality had emerged at DoubleClick. Everyone—from engineers to executives—was aligned around business service metrics, and no one wanted to breach an SLA.&lt;/p&gt;

&lt;h2&gt;
  
  
  Lessons learned and the road ahead
&lt;/h2&gt;

&lt;p&gt;Implementing a comprehensive SLM process at DoubleClick allowed us to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Manage hundreds of contracts&lt;/strong&gt; with up to five SLOs each.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Offer scalable SLAs&lt;/strong&gt; that could adapt to new products.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Reduce financial risks&lt;/strong&gt; by avoiding costly penalties.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Maintain our reputation&lt;/strong&gt; by providing accurate and meaningful SLAs.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Detect breaches in real-time&lt;/strong&gt;, allowing us to take proactive measures.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;One of the biggest advantages was knowing in advance when an SLA was at risk. For example, we could predict that adding four minutes of downtime would breach 12 contracts and result in $X in penalties. This insight helped our Ops team act—pausing releases or preventing any changes that could impact uptime.&lt;/p&gt;

&lt;p&gt;Some people dismiss SLAs, and in many cases, that skepticism is justified. Bad SLAs—those with unrealistic guarantees, no real penalties, or vague measurement criteria—undermine trust. I often see SLAs promising 0% packet loss, but when you ask how it’s measured, you quickly realize it’s meaningless. These kinds of SLAs give the entire concept a bad reputation.&lt;/p&gt;

&lt;p&gt;However, when done right, SLAs are essential. They align customers and vendors, reduce friction, and eliminate blame games. That said, customers need to demand useful SLAs—not just ones that sound good on paper. The goal isn’t to drive vendors out of business but to hold them accountable. If they fail to deliver, they should feel the impact.&lt;/p&gt;

&lt;h2&gt;
  
  
  The evolution of SLAs
&lt;/h2&gt;

&lt;p&gt;Back in 2001, we knew SLA management was critical, but could we have predicted how integral it would become in today’s cloud-driven world? SLAs have evolved from simple uptime guarantees to complex agreements that cover everything from latency to data residency. XLOs (Experience Level Objectives) are a thing—metrics that focus on the customer’s experience, not just the server’s performance. This shift in focus—from internal metrics to customer outcomes—is the future of performance management.&lt;/p&gt;

&lt;p&gt;Stay tuned for Part 2, where we’ll explore how businesses can align their internal metrics with what truly matters: the customer’s experience&lt;/p&gt;

&lt;h2&gt;
  
  
  Learn more
&lt;/h2&gt;

&lt;p&gt;New to SLAs, SLOs, and SLIs? &lt;a href="https://www.catchpoint.com/blog/mastering-ipm-protecting-revenue-through-sla-monitoring" rel="noopener noreferrer"&gt;Read this post&lt;/a&gt; to learn the fundamentals, best practices, and how they impact service reliability.&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

</description>
      <category>monitoring</category>
      <category>observability</category>
      <category>sla</category>
    </item>
    <item>
      <title>When AI tools fail: How to map your AI dependencies for proactive visibility</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Mon, 11 Aug 2025 14:32:52 +0000</pubDate>
      <link>https://forem.com/catchpoint/when-ai-tools-fail-how-to-map-your-ai-dependencies-for-proactive-visibility-ia4</link>
      <guid>https://forem.com/catchpoint/when-ai-tools-fail-how-to-map-your-ai-dependencies-for-proactive-visibility-ia4</guid>
      <description>&lt;p&gt;&lt;em&gt;(This post originally appeared on the &lt;a href="https://www.catchpoint.com/blog/when-ai-tools-fail-how-to-map-your-ai-dependencies-for-proactive-visibility" rel="noopener noreferrer"&gt;Catchpoint Blog&lt;/a&gt; )&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;AI platforms have experienced several service interruptions over the past few months.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpdgzv5s3qqhvqzrt5ylp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpdgzv5s3qqhvqzrt5ylp.png" alt="A screenshot of a social media postAI-generated content may be incorrect." width="757" height="871"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;We’ve all seen the memes fly when ChatGPT, Gemini or Perplexity go down. They’re funny at first, but then reality hits: if you rely on AI tools for work or business, these outages can grind your day to a halt. And it’s not just a glitch here or there— there’s a clear pattern of AI services failing across different platforms:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;February 5-6, 2025:&lt;/strong&gt; &lt;a href="https://www.google.com/appsstatus/dashboard/incidents/8VCWK4rfLd4RjWXDsPdn" rel="noopener noreferrer"&gt;Google’s Gemini&lt;/a&gt; experienced a 23-hour disruption that affected the "Add File" and "Link File" functionalities within Gems. The outage prevented users from attaching files to their AI-driven workflows. Users had no workaround, leading to productivity loss for businesses relying on Gemini’s file-processing capabilities.  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;January 23, 2025:&lt;/strong&gt; &lt;a href="https://status.openai.com/incidents/k934n9lt39h8" rel="noopener noreferrer"&gt;ChatGPT&lt;/a&gt; and several OpenAI APIs suffered elevated error rates, with users encountering "bad gateway" errors. Businesses relying on ChatGPT for automation, customer service, and content generation were left scrambling  &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;January 23, 2025&lt;/strong&gt;: &lt;a href="https://status.perplexity.com/cm68ne5it000crf9ewzcamb8y" rel="noopener noreferrer"&gt;Perplexity’s&lt;/a&gt; API experienced a major outage, causing timeouts and disruptions for applications relying on its AI capabilities.    &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;December 26, 2024:&lt;/strong&gt; A host of &lt;a href="https://status.openai.com/incidents/6bwlxnvdncnm" rel="noopener noreferrer"&gt;OpenAI services&lt;/a&gt; (ChatGPT, Sora video creation, plus agents, realtime speech, batch, and DALL-E APIs) suffered error rates north of 90%.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;June 4, 2024:&lt;/strong&gt; On this day, &lt;a href="https://techcrunch.com/2024/06/04/ai-apocalypse-chatgpt-claude-and-perplexity-are-all-down-at-the-same-time/" rel="noopener noreferrer"&gt;multiple AI platforms&lt;/a&gt;, including OpenAI's ChatGPT, Anthropic's Claude, and Perplexity, experienced simultaneous outages. Users worldwide reported disruptions, leading to widespread discussions on social media platforms.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;We’ve officially hit a point where our dependence on AI is no longer just a possibility; it’s an absolute. When these systems fail, we’re left scrambling. The real question is: &lt;strong&gt;how do you stay ahead of the next failure?&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The revenue impact of AI outages
&lt;/h2&gt;

&lt;p&gt;The numbers are big and growing bigger: in 2025, global &lt;a href="https://hyperight.com/countdown-to-2025-why-now-is-the-time-to-invest-in-data-and-ai/#:~:text=The%20Tipping%20Point:%20Why%202025,invest%20in%20data%20and%20AI!" rel="noopener noreferrer"&gt;AI investments are set to exceed $500 billion&lt;/a&gt;. For many companies, AI apps like ChatGPT aren’t optional anymore—they’re mission-critical. &lt;a href="https://www.gartner.com/en/newsroom/press-releases/2023-05-03-gartner-poll-finds-45-percent-of-executives-say-chatgpt-has-prompted-an-increase-in-ai-investment#:~:text=70%25%20of%20Organizations%20Currently%20in%20Exploration%20Mode%20with%20Generative%20AI" rel="noopener noreferrer"&gt;Gartner reports&lt;/a&gt; that 70% of enterprises now use large language models (LLMs) for everyday tasks like automated customer service, marketing personalization, and real-time data crunching.&lt;/p&gt;

&lt;p&gt;When these AI systems go offline, it’s not just a minor inconvenience. In finance, a few hours of AI downtime could mean millions lost due to missed trades or undetected fraud. In eCommerce, chatbots and recommendation engines going dark mean abandoned shopping carts and fewer conversions, which translates to real money left on the table.&lt;/p&gt;

&lt;p&gt;But the damage doesn’t stop at lost revenue. Companies increasingly rely on AI-powered automation to streamline workflows, meaning that outages force employees to revert to manual processes, significantly slowing down productivity. This is particularly evident in customer support, where AI chatbots handle vast volumes of inquiries. If an outage forces companies to fall back on human agents, call center queues expand, increasing response times and leading to diminished customer satisfaction.&lt;/p&gt;

&lt;p&gt;If you’re concerned about the impact of AI outages on your business, now is the time to evaluate your AI dependencies and invest in tools that can help you stay ahead of disruptions.&lt;/p&gt;

&lt;h2&gt;
  
  
  The need for visibility: Mapping your AI dependencies  
&lt;/h2&gt;

&lt;p&gt;Even with the best monitoring strategies in place, AI outages present a unique challenge: You may know something is broken, but not necessarily &lt;strong&gt;where or why&lt;/strong&gt;. To accurately pinpoint issues, you need tools that enable you to get actionable insights into your AI dependencies—whether they originate in the application layer or the underlying &lt;a href="https://www.catchpoint.com/glossary/internet-stack" rel="noopener noreferrer"&gt;Internet Stack&lt;/a&gt;.  &lt;/p&gt;

&lt;h2&gt;
  
  
  eCommerce AI dependencies: A case study
&lt;/h2&gt;

&lt;p&gt;Consider an eCommerce company relying on an AI-powered chatbot for customer support. It relies on several key components to deliver a seamless shopping experience:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Front-end CDN:&lt;/strong&gt; Ensures fast content delivery to users.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Distributed Hyperscaler:&lt;/strong&gt; Acts as the origin server for dynamic content.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Search and Seller APIs&lt;/strong&gt;: Retrieve relevant product data for users.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Chatbot Powered by OpenAI API:&lt;/strong&gt; Handles customer inquiries and provides real-time support.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The chatbot is a critical part of the customer support workflow. When a shopper interacts with the chatbot, their request is forwarded to an external API, which then interacts with the OpenAI API to generate a response. This means the chatbot’s functionality is entirely dependent on the OpenAI API.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frsnqbnn3r4rr7cga2lou.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frsnqbnn3r4rr7cga2lou.png" alt="A white rectangular sign with black textAI-generated content may be incorrect." width="800" height="105"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Flow diagram depicting the interaction between a user, an external API, and OpenAI's API in an e-commerce chatbot system&lt;/p&gt;

&lt;p&gt;If the OpenAI API experiences an outage, the chatbot fails, leaving customers without support. This not only frustrates users but can also lead to lost sales and damaged customer relationships.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to map AI dependencies and stay ahead of outages
&lt;/h2&gt;

&lt;p&gt;In the eCommerce example above, the chatbot’s dependency on the OpenAI API highlights the importance of mapping AI dependencies. When outages occur, knowing exactly where the failure lies can mean the difference between minutes of downtime and hours of lost revenue. By mapping your AI dependencies, you can quickly identify the root cause of outages, reducing downtime and minimizing revenue loss. Here's how:&lt;/p&gt;

&lt;h3&gt;
  
  
  #1 Visualize your AI dependencies
&lt;/h3&gt;

&lt;p&gt;Start by creating a map of all the services and APIs your AI tools rely on. For example, if your chatbot depends on OpenAI’s API, you need to include it in your dependency map. Tools like &lt;a href="https://www.catchpoint.com/internet-stack-map" rel="noopener noreferrer"&gt;Internet Stack Map&lt;/a&gt; can help you visualize these connections, making it easier to pinpoint where failures occur when an outage happens.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdcl983ve17mksngw5nbp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdcl983ve17mksngw5nbp.png" width="800" height="364"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Internet Stack Map view&lt;/p&gt;

&lt;p&gt;In the example above, the Internet Stack Map view of our eCommerce case study shows all other services are working as expected, except for the OpenAI API (highlighted in red), which impacts chatbot interactions.&lt;/p&gt;

&lt;h3&gt;
  
  
  #2 Customize your workflow
&lt;/h3&gt;

&lt;p&gt;Every AI system is unique, so your dependency map should reflect your specific architecture. Identify key components like CDNs, DNS providers, and origin servers, and ensure they’re included in your map. This customization ensures you’re prepared to troubleshoot issues that are specific to your setup.&lt;/p&gt;

&lt;h3&gt;
  
  
  #3 Correlate data for faster insights
&lt;/h3&gt;

&lt;p&gt;Use monitoring tools that combine &lt;a href="https://www.catchpoint.com/guide-to-synthetic-monitoring/rum-vs-synthetic-monitoring" rel="noopener noreferrer"&gt;synthetic testing&lt;/a&gt; with &lt;a href="https://www.catchpoint.com/internet-sonar" rel="noopener noreferrer"&gt;real-time outage data&lt;/a&gt;. By correlating this data, you can quickly determine whether an issue is with your AI provider (e.g., OpenAI) or your own infrastructure. This reduces the time spent diagnosing problems, helps you avoid unnecessary war rooms and saves you money.&lt;/p&gt;

&lt;h2&gt;
  
  
  Faster resolution, fewer disruptions
&lt;/h2&gt;

&lt;p&gt;AI outages remind us how vulnerable we are in this interconnected world. When these systems fail, every minute counts—particularly if you’re losing revenue or driving customers away. That’s why Internet Stack Map, recently updated with a &lt;a href="https://www.catchpoint.com/press-releases/catchpoint-helps-it-operations-teams-minimize-outages-with-revolutionary-visibility-across-the-complete-internet-stack-into-the-application-stack" rel="noopener noreferrer"&gt;groundbreaking user interface&lt;/a&gt;, is a game-changer for incident response. It offers immediate clarity about what broke and where, shrinking your Mean Time to Identify (MTTI) and Mean Time to Resolve (MTTR).  &lt;/p&gt;

&lt;p&gt;See how Internet Stack Map can help you stay ahead of disruptions—&lt;a href="https://www.catchpoint.com/learn-more" rel="noopener noreferrer"&gt;schedule a demo today&lt;/a&gt;.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>aiops</category>
      <category>monitoring</category>
      <category>observability</category>
    </item>
    <item>
      <title>Cloud Monitoring's Blind Spot: The User Perspective</title>
      <dc:creator>Leon Adato</dc:creator>
      <pubDate>Tue, 05 Aug 2025 17:52:17 +0000</pubDate>
      <link>https://forem.com/catchpoint/cloud-monitorings-blind-spot-the-user-perspective-2gnl</link>
      <guid>https://forem.com/catchpoint/cloud-monitorings-blind-spot-the-user-perspective-2gnl</guid>
      <description>&lt;p&gt;The evolution of internet-centric application delivery has worsened IT's visibility gaps into what impacts an end user's experience. This problem is exacerbated when these gaps lead to negative business consequences, such as loss of revenue or lower Net Promoter Scores (NPS). The need to address this worsening visibility gap problem is reinforced by Gartner’s recent publication of its first &lt;a href="https://www.catchpoint.com/asset/2024-gartner-magic-quadrant-for-digital-experience-monitoring" rel="noopener noreferrer"&gt;Magic Quadrant for Digital Experience Monitoring (DEM)&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;A good way to understand what visibility really should look like is through the perspective of first- versus last-mile monitoring.  &lt;/p&gt;

&lt;h2&gt;
  
  
  First- vs. Last-Mile Monitoring: Where You Monitor Matters
&lt;/h2&gt;

&lt;p&gt;The first mile represents cloud networks and platforms like AWS, Azure, Google Cloud and even “Joe’s network closet.” These environments are stable, well-optimized and critical for hosting applications. Monitoring from the first mile focuses on ensuring that the core infrastructure and code of your applications are performing as expected.&lt;/p&gt;

&lt;p&gt;The last mile, however, is where real users connect to your applications; it’s where experiences occur. This includes backbone networks (e.g., regional ISPs like BT, AT&amp;amp;T and Comcast), last-mile providers (fiber or wireless like Verizon, Sky or T-Mobile) and wireless connections. Monitoring the last mile reveals the real-world challenges users face, such as latency spikes, packet loss and internet service provider (ISP)-specific issues that are invisible from the first mile.&lt;/p&gt;

&lt;p&gt;Think of it like the Domino’s “Paving for Pizza” ad campaign — it’s not just about ensuring the pizza is perfect when it leaves the store (first mile); it’s about fixing the potholes in the roads so the pizza arrives intact at the customer’s door (last mile). The same principle applies to digital experiences: monitoring the first mile isn’t enough if the last mile isn’t delivering. Monitoring from the last mile paints the clearest picture of performance from your users’ perspective.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why First-Mile Monitoring Alone Falls Short
&lt;/h2&gt;

&lt;p&gt;Your applications are most likely hosted in a cloud provider’s data center, often within the same border gateway protocol (BGP) autonomous system (AS) as your monitoring tools. This means that monitoring so close to the source does little more than verify the availability of your infrastructure. In other words, this type of "inside of the house" setup offers limited visibility into real-world issues.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; &lt;strong&gt;User perspective lost:&lt;/strong&gt; Internet Performance Monitoring (IPM) monitors health from the user’s perspective, which cloud-only monitoring can’t do.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Observability risks:&lt;/strong&gt; When the first mile goes down — a more frequent occurrence than many realize — your observability strategy goes with it. This isn’t just a theoretical risk; it’s something we’ve seen play out time and time again in real-world outages, such as the &lt;a href="https://www.catchpoint.com/blog/dont-get-caught-in-the-dark-lessons-from-a-lumen-aws-micro-outage" rel="noopener noreferrer"&gt;Lumen and AWS micro-outage&lt;/a&gt; in August 2024. In this incident, critical systems were disrupted, rippling across interconnected ecosystems and catching businesses off guard.&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Rethinking Observability: Availability and Reachability
&lt;/h2&gt;

&lt;p&gt;When it comes to delivering a flawless digital experience, observability relies on four key pillars: &lt;strong&gt;availability, reachability, performance, and reliability&lt;/strong&gt;. Each plays a critical role in understanding how your applications are performing and how users experience them.&lt;/p&gt;

&lt;p&gt;I’ll focus on the first two: availability and reachability. Availability is about whether your application is up and running. Reachability, on the other hand, measures whether users can actually connect to your application, factoring in network latency, packet loss and the number of hops between them and your servers.&lt;/p&gt;

&lt;p&gt;I’ll illustrate the difference between monitoring from the cloud versus end-user networks, and how what looks perfect in the cloud often falls apart in the wild.&lt;/p&gt;

&lt;h2&gt;
  
  
  Visualizing the Difference: Availability Across Network Types
&lt;/h2&gt;

&lt;p&gt;Cloud monitoring data often paints an overly rosy picture. As the chart below shows, monitoring from the cloud (green line) reports near-perfect availability, consistently hovering around 99.99%. But this data tells only part of the story — it reflects the controlled environment of cloud infrastructure, not the real-world experience of users.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fecwy8wdb1qe3zjyod0eq.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fecwy8wdb1qe3zjyod0eq.jpeg" width="800" height="627"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint dashboard showing network availability trends across Backbone, Cloud, Last Mile, and Wireless networks over a one-week period&lt;/p&gt;

&lt;p&gt;Now, compare this to the backbone (the blue line), last mile (the red line) and wireless (the purple line) data. These fluctuations highlight the everyday challenges users face, from regional ISP disruptions to last-mile instability. The takeaway? While cloud monitoring data might make dashboards look good, it doesn’t account for the realities of real-world networks where your users connect. To truly understand availability, you need to monitor across all these network types.&lt;/p&gt;

&lt;h2&gt;
  
  
  Cloud vs. End-User Network Maps
&lt;/h2&gt;

&lt;p&gt;Here is another example. The top map shows monitoring results from the cloud, while the bottom map shows end-user networks. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffad33gehgsljytg1vd0p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffad33gehgsljytg1vd0p.png" alt="A screenshot of a computer screenDescription automatically generated" width="800" height="896"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint dashboard comparing cloud vs. end-user network performance&lt;/p&gt;

&lt;p&gt;The cloud shows all green, indicating near-perfect first-mile performance. The bottom map shows the reality from the end-user perspective; the red and yellow markers represent performance issues that are not visible to cloud-only monitoring.&lt;/p&gt;

&lt;p&gt;This disparity underscores the critical need for monitoring where your users actually connect. While the cloud may look pristine, end-user networks tell a very different story.&lt;/p&gt;

&lt;p&gt;Now why is this?&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fahtby98zslswaagz9y43.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fahtby98zslswaagz9y43.png" width="800" height="253"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Network path visualization showing traffic flows from multiple ISPs to an Amazon destination network, highlighting packet loss and performance metrics&lt;/p&gt;

&lt;p&gt;The above image shows that the path it takes for a user to get to the cloud from an external ISP is more volatile than a cloud ISP (as shown in the image below).. This is due to the numerous BGP autonomous systems and hops that exist between a cloud-hosted application and the user. Each AS network represents a different administrative domain. As the traffic traverses these domains, it passes through multiple network hops. These hops can include diverse routing policies, peering agreements and congestion points. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxh2fhf0ns1bt8rpmbcxs.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxh2fhf0ns1bt8rpmbcxs.png" width="800" height="229"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Network path visualization showing traffic from multiple AWS Cloud locations to an Amazon destination network, showing minimal packet loss&lt;/p&gt;

&lt;p&gt;Cloud-based monitoring lacks insight into these intermediate hops, particularly across transit providers and peering exchanges, resulting in a fragmented view of network performance and true user experience. &lt;/p&gt;

&lt;p&gt;In contrast, backbone monitoring provides a more comprehensive perspective by capturing data closer to the core of the internet, offering visibility into the paths your end user traffic takes and potential bottlenecks along the way.&lt;/p&gt;

&lt;h2&gt;
  
  
  Average Response Times: Cloud vs. Backbone ISPs
&lt;/h2&gt;

&lt;p&gt;It’s one thing to know whether your application is up and running, but what about the quality of the connection? The chart below compares response metrics between backbone networks and cloud networks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8xd2625c3fvnfzd7lmia.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8xd2625c3fvnfzd7lmia.png" alt="A screenshot of a computerDescription automatically generated" width="800" height="616"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint dashboard comparing response times between backbone and cloud networks&lt;/p&gt;

&lt;p&gt;On the left, monitoring from backbone networks reveals significant variability in key metrics like &lt;strong&gt;load time&lt;/strong&gt; and &lt;strong&gt;wait time&lt;/strong&gt;. The spikes represent the challenges users face when traversing real-world networks. Compare that to the cloud on the right, where everything looks stable, smooth and controlled. But most users aren’t connecting from the cloud. Without monitoring from backbone and last-mile networks, you’re only seeing part of the story.&lt;/p&gt;

&lt;p&gt;Here’s another example of how cloud monitoring data might make everything look perfect when the reality is far from it. The chart below, showing monitoring from AWS, reports a near-instant response time of 44.79 milliseconds. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6eclyg1707of9786gto3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6eclyg1707of9786gto3.png" alt="A screenshot of a computerDescription automatically generated" width="800" height="382"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint dashboard comparing cloud vs. backbone ISP response times&lt;/p&gt;

&lt;p&gt;But what happens when you shift the perspective to backbone ISPs? In this CenturyLink example, response times skyrocket to 730.67 milliseconds.This kind of variability isn’t an outlier — it’s the reality users face every day when connecting to your application through different networks. And unless you’re monitoring from these networks, you’re missing the full picture of your application’s reachability.&lt;/p&gt;

&lt;h2&gt;
  
  
  Putting it All Together
&lt;/h2&gt;

&lt;p&gt;The data in these charts tell a clear story. It shows what first-mile monitoring alone cannot: the variability, instability and challenges users face every day on backbone, last mile and wireless networks.The takeaway? To truly understand how your applications are performing, you need to monitor beyond the cloud. Backbone, last mile and wireless networks aren’t just part of the picture — they are the picture. The ability to monitor the entire Internet Stack, including those “eyeball” networks where your users actually connect, is what sets &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Catchpoint Internet Performance Monitoring (IPM)&lt;/a&gt; apart.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Falhmgcv65ka2ccc9x1sw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Falhmgcv65ka2ccc9x1sw.png" width="800" height="449"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A visual representation of The Internet Stack&lt;/p&gt;

&lt;p&gt;To learn more about how Catchpoint IPM can help you achieve Internet Resilience, &lt;a href="https://www.catchpoint.com/guided-product-tour" rel="noopener noreferrer"&gt;request a demo&lt;/a&gt; or &lt;a href="https://www.catchpoint.com/learn-more" rel="noopener noreferrer"&gt;schedule a chat&lt;/a&gt; with our solution engineers.&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;The evolution of internet-centric application delivery has worsened IT's visibility gaps into what impacts an end user's experience. This problem is exacerbated when these gaps lead to negative business consequences, such as loss of revenue or lower Net Promoter Scores (NPS). The need to address this worsening visibility gap problem is reinforced by Gartner’s recent publication of its first &lt;a href="https://www.catchpoint.com/asset/2024-gartner-magic-quadrant-for-digital-experience-monitoring" rel="noopener noreferrer"&gt;Magic Quadrant for Digital Experience Monitoring (DEM)&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;A good way to understand what visibility really should look like is through the perspective of first- versus last-mile monitoring.  &lt;/p&gt;

&lt;h2&gt;
  
  
  First- vs. Last-Mile Monitoring: Where You Monitor Matters
&lt;/h2&gt;

&lt;p&gt;The first mile represents cloud networks and platforms like AWS, Azure, Google Cloud and even “Joe’s network closet.” These environments are stable, well-optimized and critical for hosting applications. Monitoring from the first mile focuses on ensuring that the core infrastructure and code of your applications are performing as expected.&lt;/p&gt;

&lt;p&gt;The last mile, however, is where real users connect to your applications; it’s where experiences occur. This includes backbone networks (e.g., regional ISPs like BT, AT&amp;amp;T and Comcast), last-mile providers (fiber or wireless like Verizon, Sky or T-Mobile) and wireless connections. Monitoring the last mile reveals the real-world challenges users face, such as latency spikes, packet loss and internet service provider (ISP)-specific issues that are invisible from the first mile.&lt;/p&gt;

&lt;p&gt;Think of it like the Domino’s “Paving for Pizza” ad campaign — it’s not just about ensuring the pizza is perfect when it leaves the store (first mile); it’s about fixing the potholes in the roads so the pizza arrives intact at the customer’s door (last mile). The same principle applies to digital experiences: monitoring the first mile isn’t enough if the last mile isn’t delivering. Monitoring from the last mile paints the clearest picture of performance from your users’ perspective.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why First-Mile Monitoring Alone Falls Short
&lt;/h2&gt;

&lt;p&gt;Your applications are most likely hosted in a cloud provider’s data center, often within the same border gateway protocol (BGP) autonomous system (AS) as your monitoring tools. This means that monitoring so close to the source does little more than verify the availability of your infrastructure. In other words, this type of "inside of the house" setup offers limited visibility into real-world issues.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; &lt;strong&gt;User perspective lost:&lt;/strong&gt; Internet Performance Monitoring (IPM) monitors health from the user’s perspective, which cloud-only monitoring can’t do.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Observability risks:&lt;/strong&gt; When the first mile goes down — a more frequent occurrence than many realize — your observability strategy goes with it. This isn’t just a theoretical risk; it’s something we’ve seen play out time and time again in real-world outages, such as the &lt;a href="https://www.catchpoint.com/blog/dont-get-caught-in-the-dark-lessons-from-a-lumen-aws-micro-outage" rel="noopener noreferrer"&gt;Lumen and AWS micro-outage&lt;/a&gt; in August 2024. In this incident, critical systems were disrupted, rippling across interconnected ecosystems and catching businesses off guard.&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Rethinking Observability: Availability and Reachability
&lt;/h2&gt;

&lt;p&gt;When it comes to delivering a flawless digital experience, observability relies on four key pillars: &lt;strong&gt;availability, reachability, performance, and reliability&lt;/strong&gt;. Each plays a critical role in understanding how your applications are performing and how users experience them.&lt;/p&gt;

&lt;p&gt;I’ll focus on the first two: availability and reachability. Availability is about whether your application is up and running. Reachability, on the other hand, measures whether users can actually connect to your application, factoring in network latency, packet loss and the number of hops between them and your servers.&lt;/p&gt;

&lt;p&gt;I’ll illustrate the difference between monitoring from the cloud versus end-user networks, and how what looks perfect in the cloud often falls apart in the wild.&lt;/p&gt;

&lt;h2&gt;
  
  
  Visualizing the Difference: Availability Across Network Types
&lt;/h2&gt;

&lt;p&gt;Cloud monitoring data often paints an overly rosy picture. As the chart below shows, monitoring from the cloud (green line) reports near-perfect availability, consistently hovering around 99.99%. But this data tells only part of the story — it reflects the controlled environment of cloud infrastructure, not the real-world experience of users.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fecwy8wdb1qe3zjyod0eq.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fecwy8wdb1qe3zjyod0eq.jpeg" width="800" height="627"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint dashboard showing network availability trends across Backbone, Cloud, Last Mile, and Wireless networks over a one-week period&lt;/p&gt;

&lt;p&gt;Now, compare this to the backbone (the blue line), last mile (the red line) and wireless (the purple line) data. These fluctuations highlight the everyday challenges users face, from regional ISP disruptions to last-mile instability. The takeaway? While cloud monitoring data might make dashboards look good, it doesn’t account for the realities of real-world networks where your users connect. To truly understand availability, you need to monitor across all these network types.&lt;/p&gt;

&lt;h2&gt;
  
  
  Cloud vs. End-User Network Maps
&lt;/h2&gt;

&lt;p&gt;Here is another example. The top map shows monitoring results from the cloud, while the bottom map shows end-user networks. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffad33gehgsljytg1vd0p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffad33gehgsljytg1vd0p.png" alt="A screenshot of a computer screenDescription automatically generated" width="800" height="896"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint dashboard comparing cloud vs. end-user network performance&lt;/p&gt;

&lt;p&gt;The cloud shows all green, indicating near-perfect first-mile performance. The bottom map shows the reality from the end-user perspective; the red and yellow markers represent performance issues that are not visible to cloud-only monitoring.&lt;/p&gt;

&lt;p&gt;This disparity underscores the critical need for monitoring where your users actually connect. While the cloud may look pristine, end-user networks tell a very different story.&lt;/p&gt;

&lt;p&gt;Now why is this?&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fahtby98zslswaagz9y43.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fahtby98zslswaagz9y43.png" width="800" height="253"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Network path visualization showing traffic flows from multiple ISPs to an Amazon destination network, highlighting packet loss and performance metrics&lt;/p&gt;

&lt;p&gt;The above image shows that the path it takes for a user to get to the cloud from an external ISP is more volatile than a cloud ISP (as shown in the image below).. This is due to the numerous BGP autonomous systems and hops that exist between a cloud-hosted application and the user. Each AS network represents a different administrative domain. As the traffic traverses these domains, it passes through multiple network hops. These hops can include diverse routing policies, peering agreements and congestion points. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxh2fhf0ns1bt8rpmbcxs.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxh2fhf0ns1bt8rpmbcxs.png" width="800" height="229"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Network path visualization showing traffic from multiple AWS Cloud locations to an Amazon destination network, showing minimal packet loss&lt;/p&gt;

&lt;p&gt;Cloud-based monitoring lacks insight into these intermediate hops, particularly across transit providers and peering exchanges, resulting in a fragmented view of network performance and true user experience. &lt;/p&gt;

&lt;p&gt;In contrast, backbone monitoring provides a more comprehensive perspective by capturing data closer to the core of the internet, offering visibility into the paths your end user traffic takes and potential bottlenecks along the way.&lt;/p&gt;

&lt;h2&gt;
  
  
  Average Response Times: Cloud vs. Backbone ISPs
&lt;/h2&gt;

&lt;p&gt;It’s one thing to know whether your application is up and running, but what about the quality of the connection? The chart below compares response metrics between backbone networks and cloud networks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8xd2625c3fvnfzd7lmia.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8xd2625c3fvnfzd7lmia.png" alt="A screenshot of a computerDescription automatically generated" width="800" height="616"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint dashboard comparing response times between backbone and cloud networks&lt;/p&gt;

&lt;p&gt;On the left, monitoring from backbone networks reveals significant variability in key metrics like &lt;strong&gt;load time&lt;/strong&gt; and &lt;strong&gt;wait time&lt;/strong&gt;. The spikes represent the challenges users face when traversing real-world networks. Compare that to the cloud on the right, where everything looks stable, smooth and controlled. But most users aren’t connecting from the cloud. Without monitoring from backbone and last-mile networks, you’re only seeing part of the story.&lt;/p&gt;

&lt;p&gt;Here’s another example of how cloud monitoring data might make everything look perfect when the reality is far from it. The chart below, showing monitoring from AWS, reports a near-instant response time of 44.79 milliseconds. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6eclyg1707of9786gto3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6eclyg1707of9786gto3.png" alt="A screenshot of a computerDescription automatically generated" width="800" height="382"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Catchpoint dashboard comparing cloud vs. backbone ISP response times&lt;/p&gt;

&lt;p&gt;But what happens when you shift the perspective to backbone ISPs? In this CenturyLink example, response times skyrocket to 730.67 milliseconds.This kind of variability isn’t an outlier — it’s the reality users face every day when connecting to your application through different networks. And unless you’re monitoring from these networks, you’re missing the full picture of your application’s reachability.&lt;/p&gt;

&lt;h2&gt;
  
  
  Putting it All Together
&lt;/h2&gt;

&lt;p&gt;The data in these charts tell a clear story. It shows what first-mile monitoring alone cannot: the variability, instability and challenges users face every day on backbone, last mile and wireless networks.The takeaway? To truly understand how your applications are performing, you need to monitor beyond the cloud. Backbone, last mile and wireless networks aren’t just part of the picture — they are the picture. The ability to monitor the entire Internet Stack, including those “eyeball” networks where your users actually connect, is what sets &lt;a href="https://www.catchpoint.com/internet-performance-monitoring" rel="noopener noreferrer"&gt;Catchpoint Internet Performance Monitoring (IPM)&lt;/a&gt; apart.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Falhmgcv65ka2ccc9x1sw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Falhmgcv65ka2ccc9x1sw.png" width="800" height="449"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A visual representation of The Internet Stack&lt;/p&gt;

&lt;p&gt;To learn more about how Catchpoint IPM can help you achieve Internet Resilience, &lt;a href="https://www.catchpoint.com/guided-product-tour" rel="noopener noreferrer"&gt;request a demo&lt;/a&gt; or &lt;a href="https://www.catchpoint.com/learn-more" rel="noopener noreferrer"&gt;schedule a chat&lt;/a&gt; with our solution engineers.&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;p&gt;This is some text inside of a div block.&lt;/p&gt;

</description>
      <category>monitoring</category>
      <category>observability</category>
      <category>cloud</category>
      <category>performance</category>
    </item>
  </channel>
</rss>
