<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Scott McMahan</title>
    <description>The latest articles on Forem by Scott McMahan (@scott_mcmahan_d085ae6e508).</description>
    <link>https://forem.com/scott_mcmahan_d085ae6e508</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3762553%2Fb569e409-13f5-4f48-ae60-7caf04d6afba.png</url>
      <title>Forem: Scott McMahan</title>
      <link>https://forem.com/scott_mcmahan_d085ae6e508</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/scott_mcmahan_d085ae6e508"/>
    <language>en</language>
    <item>
      <title>Your Teams Are Using AI in Silos and It's Quietly Killing Your Productivity</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Fri, 17 Apr 2026 13:30:28 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/your-teams-are-using-ai-in-silos-and-its-quietly-killing-your-productivity-9j1</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/your-teams-are-using-ai-in-silos-and-its-quietly-killing-your-productivity-9j1</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft6p8vupcsr62ydqmm4h5.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft6p8vupcsr62ydqmm4h5.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The Problem Nobody Planned For
&lt;/h2&gt;

&lt;p&gt;Most organizations have rolled out AI tools by now. The problem is that engineering adopted one set of tools, marketing adopted another, and finance is doing something else entirely. Nobody planned for how these tools would work together across departments, and that gap is costing teams more than they realize.&lt;/p&gt;

&lt;h2&gt;
  
  
  What an AI Collaboration Model Actually Is
&lt;/h2&gt;

&lt;p&gt;An AI collaboration model is essentially a blueprint for how humans and AI divide tasks, share information, and make decisions across the whole organization. It sounds formal, but it does not have to be complex. The goal is simply to make sure AI is working with your teams instead of deepening the silos that already exist.&lt;/p&gt;

&lt;h2&gt;
  
  
  What the Research Says
&lt;/h2&gt;

&lt;p&gt;The research backs this up pretty clearly. McKinsey found that AI-enabled organizations see gains in both productivity and decision quality across functions, but only when the integration is intentional. Dropping a tool into a workflow and hoping for the best does not cut it.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Three Structures Worth Knowing
&lt;/h2&gt;

&lt;p&gt;There are three common structures: centralized, where one dedicated AI team manages everything; decentralized, where each department runs its own tools independently; and hybrid, which blends both. Davenport and Mittal found that the hybrid model tends to perform best in larger organizations because it balances local flexibility with org-wide consistency.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Part Most Teams Overlook
&lt;/h2&gt;

&lt;p&gt;Adoption stalls when employees feel excluded from the process. Short, role-specific training sessions beat long generic ones every time, and involving teams in selecting tools makes a real difference in how quickly things stick.&lt;/p&gt;

&lt;h2&gt;
  
  
  Read the Full Breakdown
&lt;/h2&gt;

&lt;p&gt;If you want to go deeper on how to actually structure this inside your org, I covered everything here: &lt;a href="https://aitransformer.online/ai-collaboration-model/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-collaboration-model/&lt;/a&gt;&lt;br&gt;
How is your team currently handling AI across departments? Would love to hear what's working and what isn't.&lt;/p&gt;

&lt;h1&gt;
  
  
  ai #productivity #teamwork #softwaredevelopment #devops
&lt;/h1&gt;

</description>
      <category>ai</category>
      <category>productivity</category>
      <category>teamwork</category>
      <category>softwaredevelopment</category>
    </item>
    <item>
      <title>AI Cloud Security Is Broken. Here Is How to Fix It.</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Thu, 16 Apr 2026 13:16:41 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/ai-cloud-security-is-broken-here-is-how-to-fix-it-fgo</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/ai-cloud-security-is-broken-here-is-how-to-fix-it-fgo</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fopvalczdp7abi64rboml.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fopvalczdp7abi64rboml.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;br&gt;
If you are shipping AI workloads to production, your security strategy is probably not keeping pace. Only 18% of teams can fix vulnerabilities as fast as they release code. That means for most engineering teams, the gap between what is deployed and what is secured keeps growing with every sprint.&lt;/p&gt;

&lt;p&gt;Last year, 99% of organizations with AI systems experienced an attack on them. Over a third were breached. These are not edge cases. This is the baseline reality for teams building and deploying AI in the cloud right now.&lt;/p&gt;

&lt;p&gt;The attack surface is growing for a few reasons that are worth understanding. Most organizations are running across multiple cloud providers, which adds complexity that security tooling was not originally designed to handle. AI workloads are being pushed into production environments that were built for traditional software, without the access controls and monitoring to match. And identity management, which should be the foundation of any cloud security posture, is still being neglected at scale. Orphaned accounts, unrotated credentials, and overpermissioned service roles are some of the most common entry points attackers are exploiting today.&lt;/p&gt;

&lt;p&gt;The encouraging part is that AI is also changing what defenders can do. Modern security platforms can monitor behavioral patterns across your entire infrastructure in real time, flag anomalies automatically, and respond to threats in seconds rather than hours. Organizations using these tools are cutting response times by up to 30%. That kind of speed matters when attackers on the other side are also running automated tooling.&lt;/p&gt;

&lt;p&gt;A solid strategy here is not about ripping everything out and starting over. It starts with locking down identity and access management, adopting a Zero Trust model so that nothing inside your network is trusted by default, and building automated detection workflows that escalate to humans for the decisions that actually require judgment.&lt;/p&gt;

&lt;p&gt;I put together a thorough breakdown of what this looks like end to end, including how to approach it if you are dealing with legacy systems or a small security team.&lt;/p&gt;

&lt;p&gt;Read it here: &lt;a href="https://aitransformer.online/ai-cloud-security-strategy/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-cloud-security-strategy/&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  security #cloud #ai #devops #zerotrust
&lt;/h1&gt;

</description>
      <category>security</category>
      <category>cloud</category>
      <category>ai</category>
      <category>devops</category>
    </item>
    <item>
      <title>AI Is Coming for Your CI Pipeline. That's a Good Thing.</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Wed, 15 Apr 2026 14:33:30 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/ai-is-coming-for-your-ci-pipeline-thats-a-good-thing-4ga</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/ai-is-coming-for-your-ci-pipeline-thats-a-good-thing-4ga</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkqah2x9c1di76248uo28.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkqah2x9c1di76248uo28.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;br&gt;
Your CI pipeline is still playing catch-up. AI is about to change that.&lt;br&gt;
73% of development teams are not using AI in their CI/CD workflows yet. That number comes from a 2025 JetBrains survey, and it is surprising given how loudly AI dominates every other corner of the software engineering conversation right now.&lt;br&gt;
The gap has practical roots. Cost is a real barrier. So is uncertainty about value. Security concerns around introducing AI into build and deployment pipelines are legitimate and worth taking seriously. But the direction the industry is moving is unmistakable.&lt;/p&gt;

&lt;p&gt;AI CI automation shifts pipelines from reactive to proactive. Predictive analytics flags failure scenarios before a build even breaks. Smarter test selection cuts run times by identifying which tests actually matter for a given code change. Self-healing workflows resolve common failures automatically without waiting for a developer to intervene. Security checks move earlier into the pipeline where catching a vulnerability costs a fraction of what it costs downstream.&lt;/p&gt;

&lt;p&gt;The CI/CD tool market sits at around $35 billion today and is projected to hit $94 billion by 2035. AI is one of the primary drivers of that growth.&lt;/p&gt;

&lt;p&gt;Getting started does not require a complete infrastructure overhaul. A focused experiment with AI-assisted test selection or predictive build analytics can deliver meaningful results quickly. From there teams can expand their use of AI with confidence at each step.&lt;/p&gt;

&lt;p&gt;The engineers building this knowledge now will have a real advantage as the tools continue to mature.&lt;/p&gt;

&lt;p&gt;Full breakdown here: &lt;a href="https://aitransformer.online/ai-ci-automation/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-ci-automation/&lt;/a&gt;&lt;br&gt;
Tags: devops, cicd, ai, automation, programming&lt;/p&gt;

</description>
      <category>devops</category>
      <category>ai</category>
      <category>cicd</category>
      <category>automation</category>
    </item>
    <item>
      <title>AI Is Writing Your Code Docs, Blog Posts, and Technical Content. What Happens When It Gets It Wrong?</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Tue, 14 Apr 2026 14:46:19 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/ai-is-writing-your-code-docs-blog-posts-and-technical-content-what-happens-when-it-gets-it-wrong-13nj</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/ai-is-writing-your-code-docs-blog-posts-and-technical-content-what-happens-when-it-gets-it-wrong-13nj</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frsatq9xjhoubnz9iqz5h.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frsatq9xjhoubnz9iqz5h.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;br&gt;
Generative AI has changed the content game for dev teams. Faster documentation, quicker blog turnaround, automated release notes. The efficiency gains are real.&lt;br&gt;
But so are the risks.&lt;/p&gt;

&lt;p&gt;Hallucinated API references. Outdated code examples presented as current. Biased outputs baked into training data that nobody ever audited. These are not edge cases. They are happening right now in organizations that moved fast without building any governance around their AI content workflows.&lt;/p&gt;

&lt;p&gt;AI content governance is the framework that keeps your team out of that situation. Clear policies, human review at key stages, regular audits, and a plan for when things go wrong. It is not overhead. It is how responsible engineering teams operate at scale.&lt;/p&gt;

&lt;p&gt;We just published a full breakdown covering what AI content governance actually is, why the regulatory landscape is shifting fast, and how to build a practical framework without grinding your workflow to a halt.&lt;/p&gt;

&lt;p&gt;Worth a read if AI is touching any part of your content pipeline.&lt;br&gt;
👉 &lt;a href="https://aitransformer.online/ai-content-governance/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-content-governance/&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>governance</category>
      <category>devops</category>
      <category>programming</category>
    </item>
    <item>
      <title>Why AI Time Series Forecasting Is Worth Your Attention Right Now</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Mon, 13 Apr 2026 14:33:15 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/why-ai-time-series-forecasting-is-worth-your-attention-right-now-1088</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/why-ai-time-series-forecasting-is-worth-your-attention-right-now-1088</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhlxm2yi3dmgki7tqnhp8.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhlxm2yi3dmgki7tqnhp8.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;br&gt;
Most developers have heard of time series forecasting. Fewer have kept up with how dramatically the tooling and underlying models have changed over the past couple of years. If your last mental model of this space involves ARIMA and seasonal decomposition, it is worth a refresh.&lt;/p&gt;

&lt;p&gt;The gap between classical statistical methods and modern AI-driven approaches has grown large enough that it changes what is practical to build, who can build it, and how much effort it takes to get production-quality results.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Problem with Classical Methods
&lt;/h2&gt;

&lt;p&gt;ARIMA and its variants were the standard for a long time for good reasons. They are interpretable, computationally cheap, and well-supported by decades of statistical theory. The problem is that they assume linearity and stationarity in the underlying data. Real-world time series rarely cooperates with those assumptions for long.&lt;/p&gt;

&lt;p&gt;When a demand signal shifts, a financial instrument spikes, or a sensor reading starts drifting, classical models degrade. They were not designed to handle nonlinear dynamics or sudden distributional shifts, and no amount of tuning changes that fundamental constraint.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Deep Learning Brought to the Table
&lt;/h2&gt;

&lt;p&gt;LSTMs and GRUs were the first architectures to make a real dent in this problem. They were specifically designed to model long-range dependencies in sequential data, which made them far better suited to the kinds of patterns that break classical models. Transformers followed, and despite being designed for language tasks, they turned out to be remarkably effective for long-horizon forecasting.&lt;/p&gt;

&lt;p&gt;A comprehensive review in the Journal of Big Data quantified the improvement: deep learning approaches outperform classical statistical methods by up to 14% on forecasting accuracy, with the gap growing as data complexity increases. That is not a marginal difference at production scale.&lt;/p&gt;

&lt;h2&gt;
  
  
  Foundation Models Changed the Deployment Story
&lt;/h2&gt;

&lt;p&gt;The bigger shift for practitioners is what foundation models have done to the cost of getting started. Google's TimesFM was pre-trained on over 100 billion time-series data points and delivers strong zero-shot performance on datasets it has never encountered. Amazon's Chronos tokenizes numerical values and applies transformer-based techniques borrowed directly from large language models, benchmarking well across 42 diverse datasets.&lt;/p&gt;

&lt;p&gt;What this means in practice is that you no longer need a large domain-specific training set to build a useful forecasting system. You start from a strong pre-trained baseline and fine-tune from there. For teams without dedicated data science resources, that is a significant change in what is feasible.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Technique Worth Knowing: Future-Guided Learning
&lt;/h2&gt;

&lt;p&gt;One of the more interesting recent developments comes from a paper published in Nature Communications. The technique, called Future-Guided Learning, runs two models in parallel. A detection model analyzes future data to identify critical events, while a forecasting model learns to predict those events from current data. When predictions diverge from detections, the forecasting model updates more aggressively to close the gap.&lt;/p&gt;

&lt;p&gt;The results were a 23% reduction in prediction error on nonlinear dynamical systems and a 44.8% improvement in AUC-ROC for seizure prediction. The interesting part is the design philosophy: rather than minimizing average error, you are training the model to recognize and actively correct its own failure modes.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Actually Breaks Forecasting Systems in Production
&lt;/h2&gt;

&lt;p&gt;Model architecture is only part of the challenge. Production forecasting systems fail for reasons that have nothing to do with which transformer variant you chose. Data quality is the most common culprit. Time series data arrives with gaps, duplicate entries, inconsistent sampling rates, and outliers that distort training in ways that are hard to detect until something downstream goes wrong.&lt;/p&gt;

&lt;p&gt;Evaluation methodology is another area where teams get tripped up. Mean squared error is the default, but it rewards models that predict the mean and discourages variance. Depending on your use case, directional accuracy, peak detection, or calibration might be far more relevant. And once a model is in production, you need monitoring in place to catch performance degradation as the underlying data distribution shifts over time.&lt;/p&gt;

&lt;h2&gt;
  
  
  Read the Full Breakdown
&lt;/h2&gt;

&lt;p&gt;The full post covers all of this in depth, including a look at real-world applications across finance, healthcare, supply chain, and energy, along with practical guidance on architecture selection and getting started without overengineering the solution.&lt;/p&gt;

&lt;p&gt;Read it here: &lt;a href="https://aitransformer.online/ai-time-series-forecasting/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-time-series-forecasting/&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>timeseries</category>
    </item>
    <item>
      <title>Your KPIs Are Already Too Late</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Fri, 10 Apr 2026 14:25:11 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/your-kpis-are-already-too-late-1118</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/your-kpis-are-already-too-late-1118</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg30og3oyxr7x1ge485kq.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg30og3oyxr7x1ge485kq.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Most project KPIs are built to explain what already happened. They show delays, budget overruns, and missed targets after the fact. That might feel useful, but it does not help teams prevent failure. It only helps them document it.&lt;/p&gt;

&lt;p&gt;That is the real limitation of traditional metrics. They are backward-looking by design.&lt;/p&gt;

&lt;h2&gt;
  
  
  AI KPI Development Changes the Timing
&lt;/h2&gt;

&lt;p&gt;AI KPI development shifts the focus from reporting to anticipation. Instead of static dashboards, AI-driven metrics analyze patterns, detect early signals, and surface risks before they become visible in standard reports.&lt;/p&gt;

&lt;p&gt;This changes how project leaders operate. Decisions are no longer based only on past performance. They are shaped by what is likely to happen next.&lt;/p&gt;

&lt;p&gt;That shift creates a real advantage, especially in complex environments where small issues can escalate quickly.&lt;/p&gt;

&lt;h2&gt;
  
  
  Better Metrics Do Not Fix Broken Systems
&lt;/h2&gt;

&lt;p&gt;There is a common misconception that better dashboards will automatically improve outcomes. That is not true.&lt;/p&gt;

&lt;p&gt;If the underlying data is inconsistent, the goals are unclear, or the team does not trust the numbers, AI will only amplify those problems. It will produce faster insights, but not better ones.&lt;/p&gt;

&lt;p&gt;AI KPI development works best when it is built on strong foundations. Clear objectives, clean data, and aligned teams are still essential.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Shift Is Operational
&lt;/h2&gt;

&lt;p&gt;The move to AI-driven KPIs is not just a technical upgrade. It is a change in how organizations think about performance and decision-making.&lt;/p&gt;

&lt;p&gt;It requires leaders to move away from static reporting and toward continuous evaluation. It requires teams to trust evolving metrics instead of fixed benchmarks.&lt;/p&gt;

&lt;p&gt;That is not a small change. It is a fundamental shift in how projects are managed.&lt;/p&gt;

&lt;h2&gt;
  
  
  Rethink What Your KPIs Are Telling You
&lt;/h2&gt;

&lt;p&gt;If your KPIs still feel like a post-mortem, they are not doing enough. The goal is not just to understand what failed. It is to prevent failure in the first place.&lt;/p&gt;

&lt;p&gt;AI KPI development is one way to move closer to that goal, but only if it is implemented with the right structure and mindset.&lt;/p&gt;

&lt;p&gt;Read the full post here: &lt;a href="https://aitransformer.online/ai-kpi-development/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-kpi-development/&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>projectmanagement</category>
      <category>datascience</category>
      <category>businessintelligence</category>
    </item>
    <item>
      <title>Hackers Are Not Breaking In Anymore; They Are Logging In</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Thu, 09 Apr 2026 14:28:22 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/hackers-are-not-breaking-in-anymore-they-are-logging-in-32hd</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/hackers-are-not-breaking-in-anymore-they-are-logging-in-32hd</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo9nyavlm3cedhw66o4p9.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo9nyavlm3cedhw66o4p9.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;br&gt;
Identity Is the Weakest Link in Modern Security&lt;/p&gt;

&lt;p&gt;For years, security strategies focused on infrastructure. Networks, endpoints, and applications were the priority.&lt;/p&gt;

&lt;p&gt;That is no longer where attackers are concentrating their efforts.&lt;/p&gt;

&lt;p&gt;They are targeting identity.&lt;/p&gt;

&lt;p&gt;Deepfakes, synthetic identities, credential abuse, and account takeovers are allowing attackers to bypass traditional defenses without triggering obvious alarms. Instead of breaking systems, they are logging in and operating as legitimate users.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Traditional Detection Is Failing
&lt;/h2&gt;

&lt;p&gt;Most detection systems still rely on rules.&lt;/p&gt;

&lt;p&gt;If a login happens from a new location, flag it. If behavior looks unusual, trigger an alert. That worked when threats were slower and easier to predict.&lt;/p&gt;

&lt;p&gt;It does not work when attackers can generate thousands of realistic identities or automate login attempts at scale.&lt;/p&gt;

&lt;p&gt;By the time a rule is triggered, the attacker may already be inside.&lt;/p&gt;

&lt;h2&gt;
  
  
  How AI Changes Identity Threat Detection
&lt;/h2&gt;

&lt;p&gt;AI introduces a different approach.&lt;/p&gt;

&lt;p&gt;Instead of relying only on predefined rules, it analyzes patterns across behavior, access, and context. It can detect subtle anomalies that would not be obvious in isolation.&lt;/p&gt;

&lt;p&gt;This allows organizations to move earlier in the attack lifecycle and reduce the window of exposure.&lt;/p&gt;

&lt;p&gt;It also enables more adaptive responses as threats evolve.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Tradeoffs You Cannot Ignore
&lt;/h2&gt;

&lt;p&gt;AI is not a perfect solution.&lt;/p&gt;

&lt;p&gt;False positives can interrupt legitimate users. Bias in models can lead to uneven outcomes. Privacy concerns grow as more identity data is analyzed and stored.&lt;/p&gt;

&lt;p&gt;The challenge is not just adopting AI, but implementing it in a way that balances detection with trust.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where This Is Going
&lt;/h2&gt;

&lt;p&gt;Identity is becoming the new security perimeter.&lt;/p&gt;

&lt;p&gt;Organizations that rethink identity protection with AI will be better positioned to handle modern threats. Those that rely only on legacy detection models will continue to react after the damage is done.&lt;/p&gt;

&lt;p&gt;If you are working in security, fraud prevention, or platform engineering, this shift is already affecting your systems.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Full breakdown here:&lt;/strong&gt; &lt;a href="https://aitransformer.online/ai-identity-threat-detection/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-identity-threat-detection/&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  devto #cybersecurity #ai #identitysecurity #fraudprevention
&lt;/h1&gt;

</description>
      <category>ai</category>
      <category>cybersecurity</category>
      <category>security</category>
      <category>identity</category>
    </item>
    <item>
      <title>Your AI Architecture Is Probably Doing Too Much</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:25:00 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/your-ai-architecture-is-probably-doing-too-much-3kbp</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/your-ai-architecture-is-probably-doing-too-much-3kbp</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhytmjccvmigk5bjfq7b2.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhytmjccvmigk5bjfq7b2.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;A lot of AI backends grow in a very predictable way.&lt;/p&gt;

&lt;p&gt;You start simple. One model call. One workflow.&lt;/p&gt;

&lt;p&gt;Then you add features.&lt;/p&gt;

&lt;p&gt;Another endpoint. Another integration. Another layer of logic. Maybe some caching. Maybe a queue. Maybe a workaround for latency. Maybe a second model.&lt;/p&gt;

&lt;p&gt;Nothing feels wrong in the moment.&lt;/p&gt;

&lt;p&gt;But over time, the system becomes harder to reason about.&lt;/p&gt;

&lt;h2&gt;
  
  
  Complexity Doesn’t Announce Itself
&lt;/h2&gt;

&lt;p&gt;AI systems rarely break with a clear failure.&lt;/p&gt;

&lt;p&gt;Instead, they become harder to operate.&lt;/p&gt;

&lt;p&gt;Small changes take longer. Costs become less predictable. Performance varies in ways that are difficult to explain. Fixes create side effects somewhere else.&lt;/p&gt;

&lt;p&gt;This is not a scaling problem yet.&lt;/p&gt;

&lt;p&gt;It is a structure problem.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Backend Starts to Drift
&lt;/h2&gt;

&lt;p&gt;Without clear patterns, AI backends tend to evolve into tightly coupled systems.&lt;/p&gt;

&lt;p&gt;Inference logic mixes with orchestration. Data handling leaks into request handling. Retry logic lives in random places. Observability becomes an afterthought instead of a built-in capability.&lt;/p&gt;

&lt;p&gt;At that point, every new feature increases risk.&lt;/p&gt;

&lt;p&gt;Not because the idea is bad.&lt;/p&gt;

&lt;p&gt;Because the system is no longer easy to extend safely.&lt;/p&gt;

&lt;h2&gt;
  
  
  Good Patterns Reduce Friction
&lt;/h2&gt;

&lt;p&gt;The goal is not to eliminate complexity.&lt;/p&gt;

&lt;p&gt;It is to contain it.&lt;/p&gt;

&lt;p&gt;Well-designed AI backends separate concerns early. They make it clear where decisions are made, where failures are handled, and how data moves through the system.&lt;/p&gt;

&lt;p&gt;That clarity makes everything else easier.&lt;/p&gt;

&lt;p&gt;Scaling becomes more predictable. Debugging becomes faster. Iteration becomes safer.&lt;/p&gt;

&lt;p&gt;This Is Where AI Products Are Won or Lost&lt;/p&gt;

&lt;p&gt;The model gets attention.&lt;/p&gt;

&lt;p&gt;The backend determines outcomes.&lt;/p&gt;

&lt;p&gt;If the system is structured well, it can absorb growth, change, and new capabilities without falling apart.&lt;/p&gt;

&lt;p&gt;If it is not, every improvement becomes harder than the last.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;I break this down in more detail here:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://aitransformer.online/ai-backend-development-patterns/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-backend-development-patterns/&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  AI #SoftwareEngineering #BackendDevelopment #MLOps #SystemDesign
&lt;/h1&gt;

</description>
      <category>ai</category>
      <category>architecture</category>
      <category>backend</category>
      <category>systemdesign</category>
    </item>
    <item>
      <title>Your AI Is a Black Box Because You Didn’t Document It</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Tue, 07 Apr 2026 14:20:16 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/your-ai-is-a-black-box-because-you-didnt-document-it-41pg</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/your-ai-is-a-black-box-because-you-didnt-document-it-41pg</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd9b2kh9jpft5581m6z34.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd9b2kh9jpft5581m6z34.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;AI Systems Are Failing for a Different Reason&lt;/p&gt;

&lt;p&gt;AI systems are not just failing because of bad models.&lt;/p&gt;

&lt;p&gt;They are failing because no one can explain them.&lt;/p&gt;

&lt;p&gt;No clear data lineage. No record of decisions. No understanding of how the model evolved over time. Just systems that work until they don’t, and when they break, no one knows why.&lt;/p&gt;

&lt;h2&gt;
  
  
  This Is Not a Modeling Problem
&lt;/h2&gt;

&lt;p&gt;This is not a modeling problem.&lt;/p&gt;

&lt;p&gt;It is a documentation problem.&lt;/p&gt;

&lt;p&gt;Most teams still treat documentation as cleanup work. Something to do after training. Something to patch together before deployment. Something to revisit only when governance or compliance forces the issue.&lt;/p&gt;

&lt;p&gt;That approach does not scale.&lt;/p&gt;

&lt;p&gt;The Lifecycle Is Where It Breaks&lt;/p&gt;

&lt;p&gt;AI documentation has to follow the full lifecycle.&lt;/p&gt;

&lt;p&gt;It starts at planning. It continues through data collection, model development, evaluation, deployment, and monitoring. It evolves as the system evolves.&lt;/p&gt;

&lt;p&gt;Without that, teams lose traceability. They lose reproducibility. They lose trust.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why This Is Now a Real Risk
&lt;/h2&gt;

&lt;p&gt;Organizations are being asked to explain how their models work, what data shaped them, and how decisions are made.&lt;/p&gt;

&lt;p&gt;If the documentation is weak, those answers do not exist.&lt;/p&gt;

&lt;p&gt;That is where systems fail, not just technically, but operationally.&lt;/p&gt;

&lt;h2&gt;
  
  
  Documentation Is Infrastructure
&lt;/h2&gt;

&lt;p&gt;Documentation is not overhead.&lt;/p&gt;

&lt;p&gt;It is infrastructure.&lt;/p&gt;

&lt;p&gt;It connects data to models, models to decisions, and decisions to accountability. Without it, everything else becomes harder to manage and easier to break.&lt;/p&gt;

&lt;h2&gt;
  
  
  Read the Full Breakdown
&lt;/h2&gt;

&lt;p&gt;I wrote a deeper breakdown of the AI documentation lifecycle and what teams need to change.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://aitransformer.online/ai-documentation-lifecycle/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-documentation-lifecycle/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Tags:&lt;br&gt;
ai, machine-learning, technical-writing, mlops, devops, data-engineering&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>AI Data Pipeline Optimization: Why Most AI Data Pipelines Are Quietly Failing</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Mon, 06 Apr 2026 14:44:52 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/ai-data-pipeline-optimization-why-most-ai-data-pipelines-are-quietly-failing-3e2c</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/ai-data-pipeline-optimization-why-most-ai-data-pipelines-are-quietly-failing-3e2c</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbpjh5tz2ns3w857yscyl.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbpjh5tz2ns3w857yscyl.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;br&gt;
Most AI data pipelines are quietly failing.&lt;/p&gt;

&lt;p&gt;They are not always breaking in obvious ways. Instead, they are slowing decisions, degrading data quality, and creating hidden risks that compound over time. That is why AI data pipeline optimization is becoming essential.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Problem With Traditional Pipelines
&lt;/h2&gt;

&lt;p&gt;As pipelines scale, complexity increases. More data sources. More transformations. More dependencies.&lt;/p&gt;

&lt;p&gt;Traditional approaches struggle to keep up because they rely on static workflows and manual fixes. When something goes wrong, teams react. By then, the damage is already done.&lt;/p&gt;

&lt;h2&gt;
  
  
  AI Changes How Pipelines Behave
&lt;/h2&gt;

&lt;p&gt;AI introduces a different model.&lt;/p&gt;

&lt;p&gt;Pipelines can detect anomalies early, adapt to changing conditions, and optimize how data moves through systems. They stop being passive infrastructure and start acting like intelligent systems.&lt;/p&gt;

&lt;p&gt;This reduces downtime, improves data quality, and removes a lot of the manual overhead that slows teams down.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why This Impacts Everything
&lt;/h2&gt;

&lt;p&gt;Your pipeline is not just a backend system. It affects every decision your business makes.&lt;/p&gt;

&lt;p&gt;If your pipeline is slow, your insights are delayed. If your data is inconsistent, your outputs cannot be trusted.&lt;/p&gt;

&lt;p&gt;AI data pipeline optimization improves reliability, speed, and accuracy at the same time.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Competitive Gap Is Growing
&lt;/h2&gt;

&lt;p&gt;Teams that adopt AI-driven pipelines are moving faster and operating with fewer failures.&lt;/p&gt;

&lt;p&gt;Teams that do not are stuck fixing issues, dealing with delays, and working with data they cannot fully trust.&lt;/p&gt;

&lt;p&gt;That gap is getting wider.&lt;/p&gt;

&lt;h2&gt;
  
  
  Final Thought
&lt;/h2&gt;

&lt;p&gt;AI data pipeline optimization is not optional anymore. It is becoming a requirement for teams that want to stay competitive.&lt;/p&gt;

&lt;p&gt;If you want a deeper breakdown of how this works in practice, read the full post:&lt;br&gt;
&lt;a href="https://aitransformer.online/ai-data-pipeline-optimization/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-data-pipeline-optimization/&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  ai #dataengineering #mlops #datascience #machinelearning
&lt;/h1&gt;

</description>
      <category>ai</category>
      <category>dataengineering</category>
      <category>mlops</category>
      <category>datascience</category>
    </item>
    <item>
      <title>AI Knows Your Project Budget Will Fail Before You Do</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Fri, 03 Apr 2026 14:37:26 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/ai-knows-your-project-budget-will-fail-before-you-do-41d7</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/ai-knows-your-project-budget-will-fail-before-you-do-41d7</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsth5tikyby22r4vj1ufx.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsth5tikyby22r4vj1ufx.jpg" alt=" " width="800" height="603"&gt;&lt;/a&gt;&lt;br&gt;
Most project budgets do not fail all at once. They drift.&lt;/p&gt;

&lt;p&gt;A small variance here. A missed assumption there. Then suddenly, the numbers no longer reflect reality and no one can clearly explain when things went wrong.&lt;/p&gt;

&lt;p&gt;If you have worked on a tech project, you have seen this pattern. It is not usually the result of poor planning. It is the result of using static forecasting in an environment that never stops changing.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Limits of Spreadsheets
&lt;/h2&gt;

&lt;p&gt;Traditional budget forecasting depends on periodic updates. Teams check in weekly or monthly and adjust based on what has already happened.&lt;/p&gt;

&lt;p&gt;The problem is that modern projects do not move in neat cycles. Costs shift in real time. Scope evolves during execution. Dependencies change without warning. By the time a spreadsheet catches up, it is already behind.&lt;/p&gt;

&lt;p&gt;That delay is where most budget problems begin.&lt;/p&gt;

&lt;h2&gt;
  
  
  What AI Changes
&lt;/h2&gt;

&lt;p&gt;AI budget forecasting works differently because it does not rely on fixed update cycles. It continuously evaluates cost signals, usage patterns, and historical trends while adapting to what is happening in the present.&lt;/p&gt;

&lt;p&gt;This creates something teams rarely have with traditional methods. Earlier visibility into where things are heading.&lt;/p&gt;

&lt;p&gt;It does not eliminate uncertainty, but it reduces the gap between reality and what teams think is happening.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why This Matters for Developers and PMs
&lt;/h2&gt;

&lt;p&gt;When budgets drift, the impact shows up quickly in delivery. Features get cut. Timelines move. Priorities shift midstream.&lt;/p&gt;

&lt;p&gt;With AI-driven forecasting, teams can see those pressures forming sooner. That makes it possible to adjust direction before problems become visible at the executive level.&lt;/p&gt;

&lt;p&gt;The result is not just better financial control. It is more stable execution.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Shift You Should Pay Attention To
&lt;/h2&gt;

&lt;p&gt;This is not about replacing project managers or finance teams. It is about improving how decisions are made when conditions are constantly changing.&lt;/p&gt;

&lt;p&gt;Teams that adopt AI forecasting earlier will not just manage budgets more effectively. They will operate with fewer surprises and more confidence in their plans.&lt;/p&gt;

&lt;p&gt;That difference compounds over time.&lt;/p&gt;

&lt;h2&gt;
  
  
  Read the Full Breakdown
&lt;/h2&gt;

&lt;p&gt;If you are working in software development, project management, or tech leadership, this is worth understanding now.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://aitransformer.online/ai-budget-forecasting/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-budget-forecasting/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Tags: ai, projectmanagement, finops, softwaredevelopment, forecasting&lt;/p&gt;

</description>
      <category>ai</category>
      <category>projectmanagement</category>
      <category>finops</category>
      <category>softwaredevelopment</category>
    </item>
    <item>
      <title>AI Is Creating Security Failures You Cannot See</title>
      <dc:creator>Scott McMahan</dc:creator>
      <pubDate>Thu, 02 Apr 2026 14:37:41 +0000</pubDate>
      <link>https://forem.com/scott_mcmahan_d085ae6e508/ai-is-creating-security-failures-you-cannot-see-2cab</link>
      <guid>https://forem.com/scott_mcmahan_d085ae6e508/ai-is-creating-security-failures-you-cannot-see-2cab</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fty7yqv7sod1ajo46l6no.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fty7yqv7sod1ajo46l6no.jpg" alt=" " width="800" height="602"&gt;&lt;/a&gt;&lt;br&gt;
AI is changing how security teams detect threats, analyze signals, and support decisions.&lt;/p&gt;

&lt;p&gt;It is also creating a new category of risk that many organizations still do not fully understand.&lt;/p&gt;

&lt;p&gt;When AI is used in security, the issue is not just model performance. It is also resilience, governance, and trust. Adversarial attacks, data poisoning, model inversion, and model drift can all undermine systems that appear effective on the surface.&lt;/p&gt;

&lt;p&gt;That is why AI model risk management matters.&lt;/p&gt;

&lt;p&gt;Security teams need to think beyond accuracy and start asking harder questions about failure, manipulation, monitoring, and accountability.&lt;/p&gt;

&lt;p&gt;I wrote a post breaking down the risks, the controls, and the governance issues surrounding AI model risk management in security.&lt;/p&gt;

&lt;p&gt;Read it here:&lt;br&gt;
&lt;a href="https://aitransformer.online/ai-model-risk-management-in-security/" rel="noopener noreferrer"&gt;https://aitransformer.online/ai-model-risk-management-in-security/&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  ai #cybersecurity #security #machinelearning #devops #infosec #riskmanagement
&lt;/h1&gt;

</description>
      <category>ai</category>
      <category>cybersecurity</category>
      <category>machinelearning</category>
      <category>security</category>
    </item>
  </channel>
</rss>
