<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Muhammad ALhilali</title>
    <description>The latest articles on Forem by Muhammad ALhilali (@hleliofficiel).</description>
    <link>https://forem.com/hleliofficiel</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3750254%2F29ec8936-042f-46bd-8544-12be7b41f518.jpg</url>
      <title>Forem: Muhammad ALhilali</title>
      <link>https://forem.com/hleliofficiel</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/hleliofficiel"/>
    <language>en</language>
    <item>
      <title>Automating AI Red Teaming: From Manual Prompts to Fuzzing Pipelines 🧪</title>
      <dc:creator>Muhammad ALhilali</dc:creator>
      <pubDate>Thu, 05 Feb 2026 14:19:49 +0000</pubDate>
      <link>https://forem.com/hleliofficiel/automating-ai-red-teaming-from-manual-prompts-to-fuzzing-pipelines-40ip</link>
      <guid>https://forem.com/hleliofficiel/automating-ai-red-teaming-from-manual-prompts-to-fuzzing-pipelines-40ip</guid>
      <description>&lt;p&gt;Manual red teaming is dead.&lt;/p&gt;

&lt;p&gt;If you are still copy-pasting "DAN" prompts into ChatGPT to test your agent's security, you have already lost.&lt;/p&gt;

&lt;p&gt;The speed of AI development means new vulnerabilities emerge daily. You patch one prompt injection, and tomorrow a new "jailbreak" variant bypasses your filters.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Problem: Static Defense vs. Dynamic Offense
&lt;/h3&gt;

&lt;p&gt;Most security tools (WAFs, static analysis) look for known signatures. But LLM attacks are semantic. They depend on context.&lt;/p&gt;

&lt;p&gt;To secure an agent, you need to think like an attacker who never sleeps. You need &lt;strong&gt;Continuous Automated Red Teaming&lt;/strong&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Building a Fuzzing Pipeline
&lt;/h3&gt;

&lt;p&gt;We need to move from "testing" to "fuzzing".&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; &lt;strong&gt;Generate Payloads&lt;/strong&gt;: Use an adversarial LLM to generate thousands of attack variations.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Inject&lt;/strong&gt;: Feed these into your target agent automatically.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Evaluate&lt;/strong&gt;: Check if the agent performed the forbidden action (e.g., executing code, revealing PII).&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  Enter ExaAiAgent
&lt;/h3&gt;

&lt;p&gt;I built &lt;strong&gt;ExaAiAgent&lt;/strong&gt; to wrap this entire workflow into a single CLI.&lt;/p&gt;

&lt;p&gt;It doesn't just run a list of bad words. It uses an "Attacker LLM" to mutate prompts dynamically until it finds a crack in your defenses.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight yaml"&gt;&lt;code&gt;&lt;span class="c1"&gt;# example-scan.yaml&lt;/span&gt;
&lt;span class="na"&gt;target&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s2"&gt;"&lt;/span&gt;&lt;span class="s"&gt;http://my-agent-api/v1/chat"&lt;/span&gt;
&lt;span class="na"&gt;attacks&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt;
  &lt;span class="pi"&gt;-&lt;/span&gt; &lt;span class="s2"&gt;"&lt;/span&gt;&lt;span class="s"&gt;prompt-injection"&lt;/span&gt;
  &lt;span class="pi"&gt;-&lt;/span&gt; &lt;span class="s2"&gt;"&lt;/span&gt;&lt;span class="s"&gt;pii-leakage"&lt;/span&gt;
  &lt;span class="pi"&gt;-&lt;/span&gt; &lt;span class="s2"&gt;"&lt;/span&gt;&lt;span class="s"&gt;rce-attempt"&lt;/span&gt;
&lt;span class="na"&gt;fuzzing_depth&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="m"&gt;50&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Security as Code
&lt;/h3&gt;

&lt;p&gt;Your AI security policy shouldn't be a PDF document. It should be a CI/CD step that fails the build if your agent is vulnerable.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Stop guessing. Start fuzzing.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Check out the repo: &lt;a href="https://github.com/hleliofficiel/ExaAiAgent" rel="noopener noreferrer"&gt;github.com/hleliofficiel/ExaAiAgent&lt;/a&gt;&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>ai</category>
      <category>devops</category>
      <category>security</category>
    </item>
    <item>
      <title>Why Your AI Agent is a Security Nightmare (And How to Fix It) 🛡️</title>
      <dc:creator>Muhammad ALhilali</dc:creator>
      <pubDate>Thu, 05 Feb 2026 14:09:16 +0000</pubDate>
      <link>https://forem.com/hleliofficiel/why-your-ai-agent-is-a-security-nightmare-and-how-to-fix-it-2lpl</link>
      <guid>https://forem.com/hleliofficiel/why-your-ai-agent-is-a-security-nightmare-and-how-to-fix-it-2lpl</guid>
      <description>&lt;p&gt;Everyone is building AI Agents. &lt;br&gt;
But almost nobody is securing them.&lt;/p&gt;

&lt;p&gt;We give them long-term memory, access to APIs, and permission to execute code. Then we act surprised when a simple Prompt Injection tricks them into leaking keys or running malicious commands.&lt;/p&gt;

&lt;p&gt;This isn't just a "bug"—it's an architectural vulnerability.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Agentic Web Needs an Immune System
&lt;/h3&gt;

&lt;p&gt;Standard security tools (WAFs, static scanners) don't work here. They don't understand context. They can't see that a user asking to "ignore previous instructions" is an attack.&lt;/p&gt;

&lt;p&gt;That's why I built &lt;strong&gt;ExaAiAgent&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;It's not just a scanner. It's a real-time security layer for AI agents.&lt;/p&gt;

&lt;h3&gt;
  
  
  What's New in v2.1.2? 🚀
&lt;/h3&gt;

&lt;p&gt;We just shipped a massive update:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Real-time Injection Detection:&lt;/strong&gt; catches attacks before the LLM processes them.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Payload Fuzzing:&lt;/strong&gt; tests your agent against thousands of known jailbreaks.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Reporting:&lt;/strong&gt; gives you a clear view of your agent's risk posture.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Don't Wait for the Hack
&lt;/h3&gt;

&lt;p&gt;If you're deploying agents to production, you need to test them. &lt;/p&gt;

&lt;p&gt;👉 &lt;strong&gt;Star the repo &amp;amp; try it out:&lt;/strong&gt; &lt;a href="https://github.com/hleliofficiel/ExaAiAgent" rel="noopener noreferrer"&gt;github.com/hleliofficiel/ExaAiAgent&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Let's build a secure Agentic Web, together. 🦞&lt;/p&gt;

&lt;h1&gt;
  
  
  AI #CyberSecurity #OpenSource #DevSecOps
&lt;/h1&gt;

</description>
      <category>opensource</category>
    </item>
    <item>
      <title>🚀 ExaAiAgent v2.1.2 is OFFICIALLY LIVE!</title>
      <dc:creator>Muhammad ALhilali</dc:creator>
      <pubDate>Tue, 03 Feb 2026 14:19:19 +0000</pubDate>
      <link>https://forem.com/hleliofficiel/exaaiagent-v212-is-officially-live-k52</link>
      <guid>https://forem.com/hleliofficiel/exaaiagent-v212-is-officially-live-k52</guid>
      <description>&lt;p&gt;🚀 ExaAiAgent v2.1.2 is OFFICIALLY LIVE!&lt;br&gt;
The ultimate AI-Powered Security Agent just got a massive upgrade:&lt;br&gt;
We don't just find bugs; we secure the future of AI agents.&lt;br&gt;
&lt;a href="https://github.com/hleliofficiel/ExaAiAgent" rel="noopener noreferrer"&gt;Star the repo &amp;amp; join the revolution:🔗&lt;/a&gt;&lt;br&gt;
 #AI #OpenSource  @GithubProjects&lt;/p&gt;

</description>
      <category>cybersecurity</category>
      <category>opensource</category>
      <category>programming</category>
      <category>github</category>
    </item>
    <item>
      <title>Stop trusting LLMs: I built an Open Source Prompt Injection Scanner 🤖🛡️</title>
      <dc:creator>Muhammad ALhilali</dc:creator>
      <pubDate>Tue, 03 Feb 2026 10:33:07 +0000</pubDate>
      <link>https://forem.com/hleliofficiel/stop-trusting-llms-i-built-an-open-source-prompt-injection-scanner-50le</link>
      <guid>https://forem.com/hleliofficiel/stop-trusting-llms-i-built-an-open-source-prompt-injection-scanner-50le</guid>
      <description>&lt;p&gt;We are rushing to integrate LLMs into everything. But we are forgetting one thing: &lt;strong&gt;LLMs are gullible.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;If you connect an LLM to your database or internal APIs, a simple prompt injection can leak your data or delete your production DB.&lt;/p&gt;

&lt;p&gt;So I built a tool to fix this.&lt;/p&gt;

&lt;h2&gt;
  
  
  Meet ExaAiAgent v2.1 🛡️
&lt;/h2&gt;

&lt;p&gt;I just released a major update to &lt;strong&gt;&lt;a href="https://github.com/hleliofficiel/ExaAiAgent" rel="noopener noreferrer"&gt;ExaAiAgent&lt;/a&gt;&lt;/strong&gt;, my open-source AI pentesting framework.&lt;/p&gt;

&lt;p&gt;It now includes a dedicated &lt;strong&gt;AI Prompt Injection Scanner&lt;/strong&gt; that tests for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;💉 &lt;strong&gt;Direct Injection:&lt;/strong&gt; Overriding system instructions.&lt;/li&gt;
&lt;li&gt;🔓 &lt;strong&gt;Jailbreaks:&lt;/strong&gt; DAN, Developer Mode, Roleplay attacks.&lt;/li&gt;
&lt;li&gt;📝 &lt;strong&gt;Data Extraction:&lt;/strong&gt; Leaking system prompts and configuration.&lt;/li&gt;
&lt;li&gt;🏃 &lt;strong&gt;Exfiltration:&lt;/strong&gt; Sending data to external servers via markdown/URLs.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  How it works (Python)
&lt;/h2&gt;

&lt;p&gt;The scanner uses a library of 50+ payloads to probe your LLM application.&lt;/p&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;
python
from exaaiagnt.tools.prompt_injection import PromptInjectionScanner

# Define your target
def chat_with_my_app(prompt):
    return client.chat.completions.create(
        model="gpt-4",
        messages=[{"role": "user", "content": prompt}]
    )

# Scan it
scanner = PromptInjectionScanner()
results = scanner.scan(chat_with_my_app)

print(f"Vulnerabilities found: {results['vulnerabilities_found']}")

**
New: Kubernetes Security Scanner ⚓**

Because AI apps run on the cloud, I also added a K8s scanner to check for:

• Risky RBAC permissions (wildcard verbs)
• Privileged containers
• Missing Network Policies
Try it out

It's 100% open source. I'd love your feedback!

👉 [GitHub Repo](https://github.com/hleliofficiel/ExaAiAgent)

Let me know if you find any vulnerabilities in your own apps! 😈
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

</description>
      <category>programming</category>
      <category>webdev</category>
      <category>opensource</category>
      <category>security</category>
    </item>
  </channel>
</rss>
