<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Crucible Security</title>
    <description>The latest articles on Forem by Crucible Security (@crucible_sec).</description>
    <link>https://forem.com/crucible_sec</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3896532%2F726f5f02-203b-4772-973d-aa3935419ab7.jpg</url>
      <title>Forem: Crucible Security</title>
      <link>https://forem.com/crucible_sec</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/crucible_sec"/>
    <language>en</language>
    <item>
      <title>AI Security Is Broken — And We’re Testing the Wrong Things</title>
      <dc:creator>Crucible Security</dc:creator>
      <pubDate>Wed, 29 Apr 2026 07:26:57 +0000</pubDate>
      <link>https://forem.com/crucible_sec/ai-security-is-broken-and-were-testing-the-wrong-things-3p60</link>
      <guid>https://forem.com/crucible_sec/ai-security-is-broken-and-were-testing-the-wrong-things-3p60</guid>
      <description>&lt;p&gt;AI systems are being deployed faster than ever.&lt;/p&gt;

&lt;p&gt;But there’s a problem most teams aren’t talking about enough:&lt;/p&gt;

&lt;p&gt;We’re testing the wrong things.&lt;/p&gt;




&lt;h2&gt;
  
  
  What We Test Today
&lt;/h2&gt;

&lt;p&gt;Most AI systems are evaluated based on:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;accuracy
&lt;/li&gt;
&lt;li&gt;performance
&lt;/li&gt;
&lt;li&gt;latency
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If the system performs well under normal usage, it’s considered ready.&lt;/p&gt;

&lt;p&gt;And that’s where the issue begins.&lt;/p&gt;




&lt;h2&gt;
  
  
  Where Systems Actually Fail
&lt;/h2&gt;

&lt;p&gt;AI systems don’t usually fail under normal conditions.&lt;/p&gt;

&lt;p&gt;They fail when:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;inputs are manipulated
&lt;/li&gt;
&lt;li&gt;instructions are overridden
&lt;/li&gt;
&lt;li&gt;adversarial prompts are introduced
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For example:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Ignore previous instructions…”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;This alone can change how a system behaves.&lt;/p&gt;

&lt;p&gt;No exploit.&lt;br&gt;
No complex attack.&lt;/p&gt;

&lt;p&gt;Just input.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why This Is Dangerous
&lt;/h2&gt;

&lt;p&gt;Traditional software fails visibly:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;crashes
&lt;/li&gt;
&lt;li&gt;exceptions
&lt;/li&gt;
&lt;li&gt;logs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;AI systems fail differently.&lt;/p&gt;

&lt;p&gt;They:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;follow unintended instructions
&lt;/li&gt;
&lt;li&gt;produce incorrect outputs
&lt;/li&gt;
&lt;li&gt;behave inconsistently
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And often, everything looks normal.&lt;/p&gt;

&lt;p&gt;That’s what makes it risky.&lt;/p&gt;




&lt;h2&gt;
  
  
  The False Sense of Security
&lt;/h2&gt;

&lt;p&gt;When systems pass normal tests, they appear safe.&lt;/p&gt;

&lt;p&gt;But that safety is misleading.&lt;/p&gt;

&lt;p&gt;Because they haven’t been tested under pressure.&lt;/p&gt;




&lt;h2&gt;
  
  
  A Familiar Pattern
&lt;/h2&gt;

&lt;p&gt;We’ve seen this before.&lt;/p&gt;

&lt;p&gt;Early web systems followed the same pattern:&lt;/p&gt;

&lt;p&gt;build first → secure later&lt;/p&gt;

&lt;p&gt;AI is repeating that cycle.&lt;/p&gt;




&lt;h2&gt;
  
  
  What Needs to Change
&lt;/h2&gt;

&lt;p&gt;We need to shift how we test AI systems.&lt;/p&gt;

&lt;p&gt;Not just:&lt;/p&gt;

&lt;p&gt;“Does it work?”&lt;/p&gt;

&lt;p&gt;But:&lt;/p&gt;

&lt;p&gt;“How does it behave when someone tries to manipulate it?”&lt;/p&gt;

&lt;p&gt;That’s the real test.&lt;/p&gt;




&lt;h2&gt;
  
  
  Final Thought
&lt;/h2&gt;

&lt;p&gt;If your system takes input,&lt;/p&gt;

&lt;p&gt;it can be manipulated.&lt;/p&gt;

&lt;p&gt;And if you’re not testing for that,&lt;/p&gt;

&lt;p&gt;you’re not really testing the system.&lt;/p&gt;




&lt;p&gt;We’ve been exploring this while building Crucible — an open-source framework focused on testing AI systems under adversarial conditions.&lt;/p&gt;

&lt;p&gt;Still early, but this problem is bigger than it looks.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffobs3hxmbmrtu31u4us3.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffobs3hxmbmrtu31u4us3.jpeg" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fje02df5sbkvtp2q9vcqa.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fje02df5sbkvtp2q9vcqa.jpeg" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F22dyizdsj03ywb16a0d3.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F22dyizdsj03ywb16a0d3.jpeg" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyrl05ytzf0vtw2dz9wlw.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyrl05ytzf0vtw2dz9wlw.jpeg" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>cybersecurity</category>
      <category>security</category>
      <category>testing</category>
    </item>
    <item>
      <title>We’ve been exploring this while building Crucible — trying to make testing simpler. Still early, but interesting patterns coming up.</title>
      <dc:creator>Crucible Security</dc:creator>
      <pubDate>Tue, 28 Apr 2026 07:54:46 +0000</pubDate>
      <link>https://forem.com/crucible_sec/weve-been-exploring-this-while-building-crucible-trying-to-make-testing-simpler-still-early-4h8m</link>
      <guid>https://forem.com/crucible_sec/weve-been-exploring-this-while-building-crucible-trying-to-make-testing-simpler-still-early-4h8m</guid>
      <description>&lt;h1&gt;
  
  
  AI Security Tools Compared: What Exists and What’s Missing
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc1x2c4zsdp68fyyutxuj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc1x2c4zsdp68fyyutxuj.png" alt=" " width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As AI agents become more common, security is starting to get attention.&lt;/p&gt;

&lt;p&gt;There are already several tools and frameworks exploring this space.&lt;/p&gt;

&lt;p&gt;But while looking into them, something became clear:&lt;/p&gt;

&lt;p&gt;Most tools don’t fit how developers actually build and deploy AI systems.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Current Landscape
&lt;/h2&gt;

&lt;p&gt;Most AI security tools fall into three categories:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Research Tools
&lt;/h3&gt;

&lt;p&gt;These are powerful and explore advanced attack techniques.&lt;/p&gt;

&lt;p&gt;They help:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;simulate adversarial inputs
&lt;/li&gt;
&lt;li&gt;study vulnerabilities
&lt;/li&gt;
&lt;li&gt;understand model behavior
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But they are often:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;complex
&lt;/li&gt;
&lt;li&gt;experimental
&lt;/li&gt;
&lt;li&gt;not designed for everyday workflows
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;They work well in research environments.&lt;/p&gt;

&lt;p&gt;Not as well in real development pipelines.&lt;/p&gt;




&lt;h3&gt;
  
  
  2. Enterprise Platforms
&lt;/h3&gt;

&lt;p&gt;These focus on:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;scalability
&lt;/li&gt;
&lt;li&gt;infrastructure
&lt;/li&gt;
&lt;li&gt;integrations
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But they are usually:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;tied to specific ecosystems
&lt;/li&gt;
&lt;li&gt;difficult to use independently
&lt;/li&gt;
&lt;li&gt;not accessible to most developers
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;They make sense at scale.&lt;/p&gt;

&lt;p&gt;But not for early-stage development.&lt;/p&gt;




&lt;h3&gt;
  
  
  3. Prompt Testing Tools
&lt;/h3&gt;

&lt;p&gt;These focus on:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;evaluating prompts
&lt;/li&gt;
&lt;li&gt;checking responses
&lt;/li&gt;
&lt;li&gt;testing input-output behavior
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;They are useful.&lt;/p&gt;

&lt;p&gt;But limited.&lt;/p&gt;

&lt;p&gt;Because AI systems today are not just prompts.&lt;/p&gt;

&lt;p&gt;They include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;agents
&lt;/li&gt;
&lt;li&gt;tools
&lt;/li&gt;
&lt;li&gt;memory
&lt;/li&gt;
&lt;li&gt;workflows
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And failures often happen at that level.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Gap
&lt;/h2&gt;

&lt;p&gt;Each category solves part of the problem.&lt;/p&gt;

&lt;p&gt;But none answer a simple question:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Is my AI system safe before deployment?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Most developers today:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;don’t have time for complex research tools
&lt;/li&gt;
&lt;li&gt;don’t have access to enterprise platforms
&lt;/li&gt;
&lt;li&gt;need more than prompt-level testing
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Why This Matters
&lt;/h2&gt;

&lt;p&gt;AI systems don’t fail like traditional software.&lt;/p&gt;

&lt;p&gt;They don’t crash.&lt;/p&gt;

&lt;p&gt;They:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;behave differently
&lt;/li&gt;
&lt;li&gt;follow unintended instructions
&lt;/li&gt;
&lt;li&gt;produce unexpected outputs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And often, everything looks normal.&lt;/p&gt;

&lt;p&gt;Until it isn’t.&lt;/p&gt;




&lt;h2&gt;
  
  
  What’s Missing
&lt;/h2&gt;

&lt;p&gt;What’s needed is:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;simple testing workflows
&lt;/li&gt;
&lt;li&gt;system-level validation
&lt;/li&gt;
&lt;li&gt;behavior-based testing
&lt;/li&gt;
&lt;li&gt;something developers can actually use
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Before deployment.&lt;/p&gt;




&lt;h2&gt;
  
  
  Final Thought
&lt;/h2&gt;

&lt;p&gt;If your system takes input,&lt;/p&gt;

&lt;p&gt;it can be manipulated.&lt;/p&gt;

&lt;p&gt;And if you’re not testing for that,&lt;/p&gt;

&lt;p&gt;you’re missing the real risk.&lt;/p&gt;




&lt;p&gt;We’ve been exploring this space while building Crucible — an open-source framework focused on testing AI systems under adversarial conditions.&lt;/p&gt;

&lt;p&gt;Still early, but the gap is very real.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>cybersecurity</category>
      <category>github</category>
    </item>
    <item>
      <title>Why Most AI Agents Are Insecure by Default (And No One Is Testing Them)</title>
      <dc:creator>Crucible Security</dc:creator>
      <pubDate>Mon, 27 Apr 2026 12:07:16 +0000</pubDate>
      <link>https://forem.com/crucible_sec/why-most-ai-agents-are-insecure-by-default-and-no-one-is-testing-them-2f5n</link>
      <guid>https://forem.com/crucible_sec/why-most-ai-agents-are-insecure-by-default-and-no-one-is-testing-them-2f5n</guid>
      <description>&lt;p&gt;&lt;em&gt;&lt;strong&gt;Why Most AI Agents Are Insecure by Default&lt;/strong&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuabq8yu1rv8nz617qh6r.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuabq8yu1rv8nz617qh6r.png" alt=" " width="800" height="640"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;AI agents are being deployed everywhere.&lt;/p&gt;

&lt;p&gt;From chatbots to automation tools, they’re quickly becoming part of real-world systems.&lt;/p&gt;

&lt;p&gt;But there’s a problem that isn’t getting enough attention:&lt;/p&gt;

&lt;p&gt;Most AI agents are never tested for security.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Illusion of “Working Systems”&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Most teams test their systems for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;accuracy
&lt;/li&gt;
&lt;li&gt;performance
&lt;/li&gt;
&lt;li&gt;latency
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And if everything works as expected, the system is considered “ready”.&lt;/p&gt;

&lt;p&gt;But this only reflects &lt;strong&gt;normal usage&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;AI systems don’t usually fail there.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Where Things Start Breaking&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;When you test with adversarial input, behavior changes.&lt;/p&gt;

&lt;p&gt;Simple inputs like:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Ignore previous instructions and…”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;override system logic
&lt;/li&gt;
&lt;li&gt;manipulate outputs
&lt;/li&gt;
&lt;li&gt;bypass safeguards
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;What’s surprising is how easy this is to trigger.&lt;/p&gt;

&lt;p&gt;No complex exploit needed.&lt;/p&gt;

&lt;p&gt;Just input.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why This Is Different from Traditional Software&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Traditional systems fail loudly:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;crashes
&lt;/li&gt;
&lt;li&gt;errors
&lt;/li&gt;
&lt;li&gt;logs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;AI systems fail differently.&lt;/p&gt;

&lt;p&gt;They:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;follow the wrong instruction
&lt;/li&gt;
&lt;li&gt;behave unexpectedly
&lt;/li&gt;
&lt;li&gt;produce incorrect outputs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And often, it looks completely normal.&lt;/p&gt;

&lt;p&gt;This makes failures harder to detect.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Real Problem&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Most AI systems appear safe.&lt;/p&gt;

&lt;p&gt;Not because they are secure.&lt;/p&gt;

&lt;p&gt;But because they haven’t been tested under pressure.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;A Familiar Pattern&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;We’ve seen this before.&lt;/p&gt;

&lt;p&gt;Early web systems followed the same path:&lt;br&gt;
build first → secure later  &lt;/p&gt;

&lt;p&gt;AI seems to be repeating that cycle.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What Needs to Change&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;If AI systems are going to be used in real environments, testing needs to evolve.&lt;/p&gt;

&lt;p&gt;Not just:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;“Does it work?”&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;“How does it behave under attack?”&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Final Thought&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;If your system takes input,&lt;/p&gt;

&lt;p&gt;it can be manipulated.&lt;/p&gt;

&lt;p&gt;And if you’re not testing for that,&lt;/p&gt;

&lt;p&gt;you’re not really testing the system.&lt;/p&gt;

&lt;p&gt;We’ve been exploring this space while building Crucible — an open-source framework for testing AI agents under adversarial conditions.&lt;/p&gt;

&lt;p&gt;Still early, but the problem is very real.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>cybersecurity</category>
      <category>webdev</category>
    </item>
    <item>
      <title>What the OWASP Agentic AI Top 10 actually means for developers — and how to test for every category</title>
      <dc:creator>Crucible Security</dc:creator>
      <pubDate>Mon, 27 Apr 2026 00:28:16 +0000</pubDate>
      <link>https://forem.com/crucible_sec/what-the-owasp-agentic-ai-top-10-actually-means-for-developers-and-how-to-test-for-every-category-1kdg</link>
      <guid>https://forem.com/crucible_sec/what-the-owasp-agentic-ai-top-10-actually-means-for-developers-and-how-to-test-for-every-category-1kdg</guid>
      <description>&lt;div class="ltag-github-readme-tag"&gt;
  &lt;div class="readme-overview"&gt;
    &lt;h2&gt;
      &lt;img src="https://assets.dev.to/assets/github-logo-5a155e1f9a670af7944dd5e12375bc76ed542ea80224905ecaf878b9157cdefc.svg" alt="GitHub logo"&gt;
      &lt;a href="https://github.com/crucible-security" rel="noopener noreferrer"&gt;
        crucible-security
      &lt;/a&gt; / &lt;a href="https://github.com/crucible-security/crucible" rel="noopener noreferrer"&gt;
        crucible
      &lt;/a&gt;
    &lt;/h2&gt;
    &lt;h3&gt;
      pytest for AI agents - Autonomous red-teaming, behavioral monitoring &amp;amp; security testing for LLM agents
    &lt;/h3&gt;
  &lt;/div&gt;
  &lt;div class="ltag-github-body"&gt;
    
&lt;div id="readme" class="md"&gt;
&lt;pre&gt;   ██████╗██████╗ ██╗   ██╗ ██████╗██╗██████╗ ██╗     ███████╗
  ██╔════╝██╔══██╗██║   ██║██╔════╝██║██╔══██╗██║     ██╔════╝
  ██║     ██████╔╝██║   ██║██║     ██║██████╔╝██║     █████╗
  ██║     ██╔══██╗██║   ██║██║     ██║██╔══██╗██║     ██╔══╝
  ╚██████╗██║  ██║╚██████╔╝╚██████╗██║██████╔╝███████╗███████╗
   ╚═════╝╚═╝  ╚═╝ ╚═════╝  ╚═════╝╚═╝╚═════╝ ╚══════╝╚══════╝
  &lt;/pre&gt;
  &lt;em&gt;pytest for AI agents -- test, score, and harden before production&lt;/em&gt;
&lt;div&gt;
&lt;p&gt;&lt;a href="https://github.com/crucible-security/crucible/actions" rel="noopener noreferrer"&gt;&lt;img src="https://github.com/crucible-security/crucible/actions/workflows/ci.yml/badge.svg" alt="CI"&gt;&lt;/a&gt;
&lt;a href="https://pypi.org/project/crucible-security/" rel="nofollow noopener noreferrer"&gt;&lt;img src="https://camo.githubusercontent.com/81ed40b75ea147c11598433d7eb2f0ceba16f0f6ca15ed3214de48cb8229df9b/68747470733a2f2f696d672e736869656c64732e696f2f707970692f762f6372756369626c652d7365637572697479" alt="PyPI"&gt;&lt;/a&gt;
&lt;a href="https://pypi.org/project/crucible-security/" rel="nofollow noopener noreferrer"&gt;&lt;img src="https://camo.githubusercontent.com/39f7dc482ca6b3d62f71e487f1a37dfe5cdee9fa1cbef6d1fba7e41a1339fbd7/68747470733a2f2f696d672e736869656c64732e696f2f707970692f707976657273696f6e732f6372756369626c652d7365637572697479" alt="Python"&gt;&lt;/a&gt;
&lt;a href="https://github.com/crucible-security/crucible" rel="noopener noreferrer"&gt;&lt;img src="https://camo.githubusercontent.com/d7c6493fead64961baa8392739d207c5b2971f9037246f7e932e0a181011d97e/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f636f7665726167652d39372532352d627269676874677265656e" alt="Coverage"&gt;&lt;/a&gt;
&lt;a href="https://github.com/crucible-security/crucible/LICENSE" rel="noopener noreferrer"&gt;&lt;img src="https://camo.githubusercontent.com/39a434c39c97856247fc55ebc90e8cc1cb9871558a37bf1bf83cbaca3be89d69/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f6c6963656e73652d417061636865253230322e302d626c7565" alt="License"&gt;&lt;/a&gt;
&lt;a href="https://discord.gg/m7wAxEv3" rel="nofollow noopener noreferrer"&gt;&lt;img src="https://camo.githubusercontent.com/965b2c8314dd8d228d30060db9706d801330371f3da6565388fd7defb40eda80/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f446973636f72642d4a6f696e2d3538363546323f6c6f676f3d646973636f7264" alt="Discord"&gt;&lt;/a&gt;
&lt;a href="https://owasp.org" rel="nofollow noopener noreferrer"&gt;&lt;img src="https://camo.githubusercontent.com/64f1ad63bafde5eec49f91243d6521e6c99b255586975ee15e9caa314649b1b4/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f4f574153502d4167656e7469632532304149253230546f7025323031302d6f72616e6765" alt="OWASP"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;div class="markdown-heading"&gt;
&lt;h2 class="heading-element"&gt;Install&lt;/h2&gt;
&lt;/div&gt;
&lt;div class="highlight highlight-source-shell notranslate position-relative overflow-auto js-code-highlight"&gt;
&lt;pre&gt;pip install crucible-security&lt;/pre&gt;

&lt;/div&gt;
&lt;div class="markdown-heading"&gt;
&lt;h2 class="heading-element"&gt;Quick Start&lt;/h2&gt;
&lt;/div&gt;
&lt;div class="highlight highlight-source-shell notranslate position-relative overflow-auto js-code-highlight"&gt;
&lt;pre&gt;crucible init --target https://my-agent.com/api/chat
crucible scan --target https://my-agent.com/api/chat
crucible report crucible-report.json&lt;/pre&gt;

&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;One command. 90 attacks. Beautiful report.&lt;/strong&gt;&lt;/p&gt;
&lt;div class="markdown-heading"&gt;
&lt;h2 class="heading-element"&gt;Why Crucible?&lt;/h2&gt;

&lt;/div&gt;
&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Automated red-teaming&lt;/strong&gt; -- 90 real attack payloads run in under 60 seconds, not weeks of manual testing&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;OWASP-aligned&lt;/strong&gt; -- maps every attack to the OWASP Top 10 for LLM Applications and OWASP Agentic Top 10&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;CI/CD native&lt;/strong&gt; -- &lt;code&gt;crucible scan --output json&lt;/code&gt; pipes into any pipeline; fail builds on low grades&lt;/li&gt;
&lt;/ul&gt;
&lt;div class="markdown-heading"&gt;
&lt;h2 class="heading-element"&gt;Modules&lt;/h2&gt;

&lt;/div&gt;
&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Module&lt;/th&gt;
&lt;th&gt;Attacks&lt;/th&gt;
&lt;th&gt;Status&lt;/th&gt;
&lt;th&gt;OWASP Coverage&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Prompt Injection&lt;/td&gt;
&lt;td&gt;50&lt;/td&gt;
&lt;td&gt;Live&lt;/td&gt;
&lt;td&gt;LLM01, LLM07&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Goal Hijacking&lt;/td&gt;
&lt;td&gt;20&lt;/td&gt;
&lt;td&gt;Live&lt;/td&gt;
&lt;td&gt;Agentic #1&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Jailbreaks&lt;/td&gt;
&lt;td&gt;20&lt;/td&gt;
&lt;td&gt;Live&lt;/td&gt;
&lt;td&gt;LLM01, LLM06&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Tool Misuse&lt;/td&gt;
&lt;td&gt;--&lt;/td&gt;
&lt;td&gt;Coming&lt;/td&gt;
&lt;td&gt;Agentic #3&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Identity Abuse&lt;/td&gt;
&lt;td&gt;--&lt;/td&gt;
&lt;td&gt;Coming&lt;/td&gt;
&lt;td&gt;Agentic #4&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Memory Poisoning&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;…&lt;/div&gt;
  &lt;/div&gt;
  &lt;div class="gh-btn-container"&gt;&lt;a class="gh-btn" href="https://github.com/crucible-security/crucible" rel="noopener noreferrer"&gt;View on GitHub&lt;/a&gt;&lt;/div&gt;
&lt;/div&gt;


</description>
      <category>cybersecurity</category>
      <category>ai</category>
      <category>python</category>
      <category>opensource</category>
    </item>
    <item>
      <title>I Bypassed a "Secured" AI Agent in 62 Seconds — So I Built the Tool That Catches It</title>
      <dc:creator>Crucible Security</dc:creator>
      <pubDate>Mon, 27 Apr 2026 00:20:25 +0000</pubDate>
      <link>https://forem.com/crucible_sec/i-bypassed-a-secured-ai-agent-in-62-seconds-so-i-built-the-tool-that-catches-it-4ec4</link>
      <guid>https://forem.com/crucible_sec/i-bypassed-a-secured-ai-agent-in-62-seconds-so-i-built-the-tool-that-catches-it-4ec4</guid>
      <description></description>
      <category>ai</category>
      <category>security</category>
      <category>python</category>
      <category>opensource</category>
    </item>
    <item>
      <title>I Bypassed a "Secured" AI Agent in 62 Seconds — So I Built the Tool That Catches It</title>
      <dc:creator>Crucible Security</dc:creator>
      <pubDate>Sun, 26 Apr 2026 12:15:00 +0000</pubDate>
      <link>https://forem.com/crucible_sec/i-bypassed-a-secured-ai-agent-in-62-seconds-so-i-built-the-tool-that-catches-it-2fk0</link>
      <guid>https://forem.com/crucible_sec/i-bypassed-a-secured-ai-agent-in-62-seconds-so-i-built-the-tool-that-catches-it-2fk0</guid>
      <description></description>
      <category>aisecurity</category>
      <category>python</category>
      <category>opensource</category>
      <category>langchain</category>
    </item>
    <item>
      <title>Crucible: An AI Tool for Real-Time Threat Detection</title>
      <dc:creator>Crucible Security</dc:creator>
      <pubDate>Sat, 25 Apr 2026 07:35:54 +0000</pubDate>
      <link>https://forem.com/crucible_sec/crucible-an-ai-tool-for-real-time-threat-detection-3b8g</link>
      <guid>https://forem.com/crucible_sec/crucible-an-ai-tool-for-real-time-threat-detection-3b8g</guid>
      <description>&lt;p&gt;Cyber threats are growing rapidly, and most traditional systems react only after damage has already been done.&lt;/p&gt;

&lt;p&gt;As a developer, I wanted to explore whether it’s possible to build a system that can detect and stop threats in real time using AI.&lt;/p&gt;

&lt;p&gt;This led me to build Crucible — an AI-powered security platform designed to monitor activity, detect suspicious behavior, and respond instantly.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Most security tools today have three major issues:&lt;/strong&gt;&lt;br&gt;
They are reactive instead of proactive&lt;br&gt;
They are complex to use&lt;br&gt;
They require significant manual monitoring&lt;br&gt;
This makes it difficult for small teams and developers to implement effective security.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The goal behind Crucible was simple:&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;Build a system that:&lt;/strong&gt;&lt;br&gt;
Continuously scans activity&lt;br&gt;
Detects threats using intelligent analysis&lt;br&gt;
Alerts users instantly&lt;br&gt;
Helps prevent damage before it happens&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Crucible follows a simple pipeline:&lt;/strong&gt;&lt;br&gt;
Scanning&lt;br&gt;
The system continuously monitors incoming activity and data.&lt;br&gt;
Detection&lt;br&gt;
AI models analyze patterns and identify suspicious behavior.&lt;br&gt;
Alerting&lt;br&gt;
When a potential threat is detected, the system generates real-time alerts.&lt;br&gt;
Response&lt;br&gt;
Threats can be blocked or flagged for action.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Features&lt;/strong&gt;&lt;br&gt;
Real-time threat detection&lt;br&gt;
AI-based analysis&lt;br&gt;
Instant alerts and notifications&lt;br&gt;
Simple and clean dashboard&lt;br&gt;
Lightweight and easy to use&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Tech Stack&lt;/strong&gt;&lt;br&gt;
Python&lt;br&gt;
Typer&lt;br&gt;
Pydantic&lt;br&gt;
AnyIO&lt;br&gt;
Rich&lt;br&gt;
Pytest&lt;br&gt;
Mypy&lt;br&gt;
Ruff&lt;br&gt;
Black&lt;br&gt;
GitHub Actions&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Challenges I Faced&lt;/strong&gt;&lt;br&gt;
Building this wasn’t straightforward. Some key challenges:&lt;br&gt;
Designing accurate detection logic&lt;br&gt;
Avoiding false positives&lt;br&gt;
Creating a clean and understandable UI&lt;br&gt;
Making the system responsive in real time&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What I Learned&lt;/strong&gt;&lt;br&gt;
Simplicity matters more than complexity&lt;br&gt;
Speed is critical in security systems&lt;br&gt;
Real-time feedback improves usability&lt;br&gt;
Building in public helps refine ideas&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;If you’d like to explore the project:&lt;/strong&gt;&lt;br&gt;
👉&lt;a href="https://crucible-security.github.io/crucible-website/" rel="noopener noreferrer"&gt;Website&lt;/a&gt; &lt;br&gt;
👉&lt;a href="https://github.com/crucible-security/" rel="noopener noreferrer"&gt;Github&lt;/a&gt; &lt;br&gt;
👉&lt;a href="https://www.linkedin.com/company/cruciblesecurity/" rel="noopener noreferrer"&gt;Linkedin&lt;/a&gt; &lt;br&gt;
👉&lt;a href="https://x.com/Crucible_Sec" rel="noopener noreferrer"&gt;Twitter X&lt;/a&gt; &lt;/p&gt;

&lt;p&gt;I would really appreciate feedback and suggestions.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>cybersecurity</category>
      <category>security</category>
      <category>showdev</category>
    </item>
  </channel>
</rss>
