<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Jaspinder Singh</title>
    <description>The latest articles on Forem by Jaspinder Singh (@jaspinder12).</description>
    <link>https://forem.com/jaspinder12</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3802906%2F901e12ec-f12b-446d-92a2-5792f949782b.jpeg</url>
      <title>Forem: Jaspinder Singh</title>
      <link>https://forem.com/jaspinder12</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/jaspinder12"/>
    <language>en</language>
    <item>
      <title>AI-Generated Python Code Is Fast — But Is It Secure?</title>
      <dc:creator>Jaspinder Singh</dc:creator>
      <pubDate>Tue, 03 Mar 2026 03:38:18 +0000</pubDate>
      <link>https://forem.com/jaspinder12/ai-generated-python-code-is-fast-but-is-it-secure-25em</link>
      <guid>https://forem.com/jaspinder12/ai-generated-python-code-is-fast-but-is-it-secure-25em</guid>
      <description>&lt;p&gt;Over the past few months, I’ve been using AI tools (ChatGPT, Copilot, etc.) to generate Python code for small features and experiments.&lt;/p&gt;

&lt;p&gt;It’s fast.&lt;br&gt;
It’s convenient.&lt;br&gt;
It often “looks correct.”&lt;/p&gt;

&lt;p&gt;But I started noticing something uncomfortable.&lt;/p&gt;

&lt;p&gt;A lot of AI-generated Python code includes patterns like:&lt;/p&gt;

&lt;p&gt;SQL queries built with string concatenation&lt;/p&gt;

&lt;p&gt;eval() used without restrictions&lt;/p&gt;

&lt;p&gt;Direct file path concatenation&lt;/p&gt;

&lt;p&gt;Hardcoded API keys&lt;/p&gt;

&lt;p&gt;Unsafe os.system() usage&lt;/p&gt;

&lt;p&gt;Nothing obviously broken.&lt;br&gt;
But potentially insecure.&lt;/p&gt;

&lt;p&gt;As someone experimenting with AI-assisted coding, I kept asking:&lt;/p&gt;

&lt;p&gt;How do we quickly sanity-check AI-generated code before shipping it?&lt;/p&gt;

&lt;p&gt;Manual review works — but it’s easy to miss things, especially for beginners.&lt;/p&gt;

&lt;p&gt;So I built a small experiment called AICodeRisk.&lt;/p&gt;

&lt;p&gt;It’s intentionally simple:&lt;/p&gt;

&lt;p&gt;Paste Python code&lt;/p&gt;

&lt;p&gt;It analyzes for common security vulnerabilities&lt;/p&gt;

&lt;p&gt;Returns a structured JSON risk report&lt;/p&gt;

&lt;p&gt;Includes severity, line numbers, and suggested fixes&lt;/p&gt;

&lt;p&gt;No accounts.&lt;br&gt;
No integrations.&lt;br&gt;
Just paste → analyze → review.&lt;/p&gt;

&lt;p&gt;You can try it here:&lt;br&gt;
&lt;a href="https://aicoderisk-v1.onrender.com/" rel="noopener noreferrer"&gt;https://aicoderisk-v1.onrender.com/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This isn’t a product launch.&lt;br&gt;
I’m validating whether this is even a real pain point.&lt;/p&gt;

&lt;p&gt;I’m curious:&lt;/p&gt;

&lt;p&gt;Do you trust AI-generated code by default?&lt;/p&gt;

&lt;p&gt;Do you manually review everything?&lt;/p&gt;

&lt;p&gt;Would you use a lightweight security sanity-check tool like this?&lt;/p&gt;

&lt;p&gt;Or is this solving the wrong problem?&lt;/p&gt;

&lt;p&gt;Brutal feedback welcome.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>python</category>
      <category>security</category>
    </item>
  </channel>
</rss>
