<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Swapneswar Sundar Ray</title>
    <description>The latest articles on Forem by Swapneswar Sundar Ray (@swapneswar_sundarray).</description>
    <link>https://forem.com/swapneswar_sundarray</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3897931%2F73304f54-efbd-4435-a6ee-0ed0b2ba2da5.png</url>
      <title>Forem: Swapneswar Sundar Ray</title>
      <link>https://forem.com/swapneswar_sundarray</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/swapneswar_sundarray"/>
    <language>en</language>
    <item>
      <title>Stop Using AI Only to Build—Start Using It to Break Your Systems</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Mon, 04 May 2026 00:45:05 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/stop-using-ai-only-to-build-start-using-it-to-break-your-systems-23gd</link>
      <guid>https://forem.com/swapneswar_sundarray/stop-using-ai-only-to-build-start-using-it-to-break-your-systems-23gd</guid>
      <description>&lt;p&gt;Most of us have gotten comfortable using AI to speed things up—write code, generate tests, clean up documentation. It’s become a productivity tool. But there’s another way to use AI that feels less obvious and, in many cases, more valuable: using it to challenge your system instead of helping it.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy1o7l7vt6dff5d2w64x1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy1o7l7vt6dff5d2w64x1.png" alt=" " width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;If you’ve worked on real production systems, you already know this—things don’t usually break in obvious ways. They break in small, annoying, hard-to-reproduce ways. A value comes in with a slightly different format, a field has an extra space, casing changes, or something gets reordered. Nothing looks “wrong,” but suddenly the system behaves differently. These are the kinds of issues that slip through testing and show up later when it’s much harder to debug.&lt;/p&gt;

&lt;p&gt;The reason this happens is simple. Most testing reflects how engineers think, not how real inputs behave. We test the expected cases, maybe a few edge cases, and call it done. Even automated tools often generate inputs that are either too clean or completely random. Neither really captures how data looks in the wild.&lt;/p&gt;

&lt;p&gt;This is where AI starts to become useful in a different way. Instead of asking it to create solutions, you ask it to create variations. Give it one valid input, and it can produce multiple versions of that same input that still mean the same thing but look slightly different. That’s exactly the kind of variation that exposes weaknesses in systems.&lt;/p&gt;

&lt;p&gt;Think about a basic API that takes an amount and a currency. You test it with something like “1000.00 USD,” and everything works. But what happens when the input becomes “1000”, or “1,000.00”, or has extra spaces, or uses lowercase for the currency? These aren’t unusual cases—they happen all the time. Yet many systems treat them differently, sometimes rejecting them, sometimes misinterpreting them, and sometimes behaving inconsistently.&lt;/p&gt;

&lt;p&gt;Instead of manually trying to think of all these possibilities, you can let AI do that work. Treat it like a mutation engine. Start with one valid input and ask for realistic variations that don’t change the meaning. Then run all of them through your system and observe what happens. You’re no longer just testing whether the system works—you’re testing how stable it is when things are slightly off.&lt;/p&gt;

&lt;p&gt;This changes what you pay attention to. Instead of only asking, “Did this pass or fail?” you start asking, “Did the system behave the same way across all these inputs?” Because if two inputs are effectively the same but produce different outcomes, that’s a deeper issue. It’s not just a bug—it’s inconsistency in how your system interprets the world.&lt;/p&gt;

&lt;p&gt;The nice part is that you don’t need a complicated setup to try this. You can start small. Generate a handful of variations using AI, run them through your existing flow, and compare the results. Even this simple exercise can reveal things that traditional testing misses.&lt;/p&gt;

&lt;p&gt;This approach becomes especially useful in systems where input variability is common. Financial applications are a good example, where formatting differences can affect validations. OCR pipelines often deal with slightly inconsistent outputs for the same text. And modern AI-driven systems themselves can behave differently based on small changes in input phrasing. In all these cases, stability matters just as much as correctness.&lt;/p&gt;

&lt;p&gt;One thing to watch out for is overusing AI without direction. If you generate too many random variations, you end up with noise instead of insight. The goal isn’t to overwhelm the system—it’s to explore meaningful differences. Another common mistake is focusing only on correctness and ignoring consistency. Both matter, but consistency is often what reveals deeper issues.&lt;/p&gt;

&lt;p&gt;A more balanced way to think about this is to combine approaches. Let your code handle strict validation and rules. Use AI to explore the gray areas—the inputs that are technically valid but slightly different. Together, they give you a much better understanding of how your system behaves.&lt;/p&gt;

&lt;p&gt;If you look back at most production issues, they rarely come from completely invalid data. They come from those edge cases that no one thought to test. Usually, a small percentage of inputs ends up causing a large share of problems. Adversarial testing is simply a way to find those cases earlier, when it’s easier to fix them.&lt;/p&gt;

&lt;p&gt;In the end, AI isn’t just a tool for building faster. It’s also a way to question whether what you’ve built actually holds up under real conditions. When you start using it to push your system instead of just supporting it, you begin to uncover things you didn’t even realize were there.&lt;/p&gt;

&lt;p&gt;And that shift—using AI not just as a helper but as something that challenges your system—is where the real learning starts.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>softwareengineering</category>
      <category>llm</category>
    </item>
    <item>
      <title>I tried using AI to build an exam system. It worked… until it didn’t.</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Sat, 02 May 2026 17:05:20 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/i-tried-using-ai-to-build-an-exam-system-it-worked-until-it-didnt-58nm</link>
      <guid>https://forem.com/swapneswar_sundarray/i-tried-using-ai-to-build-an-exam-system-it-worked-until-it-didnt-58nm</guid>
      <description>&lt;p&gt;I didn’t start with the idea of building an exam platform. This actually came from a different problem. We were using AI to generate structured data for APIs, and everything looked fine at first. The responses were correct, nothing obviously wrong. But then things started breaking in production in very strange ways. One example was a value like &lt;code&gt;120.5&lt;/code&gt; instead of &lt;code&gt;120.50&lt;/code&gt;. Same number from a human point of view, but the downstream system rejected it because it expected an exact format. These were small issues, but they took a lot of time to debug and they kept happening.&lt;/p&gt;

&lt;p&gt;That got me thinking. If AI behaves like this with structured data, what happens when we use it to generate exam questions or evaluate answers? In demos it looks impressive. It can generate questions instantly, even evaluate answers. But in real usage, consistency becomes a problem. Difficulty levels vary randomly, answers are not always structured the same way, and evaluation can feel subjective. That’s not something you can rely on for students or schools.&lt;/p&gt;

&lt;p&gt;At first, I tried fixing it the usual way—by improving prompts. Making them longer, adding more rules, being very specific. It helped a little, but it didn’t solve the core issue. You still get edge cases where the output is slightly off. That’s when I realized the problem is not the prompt. The problem is trusting AI output directly without control.&lt;/p&gt;

&lt;p&gt;So instead of trying to “fix AI,” I built a small system around it. It’s a simple Java-based application that runs as a JAR. Students can enter their details, choose subject and topic, and the system generates questions, runs a timer, collects answers, and produces a report. Nothing very new there. The important part is what happens in between.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Home Page (Landing + Features)&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fozojuh1760hanu9iqyg9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fozojuh1760hanu9iqyg9.png" alt=" " width="800" height="454"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is the main entry point of the system. It shows the overall idea clearly — an AI-powered exam platform where users can generate questions, register, and select topics.&lt;/p&gt;

&lt;p&gt;What stands out here is that the system is not just a basic form-based app. It is positioned as a complete examination framework, with features like AI question generation, evaluation, timer, and reporting already integrated.&lt;/p&gt;

&lt;p&gt;The feature section below highlights the core capabilities in a structured way. It shows that the platform is designed to handle the full exam lifecycle, not just question generation. That makes it more like a system solution rather than a small tool.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Student Registration&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4w2lviuyh5dm3ecmo3zv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4w2lviuyh5dm3ecmo3zv.png" alt=" " width="800" height="452"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This screen captures detailed student information — not just name and email, but also age, country, experience, interests, and education level.&lt;/p&gt;

&lt;p&gt;This is important because the system is trying to personalize question generation based on user context. It shows that the design is thinking beyond generic questions and moving toward adaptive exam generation.&lt;/p&gt;

&lt;p&gt;The structure is simple, but the idea behind it is strong — collecting enough context so AI can generate more relevant and meaningful questions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Topic Selection&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkcryke8otkyzrsie3svi.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkcryke8otkyzrsie3svi.png" alt=" " width="800" height="440"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This screen shows predefined exam topics like Java, Spring, System Design, etc., with clear details:&lt;/p&gt;

&lt;p&gt;difficulty level&lt;br&gt;
number of questions&lt;br&gt;
time duration&lt;/p&gt;

&lt;p&gt;This is where the system becomes more structured. Instead of random question generation, it introduces controlled exam configuration.&lt;/p&gt;

&lt;p&gt;It also shows that the system is trying to balance:&lt;/p&gt;

&lt;p&gt;flexibility (multiple topics)&lt;br&gt;
control (fixed duration, levels)&lt;/p&gt;

&lt;p&gt;This reduces randomness and makes the exam predictable.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Exam Start Screen&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx0e4h33nn5q03ghkfsn4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx0e4h33nn5q03ghkfsn4.png" alt=" " width="800" height="441"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This screen shows instructions before the exam starts. It includes rules like:&lt;/p&gt;

&lt;p&gt;time limits&lt;br&gt;
no refresh&lt;br&gt;
answers cannot be changed&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Mathematics – Easy&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxab93xln3plqm87athug.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxab93xln3plqm87athug.png" alt=" " width="800" height="452"&gt;&lt;/a&gt;&lt;br&gt;
This screen shows basic math questions like addition and multiplication. At first glance, it looks like a normal quiz, but behind the scenes the questions are generated using AI based on the selected subject (Mathematics) and difficulty (Easy).&lt;/p&gt;

&lt;p&gt;When the user selects the topic, the system sends a prompt to the AI like:&lt;/p&gt;

&lt;p&gt;“Generate easy-level math questions with multiple choice answers in a fixed format.”&lt;/p&gt;

&lt;p&gt;AI returns a response, but the system does not directly display it. Instead, it validates the structure:&lt;/p&gt;

&lt;p&gt;ensures each question has exactly 4 options&lt;br&gt;
checks formatting consistency&lt;br&gt;
extracts the correct answer&lt;/p&gt;

&lt;p&gt;For math questions specifically, the system can also verify correctness deterministically. For example:&lt;/p&gt;

&lt;p&gt;15 + 27 is recalculated by the system&lt;br&gt;
the correct answer is confirmed before being stored&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Science – Medium&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvqv1llfcenaw0xgxkgbq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvqv1llfcenaw0xgxkgbq.png" alt=" " width="800" height="451"&gt;&lt;/a&gt;&lt;br&gt;
In this screen, the questions are conceptual, like chemical symbols or physics facts.&lt;/p&gt;

&lt;p&gt;Here, the system relies more on AI knowledge, but still enforces:&lt;/p&gt;

&lt;p&gt;structured question format&lt;br&gt;
single correct answer&lt;br&gt;
valid option set&lt;/p&gt;

&lt;p&gt;Since these are not numeric, the system cannot “calculate” answers the same way as math. Instead, it:&lt;/p&gt;

&lt;p&gt;cross-checks answer format&lt;br&gt;
ensures only one correct answer is marked&lt;br&gt;
optionally re-prompts AI for clarification if response is ambiguous&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Programming – Hard&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyi3cyqw86sb6cpcr1ydm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyi3cyqw86sb6cpcr1ydm.png" alt=" " width="800" height="448"&gt;&lt;/a&gt;&lt;br&gt;
This screen shows advanced questions like time complexity and data structures.&lt;/p&gt;

&lt;p&gt;Here, the system uses AI to generate:&lt;/p&gt;

&lt;p&gt;domain-specific questions&lt;br&gt;
appropriate difficulty&lt;br&gt;
relevant answer choices&lt;/p&gt;

&lt;p&gt;To improve reliability, the system:&lt;/p&gt;

&lt;p&gt;enforces known patterns (e.g., Big-O notation format)&lt;br&gt;
validates option consistency&lt;br&gt;
ensures only one correct answer is selected&lt;/p&gt;

&lt;p&gt;For some questions, the system can also apply rule-based validation, like:&lt;/p&gt;

&lt;p&gt;valid complexity values (O(n), O(log n), etc.)&lt;br&gt;
known correct answers for standard problems&lt;/p&gt;

&lt;p&gt;This is a small but important part. It shows that the system is thinking about real-world exam conditions, not just generating questions.&lt;/p&gt;

&lt;p&gt;The “Start Exam Now” action clearly separates setup from execution, which is good design for flow control.&lt;/p&gt;

&lt;p&gt;Every AI response goes through a validation layer before it is used. That means checking structure, fixing formatting issues, ensuring required fields exist, and making sure the output is consistent. So instead of just taking what AI gives, the system adjusts it into something predictable. In simple terms, AI suggests, but the system decides.&lt;/p&gt;

&lt;p&gt;The system is designed to keep the experience simple and clear for students. Every question follows the same format, with the same number of options and a consistent layout, so students don’t get confused or distracted by changing structures.&lt;/p&gt;

&lt;p&gt;When a student clicks on “Show Answer”, it’s not just revealing whatever the AI generated. The answer has already been carefully processed by the system. It is first extracted from the AI response, then validated to make sure it is correct and properly formatted. For numerical questions, the system can even recheck the calculation before showing the answer.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjasqjop4jbyf2kltf6wj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjasqjop4jbyf2kltf6wj.png" alt=" " width="800" height="664"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This way, students can trust that the answer they see is accurate, consistent, and helpful for learning—not just a random AI output.&lt;/p&gt;

&lt;p&gt;This small change made a big difference. The system became much more stable. The outputs were consistent. The same input would lead to similar structure every time. It stopped feeling like a demo and started behaving more like something usable.&lt;/p&gt;

&lt;p&gt;I also kept the system intentionally simple. No heavy UI, no complex setup. Just a Java JAR with an in-memory database. You can run it locally and try it out. The goal was not to build a full product, but to test this idea of combining AI with strict validation.&lt;/p&gt;

&lt;p&gt;I’m sharing this because I keep seeing the same pattern everywhere. AI-first systems look great initially, but small inconsistencies show up later and cause real problems. Not big failures, just small ones that are hard to trace. Adding a control layer seems boring, but it makes the system reliable.&lt;/p&gt;

&lt;p&gt;If you’ve worked on something similar—AI-generated data, exam systems, or validation layers—I’d be interested to hear how you handled it. Did you keep improving prompts, or did you add some kind of control mechanism?&lt;/p&gt;

&lt;p&gt;I’ve put the project here:&lt;br&gt;
GitHub: &lt;a href="https://github.com/swapneswarsundarray/ai-assisted-exam" rel="noopener noreferrer"&gt;https://github.com/swapneswarsundarray/ai-assisted-exam&lt;/a&gt;&lt;br&gt;
Feeback: &lt;a href="https://swapneswarsundarray.github.io/ai-assisted-exam/feedback.html" rel="noopener noreferrer"&gt;https://swapneswarsundarray.github.io/ai-assisted-exam/feedback.html&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Still early, still evolving. If this sounds interesting, feel free to try it out or contribute. Would be good to build this with more real-world input.&lt;/p&gt;

&lt;p&gt;In the end, I don’t think AI replaces systems. It just becomes one part of it. The rest is still structure, validation, and control. That’s where things actually start working properly.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Will tools like ChatGPT replace developers?</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Thu, 30 Apr 2026 20:11:14 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/will-tools-like-chatgpt-replace-developers-55f6</link>
      <guid>https://forem.com/swapneswar_sundarray/will-tools-like-chatgpt-replace-developers-55f6</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fynjmvua7epufczqk55ib.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fynjmvua7epufczqk55ib.png" alt=" " width="800" height="640"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;We’ve been hearing this question a lot &lt;br&gt;
&lt;em&gt;“Will ChatGPT replace developers?”&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;After working with AI in real projects, my answer is: &lt;strong&gt;no… but it will change how we work.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;At first glance, it’s easy to feel like it might. ChatGPT can write code, explain concepts, fix bugs, and even generate full functions in seconds. For small tasks, it’s incredibly useful. It saves time, reduces effort, and helps you move faster.&lt;/p&gt;

&lt;p&gt;But once you step into real systems, things look very different.&lt;/p&gt;

&lt;p&gt;In real-world environments, nothing is clean. Requirements change halfway through. Systems depend on other systems that don’t always behave the way you expect. Inputs are messy. Failures happen at the worst possible time. And sometimes, the problem itself isn’t even clearly defined.&lt;/p&gt;

&lt;p&gt;This is where developer skill actually matters.&lt;/p&gt;

&lt;p&gt;Writing code is just one part of the job. A good developer spends more time:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;understanding how systems behave&lt;/li&gt;
&lt;li&gt;making decisions when things are unclear&lt;/li&gt;
&lt;li&gt;designing for failure, not just success&lt;/li&gt;
&lt;li&gt;balancing trade-offs between speed, cost, and reliability&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These are things AI doesn’t really “own.” It can assist, but it doesn’t take responsibility.&lt;/p&gt;

&lt;p&gt;I’ve personally seen cases where AI-generated code looked perfect at first, but didn’t hold up when integrated into a larger system. It missed edge cases, didn’t consider dependencies, or made assumptions that weren’t true in production. Fixing those issues still required human judgment and experience.&lt;/p&gt;

&lt;p&gt;That’s when it becomes clear—AI is powerful, but it’s not the system. It’s just a part of it.&lt;/p&gt;

&lt;p&gt;What’s changing is not the need for developers, but the &lt;strong&gt;definition of a good developer&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Earlier, writing code quickly was a big advantage. Now, it’s becoming more about:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;asking the right questions&lt;/li&gt;
&lt;li&gt;guiding AI effectively&lt;/li&gt;
&lt;li&gt;validating outputs&lt;/li&gt;
&lt;li&gt;building systems that are stable and predictable&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In a way, developers are moving from “code writers” to “system thinkers.”&lt;/p&gt;

&lt;p&gt;The developers who will struggle are not the ones being replaced by AI—but the ones who refuse to adapt to it.&lt;/p&gt;

&lt;p&gt;And the ones who will stand out are those who learn how to use AI as a tool, without depending on it blindly.&lt;/p&gt;

&lt;p&gt;So no, ChatGPT won’t replace developers.&lt;/p&gt;

&lt;p&gt;But developers who learn how to work with tools like ChatGPT will definitely have an edge.&lt;/p&gt;

&lt;p&gt;And maybe that’s the real shift happening right now.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>career</category>
      <category>chatgpt</category>
      <category>discuss</category>
    </item>
    <item>
      <title>The Hidden Cost of AI Programming (and How to Use It Mindfully)</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Mon, 27 Apr 2026 20:59:35 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/the-hidden-cost-of-ai-programming-and-how-to-use-it-mindfully-5hde</link>
      <guid>https://forem.com/swapneswar_sundarray/the-hidden-cost-of-ai-programming-and-how-to-use-it-mindfully-5hde</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fprhasa7g0gbsok7nz66m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fprhasa7g0gbsok7nz66m.png" alt=" " width="800" height="640"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;“AI didn’t break your code. You just trusted it too much.”&lt;/p&gt;

&lt;p&gt;AI tools like GitHub Copilot and ChatGPT are changing how we write software. You type a comment… and suddenly a full function appears.&lt;/p&gt;

&lt;p&gt;Feels magical.&lt;br&gt;
Feels fast.&lt;br&gt;
Feels productive.&lt;/p&gt;

&lt;p&gt;But here’s the uncomfortable truth:&lt;br&gt;
&lt;strong&gt;AI can quietly make you a worse engineer if you’re not careful.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This isn’t anti-AI. I use it every day.&lt;br&gt;
This is about using AI like a senior engineer, not like autocomplete on steroids.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Bad Side of AI Programming&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;1. You Stop Thinking Deeply&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;AI gives you answers, not understanding.&lt;/p&gt;

&lt;p&gt;def calculate_discount(price, discount):&lt;br&gt;
    return price - (price * discount)&lt;/p&gt;

&lt;p&gt;Looks correct…&lt;/p&gt;

&lt;p&gt;But:&lt;/p&gt;

&lt;p&gt;What if discount = 20 instead of 0.20?&lt;br&gt;
What if price is negative?&lt;br&gt;
What if discount &amp;gt; 1?&lt;/p&gt;

&lt;p&gt;AI doesn’t validate business logic — it just generates code.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Context Blindness&lt;/strong&gt;&lt;br&gt;
AI doesn’t know your:&lt;/p&gt;

&lt;p&gt;system architecture&lt;br&gt;
scale requirements&lt;br&gt;
domain rules&lt;br&gt;
app.get('/users', async (req, res) =&amp;gt; {&lt;br&gt;
  const users = await db.getAllUsers();&lt;br&gt;
  res.json(users);&lt;br&gt;
});&lt;/p&gt;

&lt;p&gt;Looks clean.&lt;/p&gt;

&lt;p&gt;But in production:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No pagination&lt;/li&gt;
&lt;li&gt; No rate limiting&lt;/li&gt;
&lt;li&gt; No authentication&lt;/li&gt;
&lt;li&gt; No caching&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;You just created a production risk.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Confidently Wrong Code&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;AI sounds correct — even when it’s wrong.&lt;/p&gt;

&lt;p&gt;List list = Arrays.asList("a", "b", "c");&lt;br&gt;
list.add("d");  //  Runtime error&lt;/p&gt;

&lt;p&gt;Arrays.asList() returns a fixed-size list.&lt;/p&gt;

&lt;p&gt;AI misses subtle language rules.&lt;/p&gt;

&lt;p&gt;*&lt;em&gt;4. Technical Debt Explosion&lt;br&gt;
*&lt;/em&gt;&lt;br&gt;
AI optimizes for:&lt;/p&gt;

&lt;p&gt;“Make it work”&lt;/p&gt;

&lt;p&gt;Not:&lt;/p&gt;

&lt;p&gt;“Make it scalable and maintainable”&lt;/p&gt;

&lt;p&gt;function processOrder(order) {&lt;br&gt;
  if(order.type === 'A') { ... }&lt;br&gt;
  else if(order.type === 'B') { ... }&lt;br&gt;
  else if(order.type === 'C') { ... }&lt;br&gt;
}&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt; No design pattern&lt;/li&gt;
&lt;li&gt; No extensibility&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Hard to maintain&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;ol&gt;
&lt;li&gt;Debugging Skills Get Weaker&lt;/li&gt;
&lt;/ol&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If AI writes everything, what happens when things break?&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;You’re stuck debugging code you don’t fully understand.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The Mindful Way to Use AI
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;1. AI is powerful — if used correctly.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Use AI for Drafts, Not Decisions&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt; Bad:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;“AI wrote it, ship it”&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt; Good:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;“AI wrote it, now I validate it”&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Always Add Constraints&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Instead of:&lt;/p&gt;

&lt;p&gt;“write a user API”&lt;/p&gt;

&lt;p&gt;Say:&lt;/p&gt;

&lt;p&gt;“write a paginated, rate-limited, authenticated API with error handling”&lt;/p&gt;

&lt;p&gt;Example (Better API)&lt;br&gt;
`app.get('/users', async (req, res) =&amp;gt; {&lt;br&gt;
  const { page = 1, limit = 10 } = req.query;&lt;/p&gt;

&lt;p&gt;if (limit &amp;gt; 100) {&lt;br&gt;
    return res.status(400).json({ error: "Limit too high" });&lt;br&gt;
  }&lt;/p&gt;

&lt;p&gt;const users = await db.getUsersPaginated(page, limit);&lt;/p&gt;

&lt;p&gt;res.json({&lt;br&gt;
    page,&lt;br&gt;
    limit,&lt;br&gt;
    data: users&lt;br&gt;
  });&lt;br&gt;
});&lt;br&gt;
`&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Treat AI Like a Junior Developer&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Always:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;review the code&lt;/li&gt;
&lt;li&gt;question assumptions&lt;/li&gt;
&lt;li&gt;test edge cases&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;4. Ask AI “Why”, Not Just “What”&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Instead of:&lt;/p&gt;

&lt;p&gt;“give me code”&lt;/p&gt;

&lt;p&gt;Ask:&lt;/p&gt;

&lt;p&gt;“explain trade-offs, edge cases, and risks”&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5. Use AI for Repetitive Work&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Best use cases:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;boilerplate code&lt;/li&gt;
&lt;li&gt;test cases&lt;/li&gt;
&lt;li&gt;documentation&lt;/li&gt;
&lt;li&gt;refactoring suggestions&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Not for critical architecture decisions.&lt;/p&gt;

&lt;p&gt;AI is not the problem.&lt;/p&gt;

&lt;p&gt;Blind trust is.&lt;/p&gt;

&lt;p&gt;The best engineers don’t replace thinking with AI.&lt;br&gt;
They amplify thinking with AI.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>githubcopilot</category>
      <category>coding</category>
    </item>
    <item>
      <title>LangChain and LangGraph: Building Reliable Agentic AI Workflows</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Mon, 27 Apr 2026 02:00:15 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/langchain-and-langgraph-building-reliable-agentic-ai-workflows-34b4</link>
      <guid>https://forem.com/swapneswar_sundarray/langchain-and-langgraph-building-reliable-agentic-ai-workflows-34b4</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqt49yaxps3uzijxwpjyc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqt49yaxps3uzijxwpjyc.png" alt=" " width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  LangChain and LangGraph: Building Reliable Agentic AI Workflows
&lt;/h2&gt;

&lt;p&gt;Modern AI applications are no longer simple chatbot wrappers around an LLM.&lt;/p&gt;

&lt;p&gt;Real enterprise AI systems need to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;understand user intent&lt;/li&gt;
&lt;li&gt;retrieve relevant context&lt;/li&gt;
&lt;li&gt;call tools and APIs&lt;/li&gt;
&lt;li&gt;maintain state&lt;/li&gt;
&lt;li&gt;follow business rules&lt;/li&gt;
&lt;li&gt;validate outputs&lt;/li&gt;
&lt;li&gt;retry failed steps&lt;/li&gt;
&lt;li&gt;escalate risky decisions&lt;/li&gt;
&lt;li&gt;produce auditable results&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is where &lt;strong&gt;LangChain&lt;/strong&gt; and &lt;strong&gt;LangGraph&lt;/strong&gt; are useful.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;LangChain&lt;/strong&gt; provides building blocks for connecting LLMs with tools, prompts, retrievers, vector databases, APIs, and external systems.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;LangGraph&lt;/strong&gt; provides a graph-based orchestration layer for building stateful, multi-step, controllable AI workflows.&lt;/p&gt;

&lt;p&gt;In simple terms:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;LangChain connects the AI to capabilities.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
&lt;strong&gt;LangGraph controls how those capabilities are used.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  1. What Is LangChain?
&lt;/h2&gt;

&lt;p&gt;LangChain is a framework for building applications powered by large language models.&lt;/p&gt;

&lt;p&gt;It helps developers connect LLMs with external components such as:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;prompt templates&lt;/li&gt;
&lt;li&gt;tools&lt;/li&gt;
&lt;li&gt;APIs&lt;/li&gt;
&lt;li&gt;retrievers&lt;/li&gt;
&lt;li&gt;vector stores&lt;/li&gt;
&lt;li&gt;document loaders&lt;/li&gt;
&lt;li&gt;output parsers&lt;/li&gt;
&lt;li&gt;memory&lt;/li&gt;
&lt;li&gt;agents&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;A typical LangChain-based application may look like this:&lt;/p&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;
text
User Query
   |
   v
Prompt Template
   |
   v
Retriever / Tool
   |
   v
LLM
   |
   v
Output Parser
   |
   v
Final Response
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

</description>
      <category>ai</category>
      <category>webdev</category>
      <category>programming</category>
      <category>computerscience</category>
    </item>
    <item>
      <title>AI Tools Ranked (Best to Worst) by Real-World Impact</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Sun, 26 Apr 2026 14:07:42 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/ai-tools-ranked-best-to-worst-by-real-world-impact-42j0</link>
      <guid>https://forem.com/swapneswar_sundarray/ai-tools-ranked-best-to-worst-by-real-world-impact-42j0</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq9m1uw4950ny2lio7jet.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq9m1uw4950ny2lio7jet.png" alt=" " width="800" height="1200"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  AI Tools Ranked (Best to Worst) by Real-World Impact
&lt;/h1&gt;

&lt;p&gt;There are hundreds of AI tools available today.&lt;/p&gt;

&lt;p&gt;Most demos look impressive.&lt;br&gt;&lt;br&gt;
Very few actually deliver impact in production.&lt;/p&gt;

&lt;p&gt;Instead of hype, this ranking is based on real-world impact.&lt;/p&gt;

&lt;h2&gt;
  
  
  Evaluation Criteria
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Production usability (can it be deployed)&lt;/li&gt;
&lt;li&gt;Reliability and consistency&lt;/li&gt;
&lt;li&gt;Time saved and ROI&lt;/li&gt;
&lt;li&gt;Integration capability&lt;/li&gt;
&lt;li&gt;Adoption in real teams&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Tier 1 - Highest Impact (Production-Ready)
&lt;/h1&gt;

&lt;h2&gt;
  
  
  1. ChatGPT (GPT-4/5)
&lt;/h2&gt;

&lt;p&gt;Best overall AI tool today.&lt;/p&gt;

&lt;p&gt;Where it performs well:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;System design and reasoning&lt;/li&gt;
&lt;li&gt;Code generation and debugging&lt;/li&gt;
&lt;li&gt;Writing and analysis&lt;/li&gt;
&lt;li&gt;Automation workflows&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Impact:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;3 to 10x productivity improvement&lt;/li&gt;
&lt;li&gt;Faster iteration cycles&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Limitation:&lt;br&gt;
Not perfect, but the most versatile tool in production.&lt;/p&gt;

&lt;h2&gt;
  
  
  2. GitHub Copilot
&lt;/h2&gt;

&lt;p&gt;Best for day-to-day coding.&lt;/p&gt;

&lt;p&gt;Where it performs well:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Inline code suggestions&lt;/li&gt;
&lt;li&gt;Boilerplate generation&lt;/li&gt;
&lt;li&gt;Refactoring assistance&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Impact:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;30 to 50 percent faster coding&lt;/li&gt;
&lt;li&gt;Reduced context switching&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Limitations:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Weak in architecture-level reasoning&lt;/li&gt;
&lt;li&gt;May generate incorrect logic silently&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  3. Claude
&lt;/h2&gt;

&lt;p&gt;Best for long-context reasoning.&lt;/p&gt;

&lt;p&gt;Where it performs well:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Large documents&lt;/li&gt;
&lt;li&gt;Deep reasoning tasks&lt;/li&gt;
&lt;li&gt;Safer responses&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Impact:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Strong for research and analysis workflows&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Limitations:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Not as strong for coding as Copilot&lt;/li&gt;
&lt;li&gt;Slower iteration in some cases&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Tier 2 — High Impact (Specialized Use)
&lt;/h1&gt;

&lt;h2&gt;
  
  
  4. LangChain and LLM Frameworks
&lt;/h2&gt;

&lt;p&gt;Backbone of AI applications.&lt;/p&gt;

&lt;p&gt;Where they perform well:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Orchestration&lt;/li&gt;
&lt;li&gt;Retrieval-augmented generation pipelines&lt;/li&gt;
&lt;li&gt;Agent workflows&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Impact:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Enables production AI systems&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Limitation:&lt;br&gt;
Powerful but requires engineering effort.&lt;/p&gt;

&lt;h2&gt;
  
  
  5. Perplexity AI
&lt;/h2&gt;

&lt;p&gt;Best AI-powered search.&lt;/p&gt;

&lt;p&gt;Where it performs well:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Research&lt;/li&gt;
&lt;li&gt;Citation-backed answers&lt;/li&gt;
&lt;li&gt;Quick exploration&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Impact:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Replaces traditional search in many workflows&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Limitation:&lt;br&gt;
Not ideal for deep system tasks.&lt;/p&gt;

&lt;h2&gt;
  
  
  6. Midjourney and DALL-E
&lt;/h2&gt;

&lt;p&gt;Best for image generation.&lt;/p&gt;

&lt;p&gt;Where they perform well:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Design&lt;/li&gt;
&lt;li&gt;Marketing content&lt;/li&gt;
&lt;li&gt;Creative assets&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Impact:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Reduces design cost and time&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Limitation:&lt;br&gt;
Limited use for engineering workflows.&lt;/p&gt;

&lt;h1&gt;
  
  
  Tier 3 — Moderate Impact (Context Dependent)
&lt;/h1&gt;

&lt;h2&gt;
  
  
  7. AutoGPT and Agent Tools
&lt;/h2&gt;

&lt;p&gt;High potential but low reliability.&lt;/p&gt;

&lt;p&gt;Where they perform well:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Multi-step automation&lt;/li&gt;
&lt;li&gt;Experimentation&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Reality:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Still unstable&lt;/li&gt;
&lt;li&gt;Hard to control&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Impact:&lt;br&gt;
More experimental than production-ready.&lt;/p&gt;

&lt;h2&gt;
  
  
  8. AI Coding Alternatives
&lt;/h2&gt;

&lt;p&gt;Examples include tools like Ghostwriter.&lt;/p&gt;

&lt;p&gt;Where they perform well:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Beginner-friendly environments&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Limitations:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Less mature ecosystem&lt;/li&gt;
&lt;li&gt;Lower accuracy&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Tier 4 — Low Impact (Overhyped)
&lt;/h1&gt;

&lt;h2&gt;
  
  
  9. No-Code AI Builders
&lt;/h2&gt;

&lt;p&gt;Marketed as building apps without coding.&lt;/p&gt;

&lt;p&gt;Reality:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Limited flexibility&lt;/li&gt;
&lt;li&gt;Difficult to scale&lt;/li&gt;
&lt;li&gt;Not production-ready&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  10. Generic AI Wrappers
&lt;/h2&gt;

&lt;p&gt;Simple interfaces over existing APIs.&lt;/p&gt;

&lt;p&gt;Reality:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No real differentiation&lt;/li&gt;
&lt;li&gt;Easily replaceable&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  The Real Insight
&lt;/h1&gt;

&lt;p&gt;Most people ask:&lt;/p&gt;

&lt;p&gt;Which AI tool is best?&lt;/p&gt;

&lt;p&gt;The better question is:&lt;/p&gt;

&lt;p&gt;Where does AI fit into your system?&lt;/p&gt;

&lt;h1&gt;
  
  
  What Actually Works in Production
&lt;/h1&gt;

&lt;h2&gt;
  
  
  What fails
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;LLM-only systems&lt;/li&gt;
&lt;li&gt;Lack of architecture&lt;/li&gt;
&lt;li&gt;No validation layer&lt;/li&gt;
&lt;li&gt;No monitoring&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What works
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Hybrid systems combining code and LLMs&lt;/li&gt;
&lt;li&gt;Strong data pipelines&lt;/li&gt;
&lt;li&gt;Clear business use cases&lt;/li&gt;
&lt;li&gt;Monitoring and lifecycle management&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Final Ranking Summary
&lt;/h1&gt;

&lt;p&gt;Tier 1 (Game Changers)&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;ChatGPT&lt;/li&gt;
&lt;li&gt;GitHub Copilot&lt;/li&gt;
&lt;li&gt;Claude&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Tier 2 (Specialized Tools)&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;LangChain&lt;/li&gt;
&lt;li&gt;Perplexity&lt;/li&gt;
&lt;li&gt;Midjourney&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Tier 3 (Experimental)&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;AutoGPT&lt;/li&gt;
&lt;li&gt;Other coding tools&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Tier 4 (Overhyped)&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;No-code AI builders&lt;/li&gt;
&lt;li&gt;Generic wrappers&lt;/li&gt;
&lt;/ol&gt;

&lt;h1&gt;
  
  
  Final Thought
&lt;/h1&gt;

&lt;p&gt;AI tools do not create impact.&lt;/p&gt;

&lt;p&gt;Systems do.&lt;/p&gt;

&lt;p&gt;The teams succeeding with AI are not using better tools.&lt;br&gt;&lt;br&gt;
They are using tools more effectively.&lt;/p&gt;

&lt;h1&gt;
  
  
  Tags
&lt;/h1&gt;

&lt;p&gt;ai&lt;br&gt;&lt;br&gt;
machinelearning&lt;br&gt;&lt;br&gt;
developer&lt;br&gt;&lt;br&gt;
productivity&lt;br&gt;&lt;br&gt;
softwareengineering&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>developer</category>
      <category>productivity</category>
    </item>
    <item>
      <title>Banks and AI</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Sun, 26 Apr 2026 00:43:12 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/80-of-ai-projects-in-banks-fail-heres-why-and-how-we-fixed-it-5bab</link>
      <guid>https://forem.com/swapneswar_sundarray/80-of-ai-projects-in-banks-fail-heres-why-and-how-we-fixed-it-5bab</guid>
      <description>&lt;p&gt;Banks invested billions in AI.&lt;/p&gt;

&lt;p&gt;Fraud detection.&lt;br&gt;&lt;br&gt;
Credit scoring.&lt;br&gt;&lt;br&gt;
Customer experience.&lt;br&gt;&lt;br&gt;
Risk modeling.  &lt;/p&gt;

&lt;p&gt;The promise was massive.&lt;/p&gt;

&lt;p&gt;But here’s the uncomfortable truth:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Most AI projects never make it to production.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Not because the models don’t work.&lt;/p&gt;

&lt;p&gt;But because &lt;strong&gt;everything around them fails.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;From my experience building AI systems in banking, the pattern is always the same.&lt;/p&gt;

&lt;h1&gt;
  
  
  The Real Problem
&lt;/h1&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;AI doesn’t fail at the model level.&lt;br&gt;&lt;br&gt;
It fails at the system level.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Let’s break it down.&lt;/p&gt;

&lt;h1&gt;
  
  
  Where AI Projects Break
&lt;/h1&gt;

&lt;h2&gt;
  
  
  1. The “Pilot Trap”
&lt;/h2&gt;

&lt;p&gt;Every bank has this story:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Build a model
&lt;/li&gt;
&lt;li&gt;It works in a demo
&lt;/li&gt;
&lt;li&gt;Leadership is impressed
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And then… silence.&lt;/p&gt;

&lt;h3&gt;
  
  
  Why?
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;No production infrastructure
&lt;/li&gt;
&lt;li&gt;No ownership after POC
&lt;/li&gt;
&lt;li&gt;No integration roadmap
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Result:&lt;br&gt;&lt;br&gt;
&lt;strong&gt;Great demos. Zero impact.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  2. Legacy Systems Kill Momentum
&lt;/h2&gt;

&lt;p&gt;AI needs:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Clean data
&lt;/li&gt;
&lt;li&gt;Real-time access
&lt;/li&gt;
&lt;li&gt;APIs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Banks often have:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Data silos
&lt;/li&gt;
&lt;li&gt;Batch pipelines
&lt;/li&gt;
&lt;li&gt;Fragile integrations
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;AI becomes a &lt;strong&gt;side layer&lt;/strong&gt;, not core infrastructure.&lt;/p&gt;

&lt;h2&gt;
  
  
  3. Data Reality Check
&lt;/h2&gt;

&lt;p&gt;Everyone assumes:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“We have years of data—we’re ready.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Reality:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Missing fields
&lt;/li&gt;
&lt;li&gt;Inconsistent formats
&lt;/li&gt;
&lt;li&gt;Historical bias
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Garbage in → Garbage out&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  4. Compliance Slows Everything
&lt;/h2&gt;

&lt;p&gt;Banking isn’t a startup.&lt;/p&gt;

&lt;p&gt;Every model must be:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Explainable
&lt;/li&gt;
&lt;li&gt;Auditable
&lt;/li&gt;
&lt;li&gt;Fair
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  What happens:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Models get rejected late
&lt;/li&gt;
&lt;li&gt;Legal blocks rollout
&lt;/li&gt;
&lt;li&gt;Risk teams force simplification
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Speed → Gone&lt;br&gt;&lt;br&gt;
Momentum → Gone  &lt;/p&gt;

&lt;h2&gt;
  
  
  5. Business vs Tech Misalignment
&lt;/h2&gt;

&lt;p&gt;AI teams build models.&lt;/p&gt;

&lt;p&gt;Business teams expect ROI.&lt;/p&gt;

&lt;p&gt;But:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No shared KPIs
&lt;/li&gt;
&lt;li&gt;No domain alignment
&lt;/li&gt;
&lt;li&gt;No clear success metric
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Misalignment = failure.&lt;/p&gt;

&lt;h2&gt;
  
  
  6. No MLOps = No Product
&lt;/h2&gt;

&lt;p&gt;Most teams stop at:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Model trained”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;But production needs:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Monitoring
&lt;/li&gt;
&lt;li&gt;Drift detection
&lt;/li&gt;
&lt;li&gt;Retraining
&lt;/li&gt;
&lt;li&gt;Versioning
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Without MLOps, models decay fast.&lt;/p&gt;

&lt;h1&gt;
  
  
  The Reality (Simple View)
&lt;/h1&gt;

&lt;p&gt;Typical AI Project Flow in Banks:&lt;/p&gt;

&lt;p&gt;Idea → Pilot → Demo → Approval → Stuck → Dead&lt;/p&gt;

&lt;p&gt;What Actually Works:&lt;/p&gt;

&lt;p&gt;Idea → Data → Architecture → Integration → Deployment → Monitoring → Impact&lt;/p&gt;

&lt;h1&gt;
  
  
  What Actually Worked (In Production)
&lt;/h1&gt;

&lt;p&gt;Here’s what changed everything for us:&lt;/p&gt;

&lt;h2&gt;
  
  
  1. Start With Business, Not Models
&lt;/h2&gt;

&lt;p&gt;Instead of:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Let’s build AI”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;We asked:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“What business problem matters?”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h3&gt;
  
  
  Examples:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Reduce fraud loss by X%
&lt;/li&gt;
&lt;li&gt;Improve loan approval speed
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;AI became outcome-driven, not experiment-driven.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  2. Fix Data Before Models
&lt;/h2&gt;

&lt;p&gt;We invested in:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Clean pipelines
&lt;/li&gt;
&lt;li&gt;Standard schemas
&lt;/li&gt;
&lt;li&gt;Strong governance
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Data became usable and reliable.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  3. Build for Production From Day One
&lt;/h2&gt;

&lt;p&gt;No throwaway pilots.&lt;/p&gt;

&lt;p&gt;Every model had:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;API endpoints
&lt;/li&gt;
&lt;li&gt;Integration plan
&lt;/li&gt;
&lt;li&gt;Deployment path
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;If it can’t scale, don’t build it.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  4. Bring Compliance Early
&lt;/h2&gt;

&lt;p&gt;Instead of late approvals:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Risk teams involved from day one
&lt;/li&gt;
&lt;li&gt;Explainability built-in
&lt;/li&gt;
&lt;li&gt;Documentation automated
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Compliance became a partner, not a blocker.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  5. Build Cross-Functional Teams
&lt;/h2&gt;

&lt;p&gt;We combined:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Engineers
&lt;/li&gt;
&lt;li&gt;Data scientists
&lt;/li&gt;
&lt;li&gt;Domain experts
&lt;/li&gt;
&lt;li&gt;Risk &amp;amp; legal
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Decisions got faster, clearer, and aligned.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  6. Invest in MLOps
&lt;/h2&gt;

&lt;p&gt;We implemented:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;CI/CD for models
&lt;/li&gt;
&lt;li&gt;Monitoring dashboards
&lt;/li&gt;
&lt;li&gt;Automated retraining
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Models stayed reliable in production.&lt;/strong&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  The Outcome
&lt;/h1&gt;

&lt;ul&gt;
&lt;li&gt;Faster deployments
&lt;/li&gt;
&lt;li&gt;Lower failure rates
&lt;/li&gt;
&lt;li&gt;Higher reliability
&lt;/li&gt;
&lt;li&gt;Real business impact
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Most importantly:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;AI became a capability — not an experiment.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h1&gt;
  
  
  Final Thought
&lt;/h1&gt;

&lt;p&gt;AI in banking isn’t failing because it’s too complex.&lt;/p&gt;

&lt;p&gt;It’s failing because:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Organizations treat AI like a project.&lt;br&gt;&lt;br&gt;
Not like infrastructure.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Until that changes…&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Failure rates won’t.&lt;/strong&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>fintech</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>Stop Building One Giant Prompt: A Better Way to Design LLM Systems</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Sat, 25 Apr 2026 18:56:55 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/stop-building-one-giant-prompt-a-better-way-to-design-llm-systems-3g4n</link>
      <guid>https://forem.com/swapneswar_sundarray/stop-building-one-giant-prompt-a-better-way-to-design-llm-systems-3g4n</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5d2o32cdgpkmqs5nq2t9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5d2o32cdgpkmqs5nq2t9.png" alt=" " width="800" height="533"&gt;&lt;/a&gt;## Most early LLM apps start the same way:&lt;/p&gt;

&lt;p&gt;“Let’s just put everything into one prompt and let the model handle it.”&lt;/p&gt;

&lt;p&gt;So we write a prompt that tries to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;validate input&lt;/li&gt;
&lt;li&gt;transform data&lt;/li&gt;
&lt;li&gt;generate output&lt;/li&gt;
&lt;li&gt;summarize&lt;/li&gt;
&lt;li&gt;add reasoning&lt;/li&gt;
&lt;li&gt;handle edge cases&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;…and somehow do it all in one call.&lt;/p&gt;

&lt;p&gt;It works—until it doesn’t.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Problem with “God Prompts”
&lt;/h2&gt;

&lt;p&gt;As the prompt grows:&lt;/p&gt;

&lt;p&gt;Instructions start conflicting&lt;br&gt;
Context becomes noisy&lt;br&gt;
Accuracy drops&lt;br&gt;
Outputs become inconsistent&lt;/p&gt;

&lt;p&gt;You end up with:&lt;/p&gt;

&lt;p&gt;a very expensive confusion engine&lt;/p&gt;

&lt;p&gt;I’ve hit this multiple times while building AI systems.&lt;/p&gt;

&lt;h2&gt;
  
  
  What’s Actually Happening
&lt;/h2&gt;

&lt;p&gt;You’re increasing what I call LLM cognitive load.&lt;/p&gt;

&lt;p&gt;The more responsibilities you push into a single call:&lt;/p&gt;

&lt;p&gt;the harder it is for the model to prioritize&lt;br&gt;
the easier it is to miss instructions&lt;br&gt;
the more likely it is to hallucinate&lt;/p&gt;

&lt;p&gt;Even with better models, this pattern doesn’t go away.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;A Better Approach: Think Like a System Designer&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Instead of one big prompt, break the problem into smaller, focused steps.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;Validate + transform + summarize + generate + explain everything&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Do this:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Validation step (code)&lt;/li&gt;
&lt;li&gt;Extraction step (LLM)&lt;/li&gt;
&lt;li&gt;Transformation step (code or LLM)&lt;/li&gt;
&lt;li&gt;Generation step (LLM)&lt;/li&gt;
&lt;li&gt;Formatting step (code)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Use the Right Tool for the Right Job
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Let code handle:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;validation&lt;/li&gt;
&lt;li&gt;parsing&lt;/li&gt;
&lt;li&gt;routing&lt;/li&gt;
&lt;li&gt;rules&lt;/li&gt;
&lt;li&gt;state&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Let LLM handle:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;reasoning&lt;/li&gt;
&lt;li&gt;interpretation&lt;/li&gt;
&lt;li&gt;summarization&lt;/li&gt;
&lt;li&gt;ambiguity&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Treat LLM Calls Like Microservices&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This mindset shift helped me a lot:&lt;/p&gt;

&lt;p&gt;Each LLM call should have a single responsibility&lt;/p&gt;

&lt;p&gt;Small input&lt;br&gt;
Clear task&lt;br&gt;
Predictable output&lt;/p&gt;

&lt;p&gt;Then orchestrate them together.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-World Example&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;While working on API automation systems, we initially tried:&lt;/p&gt;

&lt;p&gt;one prompt to validate specs + generate APIs + create mock data&lt;/p&gt;

&lt;p&gt;It became unstable very quickly.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Splitting it into:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;validation module&lt;/li&gt;
&lt;li&gt;generation module&lt;/li&gt;
&lt;li&gt;mock data module&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;made the system far more reliable.&lt;/p&gt;

&lt;p&gt;LLMs are powerful—but they’re not a replacement for system design.&lt;/p&gt;

&lt;p&gt;“Just add AI” is not an architecture pattern.&lt;/p&gt;

&lt;p&gt;Design your system first.&lt;br&gt;
Then use AI where it actually adds value.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>promptengineering</category>
      <category>systemdesign</category>
    </item>
  </channel>
</rss>
