<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Swapneswar Sundar Ray</title>
    <description>The latest articles on Forem by Swapneswar Sundar Ray (@swapneswar_sundarray).</description>
    <link>https://forem.com/swapneswar_sundarray</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3897931%2F73304f54-efbd-4435-a6ee-0ed0b2ba2da5.png</url>
      <title>Forem: Swapneswar Sundar Ray</title>
      <link>https://forem.com/swapneswar_sundarray</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/swapneswar_sundarray"/>
    <language>en</language>
    <item>
      <title>80% of AI Projects in Banks Fail - Here’s Why (And How We Fixed It)</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Sun, 26 Apr 2026 00:43:12 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/80-of-ai-projects-in-banks-fail-heres-why-and-how-we-fixed-it-5bab</link>
      <guid>https://forem.com/swapneswar_sundarray/80-of-ai-projects-in-banks-fail-heres-why-and-how-we-fixed-it-5bab</guid>
      <description>&lt;p&gt;Banks invested billions in AI.&lt;/p&gt;

&lt;p&gt;Fraud detection.&lt;br&gt;&lt;br&gt;
Credit scoring.&lt;br&gt;&lt;br&gt;
Customer experience.&lt;br&gt;&lt;br&gt;
Risk modeling.  &lt;/p&gt;

&lt;p&gt;The promise was massive.&lt;/p&gt;

&lt;p&gt;But here’s the uncomfortable truth:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Most AI projects never make it to production.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Not because the models don’t work.&lt;/p&gt;

&lt;p&gt;But because &lt;strong&gt;everything around them fails.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;From my experience building AI systems in banking, the pattern is always the same.&lt;/p&gt;

&lt;h1&gt;
  
  
  The Real Problem
&lt;/h1&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;AI doesn’t fail at the model level.&lt;br&gt;&lt;br&gt;
It fails at the system level.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Let’s break it down.&lt;/p&gt;

&lt;h1&gt;
  
  
  Where AI Projects Break
&lt;/h1&gt;

&lt;h2&gt;
  
  
  1. The “Pilot Trap”
&lt;/h2&gt;

&lt;p&gt;Every bank has this story:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Build a model
&lt;/li&gt;
&lt;li&gt;It works in a demo
&lt;/li&gt;
&lt;li&gt;Leadership is impressed
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And then… silence.&lt;/p&gt;

&lt;h3&gt;
  
  
  Why?
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;No production infrastructure
&lt;/li&gt;
&lt;li&gt;No ownership after POC
&lt;/li&gt;
&lt;li&gt;No integration roadmap
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Result:&lt;br&gt;&lt;br&gt;
&lt;strong&gt;Great demos. Zero impact.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  2. Legacy Systems Kill Momentum
&lt;/h2&gt;

&lt;p&gt;AI needs:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Clean data
&lt;/li&gt;
&lt;li&gt;Real-time access
&lt;/li&gt;
&lt;li&gt;APIs
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Banks often have:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Data silos
&lt;/li&gt;
&lt;li&gt;Batch pipelines
&lt;/li&gt;
&lt;li&gt;Fragile integrations
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;AI becomes a &lt;strong&gt;side layer&lt;/strong&gt;, not core infrastructure.&lt;/p&gt;

&lt;h2&gt;
  
  
  3. Data Reality Check
&lt;/h2&gt;

&lt;p&gt;Everyone assumes:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“We have years of data—we’re ready.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Reality:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Missing fields
&lt;/li&gt;
&lt;li&gt;Inconsistent formats
&lt;/li&gt;
&lt;li&gt;Historical bias
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Garbage in → Garbage out&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  4. Compliance Slows Everything
&lt;/h2&gt;

&lt;p&gt;Banking isn’t a startup.&lt;/p&gt;

&lt;p&gt;Every model must be:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Explainable
&lt;/li&gt;
&lt;li&gt;Auditable
&lt;/li&gt;
&lt;li&gt;Fair
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  What happens:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Models get rejected late
&lt;/li&gt;
&lt;li&gt;Legal blocks rollout
&lt;/li&gt;
&lt;li&gt;Risk teams force simplification
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Speed → Gone&lt;br&gt;&lt;br&gt;
Momentum → Gone  &lt;/p&gt;

&lt;h2&gt;
  
  
  5. Business vs Tech Misalignment
&lt;/h2&gt;

&lt;p&gt;AI teams build models.&lt;/p&gt;

&lt;p&gt;Business teams expect ROI.&lt;/p&gt;

&lt;p&gt;But:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No shared KPIs
&lt;/li&gt;
&lt;li&gt;No domain alignment
&lt;/li&gt;
&lt;li&gt;No clear success metric
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Misalignment = failure.&lt;/p&gt;

&lt;h2&gt;
  
  
  6. No MLOps = No Product
&lt;/h2&gt;

&lt;p&gt;Most teams stop at:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Model trained”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;But production needs:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Monitoring
&lt;/li&gt;
&lt;li&gt;Drift detection
&lt;/li&gt;
&lt;li&gt;Retraining
&lt;/li&gt;
&lt;li&gt;Versioning
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Without MLOps, models decay fast.&lt;/p&gt;

&lt;h1&gt;
  
  
  The Reality (Simple View)
&lt;/h1&gt;

&lt;p&gt;Typical AI Project Flow in Banks:&lt;/p&gt;

&lt;p&gt;Idea → Pilot → Demo → Approval → Stuck → Dead&lt;/p&gt;

&lt;p&gt;What Actually Works:&lt;/p&gt;

&lt;p&gt;Idea → Data → Architecture → Integration → Deployment → Monitoring → Impact&lt;/p&gt;

&lt;h1&gt;
  
  
  What Actually Worked (In Production)
&lt;/h1&gt;

&lt;p&gt;Here’s what changed everything for us:&lt;/p&gt;

&lt;h2&gt;
  
  
  1. Start With Business, Not Models
&lt;/h2&gt;

&lt;p&gt;Instead of:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Let’s build AI”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;We asked:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“What business problem matters?”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h3&gt;
  
  
  Examples:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Reduce fraud loss by X%
&lt;/li&gt;
&lt;li&gt;Improve loan approval speed
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;AI became outcome-driven, not experiment-driven.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  2. Fix Data Before Models
&lt;/h2&gt;

&lt;p&gt;We invested in:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Clean pipelines
&lt;/li&gt;
&lt;li&gt;Standard schemas
&lt;/li&gt;
&lt;li&gt;Strong governance
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Data became usable and reliable.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  3. Build for Production From Day One
&lt;/h2&gt;

&lt;p&gt;No throwaway pilots.&lt;/p&gt;

&lt;p&gt;Every model had:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;API endpoints
&lt;/li&gt;
&lt;li&gt;Integration plan
&lt;/li&gt;
&lt;li&gt;Deployment path
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;If it can’t scale, don’t build it.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  4. Bring Compliance Early
&lt;/h2&gt;

&lt;p&gt;Instead of late approvals:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Risk teams involved from day one
&lt;/li&gt;
&lt;li&gt;Explainability built-in
&lt;/li&gt;
&lt;li&gt;Documentation automated
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Compliance became a partner, not a blocker.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  5. Build Cross-Functional Teams
&lt;/h2&gt;

&lt;p&gt;We combined:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Engineers
&lt;/li&gt;
&lt;li&gt;Data scientists
&lt;/li&gt;
&lt;li&gt;Domain experts
&lt;/li&gt;
&lt;li&gt;Risk &amp;amp; legal
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Decisions got faster, clearer, and aligned.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  6. Invest in MLOps
&lt;/h2&gt;

&lt;p&gt;We implemented:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;CI/CD for models
&lt;/li&gt;
&lt;li&gt;Monitoring dashboards
&lt;/li&gt;
&lt;li&gt;Automated retraining
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Models stayed reliable in production.&lt;/strong&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  The Outcome
&lt;/h1&gt;

&lt;ul&gt;
&lt;li&gt;Faster deployments
&lt;/li&gt;
&lt;li&gt;Lower failure rates
&lt;/li&gt;
&lt;li&gt;Higher reliability
&lt;/li&gt;
&lt;li&gt;Real business impact
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Most importantly:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;AI became a capability — not an experiment.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h1&gt;
  
  
  Final Thought
&lt;/h1&gt;

&lt;p&gt;AI in banking isn’t failing because it’s too complex.&lt;/p&gt;

&lt;p&gt;It’s failing because:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Organizations treat AI like a project.&lt;br&gt;&lt;br&gt;
Not like infrastructure.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Until that changes…&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Failure rates won’t.&lt;/strong&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>fintech</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>Stop Building One Giant Prompt: A Better Way to Design LLM Systems</title>
      <dc:creator>Swapneswar Sundar Ray</dc:creator>
      <pubDate>Sat, 25 Apr 2026 18:56:55 +0000</pubDate>
      <link>https://forem.com/swapneswar_sundarray/stop-building-one-giant-prompt-a-better-way-to-design-llm-systems-3g4n</link>
      <guid>https://forem.com/swapneswar_sundarray/stop-building-one-giant-prompt-a-better-way-to-design-llm-systems-3g4n</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5d2o32cdgpkmqs5nq2t9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5d2o32cdgpkmqs5nq2t9.png" alt=" " width="800" height="533"&gt;&lt;/a&gt;## Most early LLM apps start the same way:&lt;/p&gt;

&lt;p&gt;“Let’s just put everything into one prompt and let the model handle it.”&lt;/p&gt;

&lt;p&gt;So we write a prompt that tries to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;validate input&lt;/li&gt;
&lt;li&gt;transform data&lt;/li&gt;
&lt;li&gt;generate output&lt;/li&gt;
&lt;li&gt;summarize&lt;/li&gt;
&lt;li&gt;add reasoning&lt;/li&gt;
&lt;li&gt;handle edge cases&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;…and somehow do it all in one call.&lt;/p&gt;

&lt;p&gt;It works—until it doesn’t.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Problem with “God Prompts”
&lt;/h2&gt;

&lt;p&gt;As the prompt grows:&lt;/p&gt;

&lt;p&gt;Instructions start conflicting&lt;br&gt;
Context becomes noisy&lt;br&gt;
Accuracy drops&lt;br&gt;
Outputs become inconsistent&lt;/p&gt;

&lt;p&gt;You end up with:&lt;/p&gt;

&lt;p&gt;a very expensive confusion engine&lt;/p&gt;

&lt;p&gt;I’ve hit this multiple times while building AI systems.&lt;/p&gt;

&lt;h2&gt;
  
  
  What’s Actually Happening
&lt;/h2&gt;

&lt;p&gt;You’re increasing what I call LLM cognitive load.&lt;/p&gt;

&lt;p&gt;The more responsibilities you push into a single call:&lt;/p&gt;

&lt;p&gt;the harder it is for the model to prioritize&lt;br&gt;
the easier it is to miss instructions&lt;br&gt;
the more likely it is to hallucinate&lt;/p&gt;

&lt;p&gt;Even with better models, this pattern doesn’t go away.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;A Better Approach: Think Like a System Designer&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Instead of one big prompt, break the problem into smaller, focused steps.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;Validate + transform + summarize + generate + explain everything&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Do this:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Validation step (code)&lt;/li&gt;
&lt;li&gt;Extraction step (LLM)&lt;/li&gt;
&lt;li&gt;Transformation step (code or LLM)&lt;/li&gt;
&lt;li&gt;Generation step (LLM)&lt;/li&gt;
&lt;li&gt;Formatting step (code)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Use the Right Tool for the Right Job
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Let code handle:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;validation&lt;/li&gt;
&lt;li&gt;parsing&lt;/li&gt;
&lt;li&gt;routing&lt;/li&gt;
&lt;li&gt;rules&lt;/li&gt;
&lt;li&gt;state&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Let LLM handle:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;reasoning&lt;/li&gt;
&lt;li&gt;interpretation&lt;/li&gt;
&lt;li&gt;summarization&lt;/li&gt;
&lt;li&gt;ambiguity&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Treat LLM Calls Like Microservices&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This mindset shift helped me a lot:&lt;/p&gt;

&lt;p&gt;Each LLM call should have a single responsibility&lt;/p&gt;

&lt;p&gt;Small input&lt;br&gt;
Clear task&lt;br&gt;
Predictable output&lt;/p&gt;

&lt;p&gt;Then orchestrate them together.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-World Example&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;While working on API automation systems, we initially tried:&lt;/p&gt;

&lt;p&gt;one prompt to validate specs + generate APIs + create mock data&lt;/p&gt;

&lt;p&gt;It became unstable very quickly.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Splitting it into:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;validation module&lt;/li&gt;
&lt;li&gt;generation module&lt;/li&gt;
&lt;li&gt;mock data module&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;made the system far more reliable.&lt;/p&gt;

&lt;p&gt;LLMs are powerful—but they’re not a replacement for system design.&lt;/p&gt;

&lt;p&gt;“Just add AI” is not an architecture pattern.&lt;/p&gt;

&lt;p&gt;Design your system first.&lt;br&gt;
Then use AI where it actually adds value.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>promptengineering</category>
      <category>systemdesign</category>
    </item>
  </channel>
</rss>
