<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Dixit Angiras</title>
    <description>The latest articles on Forem by Dixit Angiras (@dixit_angiras_1f2a7cb300d).</description>
    <link>https://forem.com/dixit_angiras_1f2a7cb300d</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3900046%2F25d03696-e248-4406-8aab-1d9edfbb141e.jpg</url>
      <title>Forem: Dixit Angiras</title>
      <link>https://forem.com/dixit_angiras_1f2a7cb300d</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/dixit_angiras_1f2a7cb300d"/>
    <language>en</language>
    <item>
      <title>Machine Learning Developers: Why Most ML Projects Fail After the Model Stage</title>
      <dc:creator>Dixit Angiras</dc:creator>
      <pubDate>Thu, 30 Apr 2026 12:02:46 +0000</pubDate>
      <link>https://forem.com/dixit_angiras_1f2a7cb300d/machine-learning-developers-why-most-ml-projects-fail-after-the-model-stage-3320</link>
      <guid>https://forem.com/dixit_angiras_1f2a7cb300d/machine-learning-developers-why-most-ml-projects-fail-after-the-model-stage-3320</guid>
      <description>&lt;p&gt;Training a model is easy.&lt;br&gt;
Getting 85–90% accuracy in a notebook? Also doable.&lt;br&gt;
But getting that model to run reliably in production and drive real outcomes?&lt;br&gt;
That’s where most teams fail.&lt;/p&gt;

&lt;p&gt;The Real Gap: Model vs System&lt;br&gt;
A trained model ≠ a working ML system.&lt;br&gt;
And this is exactly where machine learning developers come in.&lt;br&gt;
They don’t just build models.&lt;br&gt;
They build systems that:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Ingest data continuously&lt;/li&gt;
&lt;li&gt;Serve predictions in real time&lt;/li&gt;
&lt;li&gt;Integrate with applications&lt;/li&gt;
&lt;li&gt;Improve over time&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;What ML Developers Actually Work On&lt;br&gt;
If you’re building anything serious, expect these layers.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Data Pipeline (Everything starts here)
Before modeling:&lt;/li&gt;
&lt;li&gt;Data ingestion (batch/stream)&lt;/li&gt;
&lt;li&gt;Cleaning &amp;amp; normalization&lt;/li&gt;
&lt;li&gt;Feature engineering&lt;/li&gt;
&lt;li&gt;Storage (data lake / warehouse)
Tools:&lt;/li&gt;
&lt;li&gt;Pandas, Spark&lt;/li&gt;
&lt;li&gt;Airflow / Prefect&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Kafka (for streaming)&lt;br&gt;
Bad pipeline → unstable system.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Model Training (Only ~20% of the work)&lt;br&gt;
This is the visible part:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Algorithm selection (XGBoost, Neural Nets, etc.)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Training &amp;amp; validation&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Hyperparameter tuning&lt;br&gt;
Frameworks:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Scikit-learn&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;TensorFlow / PyTorch&lt;br&gt;
Important: accuracy alone is not the goal.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Model Deployment (Where things break)&lt;br&gt;
Moving from notebook → production:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;REST APIs (FastAPI / Flask)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Model serialization (Pickle, ONNX)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Containerization (Docker)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cloud deployment (AWS/GCP/Azure)&lt;br&gt;
If this layer is weak → your model never gets used.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Inference Layer (Real-time or batch)&lt;br&gt;
Decide:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Real-time predictions (low latency)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Batch predictions (scheduled jobs)&lt;br&gt;
Trade-offs:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cost vs speed&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Complexity vs scalability&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;MLOps &amp;amp; Monitoring (Non-negotiable)&lt;br&gt;
Models degrade.&lt;br&gt;
You need:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Performance tracking&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data drift detection&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Logging&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Retraining pipelines&lt;br&gt;
Tools:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;MLflow&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Prometheus / Grafana&lt;br&gt;
No monitoring → silent failure.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Integration with Business Logic&lt;br&gt;
This is where value is created.&lt;br&gt;
Predictions must trigger actions:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Send recommendation&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Flag fraud&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Adjust pricing&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Trigger workflows&lt;br&gt;
Without this, ML is just analytics.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;A Practical ML System Flow&lt;/p&gt;

&lt;p&gt;Raw Data&lt;br&gt;
   ↓&lt;br&gt;
Data Pipeline (ETL)&lt;br&gt;
   ↓&lt;br&gt;
Feature Store&lt;br&gt;
   ↓&lt;br&gt;
Model Training&lt;br&gt;
   ↓&lt;br&gt;
Model Registry&lt;br&gt;
   ↓&lt;br&gt;
Deployment (API)&lt;br&gt;
   ↓&lt;br&gt;
Inference Layer&lt;br&gt;
   ↓&lt;br&gt;
Application / Workflow&lt;br&gt;
   ↓&lt;br&gt;
Monitoring &amp;amp; Retraining&lt;/p&gt;

&lt;p&gt;Where Most Teams Go Wrong&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Focusing only on model accuracy&lt;/li&gt;
&lt;li&gt;Ignoring deployment until the end&lt;/li&gt;
&lt;li&gt;No data versioning&lt;/li&gt;
&lt;li&gt;No monitoring strategy&lt;/li&gt;
&lt;li&gt;Treating ML as a one-time project
That’s why many ML initiatives never leave the prototype stage.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Real Use Cases Built This Way&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Recommendation systems (e-commerce, streaming)&lt;/li&gt;
&lt;li&gt;Fraud detection (finance)&lt;/li&gt;
&lt;li&gt;Demand forecasting (supply chain)&lt;/li&gt;
&lt;li&gt;Predictive maintenance (manufacturing)
These systems aren’t just models.
They’re continuous pipelines.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;When Do You Actually Need ML Developers?&lt;br&gt;
Not every project needs ML.&lt;br&gt;
But you do if:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Rules aren’t enough anymore&lt;/li&gt;
&lt;li&gt;Data is growing fast&lt;/li&gt;
&lt;li&gt;You need predictions, not reports&lt;/li&gt;
&lt;li&gt;You want automation at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Where Services Fit In&lt;br&gt;
If you're building production-grade systems or scaling across teams, structured support can help with:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Architecture design&lt;/li&gt;
&lt;li&gt;Deployment pipelines&lt;/li&gt;
&lt;li&gt;MLOps setup&lt;/li&gt;
&lt;li&gt;Optimization
If you want to see how such systems are implemented in real scenarios: &lt;a href="https://artificialintelligence.oodles.io/services/machine-learning-development-services/machine-learning-developers/" rel="noopener noreferrer"&gt;https://artificialintelligence.oodles.io/services/machine-learning-development-services/machine-learning-developers/&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Final Thought&lt;br&gt;
Machine learning is easy to prototype.&lt;br&gt;
Hard to productionize.&lt;br&gt;
The difference isn’t the model.&lt;br&gt;
It’s everything around it.&lt;br&gt;
If you’re building ML, optimize for: → reliability → integration → continuous improvement&lt;br&gt;
That’s what turns a model into a system.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>architecture</category>
      <category>dataengineering</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>Machine Learning Developers: What It Actually Takes to Build ML Systems That Work</title>
      <dc:creator>Dixit Angiras</dc:creator>
      <pubDate>Wed, 29 Apr 2026 10:24:07 +0000</pubDate>
      <link>https://forem.com/dixit_angiras_1f2a7cb300d/machine-learning-developers-what-it-actually-takes-to-build-ml-systems-that-work-53l</link>
      <guid>https://forem.com/dixit_angiras_1f2a7cb300d/machine-learning-developers-what-it-actually-takes-to-build-ml-systems-that-work-53l</guid>
      <description>&lt;p&gt;A lot of teams say they’re “doing machine learning.”&lt;br&gt;
What they often mean is:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Training a model in a notebook&lt;/li&gt;
&lt;li&gt;Getting decent accuracy&lt;/li&gt;
&lt;li&gt;Calling it done
That’s not machine learning in production.
That’s experimentation.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The Gap Between Models and Systems&lt;br&gt;
Building a model is one step.&lt;br&gt;
Building a machine learning system is something else entirely.&lt;br&gt;
And this is where machine learning developers come in.&lt;br&gt;
They don’t just train models.&lt;br&gt;
They make them usable, reliable, and scalable.&lt;/p&gt;

&lt;p&gt;What Machine Learning Developers Actually Do&lt;br&gt;
If you strip away the buzzwords, their job is to build end-to-end pipelines:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Data Engineering (The Real Heavy Lifting)
Before any model:&lt;/li&gt;
&lt;li&gt;Data collection&lt;/li&gt;
&lt;li&gt;Cleaning&lt;/li&gt;
&lt;li&gt;Feature engineering&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Pipeline creation&lt;br&gt;
Bad data = useless model.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Model Development&lt;br&gt;
This is the visible part:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Choosing algorithms&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Training models&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Hyperparameter tuning&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Evaluation&lt;br&gt;
But this is only a fraction of the work.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Deployment (Where Most Projects Fail)&lt;br&gt;
A model in a notebook has zero business value.&lt;br&gt;
Deployment involves:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;APIs (FastAPI, Flask)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Batch or real-time inference&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Containerization (Docker)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cloud setup (AWS/GCP/Azure)&lt;br&gt;
This is where many teams get stuck.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;MLOps &amp;amp; Monitoring&lt;br&gt;
Models degrade over time.&lt;br&gt;
You need:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Logging&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Performance tracking&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data drift detection&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Retraining pipelines&lt;br&gt;
Without this, accuracy drops silently.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Integration with Business Systems&lt;br&gt;
Predictions need to trigger actions.&lt;br&gt;
That means connecting ML outputs to:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;CRMs&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;ERPs&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Internal tools&lt;br&gt;
Otherwise, it’s just another dashboard.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;A Simple ML System Architecture&lt;/p&gt;

&lt;p&gt;Data Sources&lt;br&gt;
   ↓&lt;br&gt;
Data Pipeline (ETL)&lt;br&gt;
   ↓&lt;br&gt;
Feature Engineering&lt;br&gt;
   ↓&lt;br&gt;
Model Training&lt;br&gt;
   ↓&lt;br&gt;
Model Deployment (API)&lt;br&gt;
   ↓&lt;br&gt;
Inference Layer&lt;br&gt;
   ↓&lt;br&gt;
Business Application&lt;br&gt;
   ↓&lt;br&gt;
Monitoring &amp;amp; Retraining&lt;/p&gt;

&lt;p&gt;Where Most Teams Go Wrong&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Focusing only on model accuracy&lt;/li&gt;
&lt;li&gt;Ignoring data pipelines&lt;/li&gt;
&lt;li&gt;Skipping deployment planning&lt;/li&gt;
&lt;li&gt;No monitoring or retraining&lt;/li&gt;
&lt;li&gt;Treating ML as a one-time project
Machine learning is not static.
It’s a continuous system.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Real-World Use Cases&lt;br&gt;
Machine learning developers are building systems like:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Recommendation engines (Netflix/Amazon style)&lt;/li&gt;
&lt;li&gt;Fraud detection systems&lt;/li&gt;
&lt;li&gt;Demand forecasting models&lt;/li&gt;
&lt;li&gt;Predictive maintenance systems
These aren’t “models.”
They’re production systems that evolve over time.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;When Do You Actually Need ML Developers?&lt;br&gt;
Not every project needs ML.&lt;br&gt;
But you do if:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;You have large, growing datasets&lt;/li&gt;
&lt;li&gt;You need predictions or automation&lt;/li&gt;
&lt;li&gt;Rule-based systems aren’t enough&lt;/li&gt;
&lt;li&gt;You want systems that improve with data&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Where Services Fit In&lt;br&gt;
If you’re building something complex or scaling across teams, structured support can help.&lt;br&gt;
Teams offering machine learning development services typically handle:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Architecture design&lt;/li&gt;
&lt;li&gt;Model development&lt;/li&gt;
&lt;li&gt;Deployment&lt;/li&gt;
&lt;li&gt;MLOps
If you want to see how these systems are implemented in real scenarios, this is a useful reference: &lt;a href="https://artificialintelligence.oodles.io/services/machine-learning-development-services/machine-learning-developers/" rel="noopener noreferrer"&gt;https://artificialintelligence.oodles.io/services/machine-learning-development-services/machine-learning-developers/&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Final Thoughts&lt;br&gt;
Machine learning is easy to prototype.&lt;br&gt;
Hard to productionize.&lt;br&gt;
The difference isn’t the algorithm.&lt;br&gt;
It’s the system around it.&lt;br&gt;
If you're building ML, don’t just aim for accuracy.&lt;br&gt;
Aim for something that actually runs, scales, and improves over time.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Generative AI Development Services: What It Actually Takes to Move from Demo to Production</title>
      <dc:creator>Dixit Angiras</dc:creator>
      <pubDate>Tue, 28 Apr 2026 12:33:03 +0000</pubDate>
      <link>https://forem.com/dixit_angiras_1f2a7cb300d/generative-ai-development-services-what-it-actually-takes-to-move-from-demo-to-production-2a6h</link>
      <guid>https://forem.com/dixit_angiras_1f2a7cb300d/generative-ai-development-services-what-it-actually-takes-to-move-from-demo-to-production-2a6h</guid>
      <description>&lt;p&gt;Most developers have already experimented with generative AI.&lt;br&gt;
You call an API, send a prompt, and get a response. It works surprisingly well.&lt;br&gt;
Until you try to use it in a real product.&lt;br&gt;
That’s where things start to break.&lt;/p&gt;

&lt;p&gt;The Problem with “API-First AI”&lt;br&gt;
The default approach looks like this:&lt;/p&gt;

&lt;p&gt;Use OpenAI / other LLM APIs&lt;/p&gt;

&lt;p&gt;Add prompt templates&lt;/p&gt;

&lt;p&gt;Ship a feature&lt;/p&gt;

&lt;p&gt;For simple use cases, that’s fine.&lt;br&gt;
But in production, you quickly run into issues:&lt;/p&gt;

&lt;p&gt;Responses lack domain context&lt;/p&gt;

&lt;p&gt;Hallucinations become risky&lt;/p&gt;

&lt;p&gt;No access to internal knowledge&lt;/p&gt;

&lt;p&gt;Latency and cost increase with scale&lt;/p&gt;

&lt;p&gt;Limited control over outputs&lt;/p&gt;

&lt;p&gt;At that point, you realize:&lt;br&gt;
You’re not building an AI system.&lt;br&gt;
You’re wrapping an API.&lt;/p&gt;

&lt;p&gt;What Generative AI Development Actually Involves&lt;br&gt;
If you're building something that needs to scale, you need more than prompts.&lt;br&gt;
You need a system architecture.&lt;br&gt;
That’s where generative AI development services come in—not as a buzzword, but as a structured way to build production-ready AI.&lt;/p&gt;

&lt;p&gt;Core Components of a Production-Ready AI System&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Data Layer (The Real Differentiator)
Your advantage isn’t the model.
It’s your data.
This includes:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Internal documents&lt;/p&gt;

&lt;p&gt;Customer interactions&lt;/p&gt;

&lt;p&gt;Structured + unstructured datasets&lt;/p&gt;

&lt;p&gt;Without this layer, your AI stays generic.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Retrieval-Augmented Generation (RAG)
Instead of relying purely on model memory, use retrieval.
Basic flow:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;User query&lt;/p&gt;

&lt;p&gt;Retrieve relevant documents (vector DB)&lt;/p&gt;

&lt;p&gt;Inject context into prompt&lt;/p&gt;

&lt;p&gt;Generate response&lt;/p&gt;

&lt;p&gt;Tools:&lt;/p&gt;

&lt;p&gt;FAISS / Pinecone / Weaviate&lt;/p&gt;

&lt;p&gt;LangChain / LlamaIndex&lt;/p&gt;

&lt;p&gt;This reduces hallucinations and improves accuracy.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Model Strategy
You don’t always need to train from scratch.
Options:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;API-based models (fast to start)&lt;/p&gt;

&lt;p&gt;Open-source models (more control)&lt;/p&gt;

&lt;p&gt;Fine-tuned models (better relevance)&lt;/p&gt;

&lt;p&gt;Trade-offs:&lt;/p&gt;

&lt;p&gt;Cost vs control&lt;/p&gt;

&lt;p&gt;Speed vs customization&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Prompt Engineering + Guardrails
Prompts alone aren’t enough.
You need:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Structured prompts&lt;/p&gt;

&lt;p&gt;Output formatting&lt;/p&gt;

&lt;p&gt;Validation layers&lt;/p&gt;

&lt;p&gt;Safety filters&lt;/p&gt;

&lt;p&gt;Think of prompts as logic, not just text.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Workflow Integration
AI doesn’t create value in isolation.
It needs to connect with:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Backend services&lt;/p&gt;

&lt;p&gt;CRMs / ERPs&lt;/p&gt;

&lt;p&gt;Internal tools&lt;/p&gt;

&lt;p&gt;This is where most “AI features” fail—they stop at output, not action.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Monitoring &amp;amp; Feedback Loops
Production AI requires:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Logging outputs&lt;/p&gt;

&lt;p&gt;Tracking errors&lt;/p&gt;

&lt;p&gt;Human-in-the-loop corrections&lt;/p&gt;

&lt;p&gt;Continuous improvement&lt;/p&gt;

&lt;p&gt;Without this, quality degrades over time.&lt;/p&gt;

&lt;p&gt;A Simplified Architecture&lt;br&gt;
User Input   ↓API Layer   ↓Retriever (Vector DB)   ↓LLM (API / Fine-tuned Model)   ↓Post-processing &amp;amp; Validation   ↓Business Logic / Workflow   ↓Response / Action&lt;/p&gt;

&lt;p&gt;Real-World Use Cases&lt;br&gt;
This approach is already being used to build:&lt;/p&gt;

&lt;p&gt;AI copilots for internal teams&lt;/p&gt;

&lt;p&gt;Knowledge-based chat systems&lt;/p&gt;

&lt;p&gt;Content generation pipelines&lt;/p&gt;

&lt;p&gt;Automated support workflows&lt;/p&gt;

&lt;p&gt;These systems go beyond “text generation” and actually drive operations.&lt;/p&gt;

&lt;p&gt;Where Most Teams Go Wrong&lt;/p&gt;

&lt;p&gt;Over-relying on prompts&lt;/p&gt;

&lt;p&gt;Ignoring data quality&lt;/p&gt;

&lt;p&gt;Skipping retrieval systems&lt;/p&gt;

&lt;p&gt;Not designing for scale&lt;/p&gt;

&lt;p&gt;Treating AI as a feature, not infrastructure&lt;/p&gt;

&lt;p&gt;Where Development Services Fit In&lt;br&gt;
If you’re building something simple, you don’t need external help.&lt;br&gt;
But if you're:&lt;/p&gt;

&lt;p&gt;Handling sensitive data&lt;/p&gt;

&lt;p&gt;Scaling across teams&lt;/p&gt;

&lt;p&gt;Building complex workflows&lt;/p&gt;

&lt;p&gt;Then structured generative AI development services can help design, build, and optimize these systems properly.&lt;br&gt;
If you want to see how such systems are implemented in real business scenarios, this is a useful reference:&lt;br&gt;
&lt;a href="https://artificialintelligence.oodles.io/services/generative-ai/generative-ai-development-services/" rel="noopener noreferrer"&gt;https://artificialintelligence.oodles.io/services/generative-ai/generative-ai-development-services/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Final Thoughts&lt;br&gt;
Generative AI is easy to demo.&lt;br&gt;
Hard to productionize.&lt;br&gt;
The difference comes down to one thing:&lt;br&gt;
Are you just generating outputs?&lt;br&gt;
Or building systems that use them?&lt;br&gt;
If it's the second, you need to think beyond APIs—and start thinking in architecture.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>softwareengineering</category>
      <category>systemdesign</category>
    </item>
    <item>
      <title>Building OCR Solutions That Actually Work in Production (Not Just Demos)</title>
      <dc:creator>Dixit Angiras</dc:creator>
      <pubDate>Mon, 27 Apr 2026 08:54:13 +0000</pubDate>
      <link>https://forem.com/dixit_angiras_1f2a7cb300d/building-ocr-solutions-that-actually-work-in-production-not-just-demos-1gfp</link>
      <guid>https://forem.com/dixit_angiras_1f2a7cb300d/building-ocr-solutions-that-actually-work-in-production-not-just-demos-1gfp</guid>
      <description>&lt;p&gt;Most developers have tried OCR at some point.&lt;/p&gt;

&lt;p&gt;You pick a library, run it on a PDF, extract text… and it works.&lt;/p&gt;

&lt;p&gt;Until you try to use it in a real system.&lt;/p&gt;

&lt;p&gt;That’s where things start breaking.&lt;/p&gt;

&lt;p&gt;The Problem with “Basic OCR”&lt;/p&gt;

&lt;p&gt;Out-of-the-box OCR (like Tesseract or simple APIs) works fine for:&lt;/p&gt;

&lt;p&gt;Clean documents&lt;br&gt;
Standard fonts&lt;br&gt;
Structured layouts&lt;/p&gt;

&lt;p&gt;But real-world documents are messy:&lt;/p&gt;

&lt;p&gt;Different invoice formats&lt;br&gt;
Skewed scans&lt;br&gt;
Low-quality images&lt;br&gt;
Handwritten fields&lt;br&gt;
Multi-language content&lt;/p&gt;

&lt;p&gt;So what happens?&lt;/p&gt;

&lt;p&gt;You get:&lt;/p&gt;

&lt;p&gt;Incorrect extraction&lt;br&gt;
Missing fields&lt;br&gt;
Broken pipelines&lt;br&gt;
Manual fallback (again)&lt;/p&gt;

&lt;p&gt;At that point, OCR becomes a partial solution, not automation.&lt;/p&gt;

&lt;p&gt;What Production-Ready OCR Actually Requires&lt;/p&gt;

&lt;p&gt;If you're building OCR for real use cases (invoices, KYC, forms), think beyond text extraction.&lt;/p&gt;

&lt;p&gt;You need a pipeline, not a tool.&lt;/p&gt;

&lt;p&gt;Step 1: Image Preprocessing (Critical but Ignored)&lt;/p&gt;

&lt;p&gt;Before OCR, clean the input.&lt;/p&gt;

&lt;p&gt;Typical steps:&lt;/p&gt;

&lt;p&gt;Deskewing&lt;br&gt;
Noise removal&lt;br&gt;
Binarization&lt;br&gt;
Contrast enhancement&lt;/p&gt;

&lt;p&gt;Libraries:&lt;/p&gt;

&lt;p&gt;OpenCV&lt;br&gt;
Pillow&lt;/p&gt;

&lt;p&gt;Without this, accuracy drops significantly.&lt;/p&gt;

&lt;p&gt;Step 2: OCR Engine Selection&lt;/p&gt;

&lt;p&gt;Options depend on your use case:&lt;/p&gt;

&lt;p&gt;Tesseract → Open-source, customizable&lt;br&gt;
EasyOCR / PaddleOCR → Better for deep learning-based extraction&lt;br&gt;
Cloud APIs (AWS Textract, Google Vision) → Higher accuracy, less control&lt;/p&gt;

&lt;p&gt;There’s no “best” option—only trade-offs.&lt;/p&gt;

&lt;p&gt;Step 3: Layout &amp;amp; Document Understanding&lt;/p&gt;

&lt;p&gt;Raw text is useless without structure.&lt;/p&gt;

&lt;p&gt;You need to identify:&lt;/p&gt;

&lt;p&gt;Headers&lt;br&gt;
Tables&lt;br&gt;
Key-value pairs&lt;/p&gt;

&lt;p&gt;Tools:&lt;/p&gt;

&lt;p&gt;LayoutLM&lt;br&gt;
Detectron2&lt;br&gt;
Donut (for document understanding)&lt;/p&gt;

&lt;p&gt;This is where most OCR systems fail.&lt;/p&gt;

&lt;p&gt;Step 4: Field Extraction (The Real Value Layer)&lt;/p&gt;

&lt;p&gt;Instead of returning full text, extract:&lt;/p&gt;

&lt;p&gt;Invoice number&lt;br&gt;
Date&lt;br&gt;
Amount&lt;br&gt;
Name&lt;/p&gt;

&lt;p&gt;Approaches:&lt;/p&gt;

&lt;p&gt;Rule-based (regex)&lt;br&gt;
ML models&lt;br&gt;
LLM-assisted extraction&lt;/p&gt;

&lt;p&gt;LLMs are increasingly useful here for flexible parsing.&lt;/p&gt;

&lt;p&gt;Step 5: Post-Processing &amp;amp; Validation&lt;/p&gt;

&lt;p&gt;Even good OCR isn’t perfect.&lt;/p&gt;

&lt;p&gt;Add:&lt;/p&gt;

&lt;p&gt;Confidence thresholds&lt;br&gt;
Validation rules&lt;br&gt;
Human-in-the-loop fallback&lt;/p&gt;

&lt;p&gt;This ensures reliability.&lt;/p&gt;

&lt;p&gt;Step 6: Integration into Workflows&lt;/p&gt;

&lt;p&gt;OCR alone doesn’t create value.&lt;/p&gt;

&lt;p&gt;It needs to connect with:&lt;/p&gt;

&lt;p&gt;ERP systems&lt;br&gt;
CRMs&lt;br&gt;
Databases&lt;br&gt;
Internal tools&lt;/p&gt;

&lt;p&gt;This is where automation actually happens.&lt;/p&gt;

&lt;p&gt;Real-World Architecture (Simplified)&lt;br&gt;
Input (PDF/Image)&lt;br&gt;
   ↓&lt;br&gt;
Preprocessing (OpenCV)&lt;br&gt;
   ↓&lt;br&gt;
OCR Engine (Tesseract / API)&lt;br&gt;
   ↓&lt;br&gt;
Layout Detection (LayoutLM)&lt;br&gt;
   ↓&lt;br&gt;
Field Extraction (ML / LLM)&lt;br&gt;
   ↓&lt;br&gt;
Validation Layer&lt;br&gt;
   ↓&lt;br&gt;
API / Database / CRM&lt;br&gt;
Where Most Teams Go Wrong&lt;br&gt;
Treating OCR as a one-step process&lt;br&gt;
Ignoring preprocessing&lt;br&gt;
Expecting 100% accuracy&lt;br&gt;
Not designing fallback systems&lt;br&gt;
Skipping integration&lt;/p&gt;

&lt;p&gt;OCR isn’t hard because of text extraction.&lt;/p&gt;

&lt;p&gt;It’s hard because of variability.&lt;/p&gt;

&lt;p&gt;Where Modern OCR Is Heading&lt;/p&gt;

&lt;p&gt;The shift is clear:&lt;/p&gt;

&lt;p&gt;From:&lt;br&gt;
Text extraction&lt;/p&gt;

&lt;p&gt;To:&lt;br&gt;
Document understanding&lt;/p&gt;

&lt;p&gt;With:&lt;/p&gt;

&lt;p&gt;AI models&lt;br&gt;
Context-aware parsing&lt;br&gt;
Continuous learning&lt;/p&gt;

&lt;p&gt;This is what enables near full automation.&lt;/p&gt;

&lt;p&gt;Real Implementation Insight&lt;/p&gt;

&lt;p&gt;In production systems, OCR is often combined with:&lt;/p&gt;

&lt;p&gt;AI models for classification&lt;br&gt;
LLMs for flexible data extraction&lt;br&gt;
RAG systems for validation&lt;/p&gt;

&lt;p&gt;This creates end-to-end automation instead of partial solutions.&lt;/p&gt;

&lt;p&gt;If you want to explore how such systems are built in real business scenarios, this is a useful reference:&lt;br&gt;
&lt;a href="https://artificialintelligence.oodles.io/optical-character-recognition-services" rel="noopener noreferrer"&gt;https://artificialintelligence.oodles.io/optical-character-recognition-services&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Final Thoughts&lt;/p&gt;

&lt;p&gt;OCR is easy to demo.&lt;/p&gt;

&lt;p&gt;Hard to scale.&lt;/p&gt;

&lt;p&gt;If you're building one:&lt;br&gt;
Don’t optimize for extraction.&lt;/p&gt;

&lt;p&gt;Optimize for accuracy + structure + integration.&lt;/p&gt;

&lt;p&gt;That’s what turns OCR into a real system—not just a feature.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>architecture</category>
      <category>machinelearning</category>
      <category>softwareengineering</category>
    </item>
  </channel>
</rss>
