<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Tahmid Khan A</title>
    <description>The latest articles on Forem by Tahmid Khan A (@tahmid_khana_alim).</description>
    <link>https://forem.com/tahmid_khana_alim</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3760938%2Fbf7554e8-f372-43c9-9c5d-4d580a0b018f.jpg</url>
      <title>Forem: Tahmid Khan A</title>
      <link>https://forem.com/tahmid_khana_alim</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/tahmid_khana_alim"/>
    <language>en</language>
    <item>
      <title>The AI Code Deluge: Why We're Drowning in Technical Debt</title>
      <dc:creator>Tahmid Khan A</dc:creator>
      <pubDate>Wed, 11 Feb 2026 13:36:40 +0000</pubDate>
      <link>https://forem.com/tahmid_khana_alim/the-ai-code-deluge-why-were-drowning-in-technical-debt-18he</link>
      <guid>https://forem.com/tahmid_khana_alim/the-ai-code-deluge-why-were-drowning-in-technical-debt-18he</guid>
      <description>&lt;p&gt;It’s February 2026. The hype cycle has shifted again. We’ve moved past the "Chatbot" phase and firmly into the "Agentic" era. OpenAI, Anthropic, and Microsoft are selling us a dream: autonomous coding agents that live in your IDE, understand your entire repo, and ship features while you sleep.&lt;/p&gt;

&lt;p&gt;If you believe the marketing, software engineering as a discipline is about to be "solved."&lt;/p&gt;

&lt;p&gt;But if you look at the actual state of codebases in 2026, a different reality is emerging. We aren't just shipping features faster; we are generating technical debt at a velocity that human teams cannot sustain.&lt;/p&gt;

&lt;p&gt;We are witnessing the rise of &lt;strong&gt;AI Productivity Theater&lt;/strong&gt;. And if we don't change how we use these tools, we're going to drown in a sea of mediocre, unmaintainable code.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Illusion of Velocity
&lt;/h2&gt;

&lt;p&gt;I spent the last week auditing a project that had "heavily adopted" AI workflows. The team was proud. Their commit volume was up 300%. Tickets were moving from "In Progress" to "Done" in record time.&lt;/p&gt;

&lt;p&gt;On the surface, it looked like a hyper-efficient machine.&lt;/p&gt;

&lt;p&gt;Then I opened the code.&lt;/p&gt;

&lt;p&gt;It was a sprawling mess of disconnected logic. Functions were duplicated because the AI didn't know a utility for that already existed in &lt;code&gt;utils/&lt;/code&gt;. Components had slightly different styling implementations because three different prompts generated them. Error handling was generic—lots of &lt;code&gt;try/catch&lt;/code&gt; blocks wrapping massive chunks of logic, logging &lt;code&gt;e&lt;/code&gt; to the console, and failing silently.&lt;/p&gt;

&lt;p&gt;The code &lt;em&gt;worked&lt;/em&gt;. The tests (also written by AI) passed.&lt;/p&gt;

&lt;p&gt;But the &lt;strong&gt;architecture&lt;/strong&gt; was rotting.&lt;/p&gt;

&lt;p&gt;This is the trap. AI is fantastic at the &lt;strong&gt;micro&lt;/strong&gt; (writing a function, fixing a regex, generating a test case). It is terrible at the &lt;strong&gt;macro&lt;/strong&gt; (system cohesion, data flow consistency, long-term maintainability).&lt;/p&gt;

&lt;p&gt;When you let an autocomplete engine drive your architecture, you get exactly what you'd expect: a system that looks like a patchwork quilt of Stack Overflow answers, stitched together with confidence but no comprehension.&lt;/p&gt;

&lt;h2&gt;
  
  
  The "Senior Review" Bottleneck
&lt;/h2&gt;

&lt;p&gt;Here is the dirty secret of 2026: &lt;strong&gt;Code Review is dead.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Or rather, &lt;em&gt;effective&lt;/em&gt; code review is dying.&lt;/p&gt;

&lt;p&gt;In the old days (circa 2023), a Junior Dev would submit a PR with 50 lines of code. A Senior would read it, spot a logic error, and explain &lt;em&gt;why&lt;/em&gt; it was wrong. That was the feedback loop. That was mentorship.&lt;/p&gt;

&lt;p&gt;Today, an "AI-Augmented" Junior submits a PR with 500 lines of code. It looks clean. It follows the linter rules. The variable names are descriptive.&lt;/p&gt;

&lt;p&gt;But to verify if it's &lt;em&gt;actually&lt;/em&gt; correct—to check for race conditions, edge cases in state management, or security holes—the Senior has to mentally reconstruct the entire logic flow.&lt;/p&gt;

&lt;p&gt;And they don't.&lt;/p&gt;

&lt;p&gt;It takes too much energy. When the code &lt;em&gt;looks&lt;/em&gt; right, the brain skims. We approve the PR. We merge the debt.&lt;/p&gt;

&lt;p&gt;We are replacing "Junior Developers" with "AI Agents," but we are forgetting that Juniors grow up to be Seniors. Agents don't. Agents don't learn from your architecture; they just statistically predict the next token. They will make the same architectural mistake a thousand times if you let them.&lt;/p&gt;

&lt;h2&gt;
  
  
  Case Study: The "Refactor from Hell"
&lt;/h2&gt;

&lt;p&gt;Let me tell you a story about a startup I consulted for last month (names changed to protect the innocent). They were building a fintech dashboard. They used a popular "Agentic IDE" to scaffold the entire backend in a week.&lt;/p&gt;

&lt;p&gt;"It saved us two months of dev time!" the CTO told me.&lt;/p&gt;

&lt;p&gt;The backend was built using Python and FastAPI. It had 40 endpoints.&lt;/p&gt;

&lt;p&gt;When I looked closer, I found &lt;strong&gt;14 different ways&lt;/strong&gt; of connecting to the database. Some endpoints used an ORM. Some used raw SQL strings (generated by the AI). Some used a deprecated driver that the AI hallucinated was the "standard."&lt;/p&gt;

&lt;p&gt;Why? Because different team members used different prompts at different times, and the AI just gave them whatever was statistically likely for that specific prompt context. It didn't look at the other files to see how the connection was &lt;em&gt;already&lt;/em&gt; established.&lt;/p&gt;

&lt;p&gt;When they tried to migrate the database schema, everything exploded.&lt;/p&gt;

&lt;p&gt;The "two months saved" were immediately spent on a painful, month-long rewrite where humans had to go in and untangle the AI's spaghetti.&lt;/p&gt;

&lt;p&gt;This is the &lt;strong&gt;Hidden Tax&lt;/strong&gt; of AI code. You don't pay it when you write the code. You pay it with high interest when you try to change it.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Security &amp;amp; Compliance Minefield
&lt;/h2&gt;

&lt;p&gt;Beyond architecture, there's a looming security crisis.&lt;/p&gt;

&lt;p&gt;I recently saw a codebase where an AI agent had helpfully imported a package called &lt;code&gt;azure-sdk-python-core&lt;/code&gt;. Sounds official, right?&lt;/p&gt;

&lt;p&gt;It wasn't. It was a typosquatted malware package.&lt;/p&gt;

&lt;p&gt;The AI didn't "know" it was malware. It just saw that &lt;code&gt;azure-sdk&lt;/code&gt; and &lt;code&gt;python-core&lt;/code&gt; often appear together in its training data, and hallucinated a package name that sounded plausible. Because the package actually existed on PyPI (registered by an attacker), the install succeeded.&lt;/p&gt;

&lt;p&gt;Furthermore, we are pasting massive amounts of proprietary business logic into context windows. "Sovereign AI" is the buzzword of 2026 for a reason—enterprises are terrified. But developers are lazy. If the local model is dumb, they will paste the code into the smart cloud model.&lt;/p&gt;

&lt;p&gt;We are leaking our IP, bit by bit, into the training datasets of the future.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Death of the "Tinkerer"
&lt;/h2&gt;

&lt;p&gt;This brings me to the saddest trend I see on Reddit and Hacker News: the despair of the entry-level market.&lt;/p&gt;

&lt;p&gt;Companies are freezing hiring for juniors because "AI can do that work."&lt;/p&gt;

&lt;p&gt;This is short-sighted suicide.&lt;/p&gt;

&lt;p&gt;A Junior Developer isn't just a "ticket closer." They are a future architect in training. They learn by breaking things. They learn by struggling through a weird dependency conflict. They learn by writing a slow query, taking down production, and fixing it.&lt;/p&gt;

&lt;p&gt;If you outsource the "struggle" to an AI, you outsource the learning.&lt;/p&gt;

&lt;p&gt;We are raising a generation of "Prompt Engineers" who can summon a React component in seconds but can't debug a memory leak because they have no mental model of how the DOM actually works. They know the &lt;em&gt;syntax&lt;/em&gt; of the solution, but not the &lt;em&gt;mechanics&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;When the abstraction leaks—and it always does—they are helpless.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Survival Guide for the AI Age
&lt;/h2&gt;

&lt;p&gt;So, am I a luddite? Am I saying we should go back to writing assembly on stone tablets?&lt;/p&gt;

&lt;p&gt;No. I use these tools every single day. My productivity &lt;em&gt;is&lt;/em&gt; higher. But my &lt;strong&gt;process&lt;/strong&gt; has changed. And yours needs to, too.&lt;/p&gt;

&lt;p&gt;Here is the &lt;strong&gt;Senior Engineer's Manifesto for AI&lt;/strong&gt;:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Treat AI as a Junior, Not a Guru
&lt;/h3&gt;

&lt;p&gt;Never ask an AI to "design" a system. Design it yourself. Draw the boxes. Define the interfaces. Then, and only then, ask the AI to implement the specific implementation details inside those boxes.&lt;/p&gt;

&lt;p&gt;You are the Architect. The AI is the bricklayer. If you let the bricklayer decide where the walls go, you're going to end up with a house that has no doors.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. The "Explain It To Me" Rule
&lt;/h3&gt;

&lt;p&gt;When an AI generates code for you, do not commit it until you can explain &lt;em&gt;exactly&lt;/em&gt; what every line does. If there is a regex you don't understand, ask the AI to break it down. If there is a library import you don't recognize, Google it (don't ask the AI, verify it externally).&lt;/p&gt;

&lt;p&gt;If you commit code you don't understand, you are not a developer. You are a copy-paster.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Mentorship &amp;gt; Automation
&lt;/h3&gt;

&lt;p&gt;If you manage a team, don't let your juniors use AI to bypass the struggle. Encourage them to use it to &lt;em&gt;explain&lt;/em&gt; concepts ("Explain why this useEffect is triggering twice"), not to &lt;em&gt;solve&lt;/em&gt; them ("Fix this useEffect").&lt;/p&gt;

&lt;p&gt;The goal is to build &lt;strong&gt;mental models&lt;/strong&gt;, not just software.&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Code Deletion is the New Productivity Metric
&lt;/h3&gt;

&lt;p&gt;In an age where generating code is free, the value of code approaches zero. The liability of code approaches infinity.&lt;/p&gt;

&lt;p&gt;Celebrate the PRs that &lt;em&gt;delete&lt;/em&gt; code. Celebrate the refactors that &lt;em&gt;simplify&lt;/em&gt; logic. Use AI to find dead code, to consolidate duplicates, to generate documentation—tasks that &lt;em&gt;reduce&lt;/em&gt; entropy, not increase it.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Verdict
&lt;/h2&gt;

&lt;p&gt;The flood of AI-generated code isn't going to stop. Tools like the Model Context Protocol (MCP) are only going to make it easier to generate massive PRs with a single click.&lt;/p&gt;

&lt;p&gt;But software engineering isn't about &lt;strong&gt;writing code&lt;/strong&gt;. It never was.&lt;/p&gt;

&lt;p&gt;It's about &lt;strong&gt;managing complexity&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Right now, AI is the greatest complexity &lt;em&gt;generator&lt;/em&gt; we have ever invented. Our job, more than ever, is to be the filter. To be the ruthless editor. To say "No" to the easy, generated path and "Yes" to the hard, thoughtful architecture.&lt;/p&gt;

&lt;p&gt;Don't let the tools fool you. Speed is not quality. And in 2026, the most valuable developer isn't the one who writes the most code—it's the one who knows what code &lt;em&gt;not&lt;/em&gt; to write.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://reboundbytes.top/blog/2026-02-11-ai-technical-debt-crisis" rel="noopener noreferrer"&gt;Rebound Bytes&lt;/a&gt;. No fluff, just code.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>codequality</category>
      <category>discuss</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>AI Agents Are Productivity Theater (And That's Fine)</title>
      <dc:creator>Tahmid Khan A</dc:creator>
      <pubDate>Tue, 10 Feb 2026 19:14:15 +0000</pubDate>
      <link>https://forem.com/tahmid_khana_alim/ai-agents-are-productivity-theater-and-thats-fine-geh</link>
      <guid>https://forem.com/tahmid_khana_alim/ai-agents-are-productivity-theater-and-thats-fine-geh</guid>
      <description>&lt;p&gt;The era of AI agents has arrived. Or so we're told.&lt;/p&gt;

&lt;p&gt;If you've been anywhere near tech Twitter, Hacker News, or Reddit in the past year, you've seen the hype train. Autonomous AI agents will replace your workforce. They'll book your travel, answer your emails, write your code, and probably do your laundry if you ask nicely enough.&lt;/p&gt;

&lt;p&gt;The reality? Most AI agents are expensive toys solving imaginary problems.&lt;/p&gt;

&lt;p&gt;But here's the twist: &lt;strong&gt;the ones that actually work are changing how we build software&lt;/strong&gt;—just not in the way the marketing teams want you to believe.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Agent Hype Cycle: Where We Are Now
&lt;/h2&gt;

&lt;p&gt;February 2025 marked the peak of what I call "agent theater." xAI launched Grok 3, Google DeepMind shipped Veo 2, and every startup with a ChatGPT wrapper pivoted to calling themselves an "agentic AI platform."&lt;/p&gt;

&lt;p&gt;The demos were slick. The valuations were insane. The actual utility? Questionable.&lt;/p&gt;

&lt;p&gt;Here's what the research showed (and yes, I pulled this from real discussions on HN and Reddit):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Hardware is the real story:&lt;/strong&gt; Microsoft's quantum chip progress and Toyota's solid-state battery breakthrough got buried under AI noise&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Developer fatigue is real:&lt;/strong&gt; AI-generated documentation began outranking official docs in search results, making Stack Overflow practically unusable&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Security nightmare:&lt;/strong&gt; Gmail phishing scams using AI-cloned voices jumped &lt;strong&gt;300%&lt;/strong&gt; in Q1 2025&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The job market told the real story: entry-level generalist dev roles dropped &lt;strong&gt;25%&lt;/strong&gt; year-over-year, while specialized AI/ML and cloud security positions increased &lt;strong&gt;15%&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Translation: Companies aren't replacing developers with agents. They're hiring fewer generalists and more specialists to &lt;strong&gt;build and secure&lt;/strong&gt; agent systems.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Actually Works: The Boring Stuff
&lt;/h2&gt;

&lt;p&gt;Strip away the hype, and AI agents excel at three things:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Glorified Automation Scripts (With Context)
&lt;/h3&gt;

&lt;p&gt;The best agents I've seen aren't sentient workers. They're context-aware automation layers.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example:&lt;/strong&gt; A customer support agent that:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Reads your ticket history&lt;/li&gt;
&lt;li&gt;Checks your account status&lt;/li&gt;
&lt;li&gt;Pulls relevant docs&lt;/li&gt;
&lt;li&gt;Drafts a reply for a human to review&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Is this revolutionary? No. It's a smart database query + template engine.&lt;/p&gt;

&lt;p&gt;Is it useful? Hell yes. It cuts response time from 45 minutes to 3 minutes.&lt;/p&gt;

&lt;p&gt;The difference between this and a traditional workflow automation tool? The agent understands intent. You don't need to hardcode every possible ticket type—it generalizes from examples.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Natural Language as an Interface Layer
&lt;/h3&gt;

&lt;p&gt;This is where agents shine: &lt;strong&gt;making complex systems accessible without learning SQL, regex, or whatever arcane syntax your enterprise dashboard requires&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;You want last quarter's revenue broken down by region? Just ask.&lt;/p&gt;

&lt;p&gt;Previously, this required:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Finding the right dashboard (30 min)&lt;/li&gt;
&lt;li&gt;Remembering the filter syntax (10 min)&lt;/li&gt;
&lt;li&gt;Exporting to Excel because the UI is garbage (5 min)&lt;/li&gt;
&lt;li&gt;Manually aggregating because the export format is different than expected (15 min)&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Now? "Show me Q4 revenue by region" → instant Markdown table.&lt;/p&gt;

&lt;p&gt;The underlying data pipeline hasn't changed. The &lt;strong&gt;interface friction&lt;/strong&gt; disappeared.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Tedious, High-Volume Tasks Nobody Wants
&lt;/h3&gt;

&lt;p&gt;PR reviews for style violations. Scheduling meetings across six time zones. Parsing vendor invoices.&lt;/p&gt;

&lt;p&gt;These tasks don't need AGI. They need a tireless junior employee who doesn't get bored.&lt;/p&gt;

&lt;p&gt;Agents are perfect for this. They're consistent, they don't complain, and they cost pennies compared to human hours.&lt;/p&gt;

&lt;p&gt;But here's the catch: &lt;strong&gt;you still need humans to define "good."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;An agent can flag PRs with inconsistent naming. It can't decide whether your team's naming convention is stupid in the first place. That's still a human judgment call.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Most Agent Startups Will Fail
&lt;/h2&gt;

&lt;p&gt;The problem isn't technical capability. It's &lt;strong&gt;use case mismatch&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Most agent platforms are built like Swiss Army knives: technically impressive, but not actually great at anything specific.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example:&lt;/strong&gt; A "general-purpose" scheduling agent that can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Book flights&lt;/li&gt;
&lt;li&gt;Reserve restaurants&lt;/li&gt;
&lt;li&gt;Schedule meetings&lt;/li&gt;
&lt;li&gt;Order groceries&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Sounds amazing. In practice?&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Flight booking requires accessing your loyalty accounts (security nightmare)&lt;/li&gt;
&lt;li&gt;Restaurant preferences are hyper-personal and change based on mood/context&lt;/li&gt;
&lt;li&gt;Meeting scheduling needs org-specific rules (who can decline whom, internal vs external protocols)&lt;/li&gt;
&lt;li&gt;Grocery shopping involves dietary restrictions, brand preferences, and the fact that sometimes you just want junk food&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Each of these is a &lt;strong&gt;deep vertical problem&lt;/strong&gt;. A horizontal solution will be mediocre at all of them.&lt;/p&gt;

&lt;p&gt;The winners will be specialized agents solving &lt;strong&gt;one specific workflow&lt;/strong&gt; better than any human could.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Innovation: Agents as Infrastructure
&lt;/h2&gt;

&lt;p&gt;Here's where it gets interesting.&lt;/p&gt;

&lt;p&gt;The best use of AI agents isn't replacing jobs—it's &lt;strong&gt;replacing middleware&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Think about how modern web apps work:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Frontend calls API&lt;/li&gt;
&lt;li&gt;API validates request&lt;/li&gt;
&lt;li&gt;API queries database&lt;/li&gt;
&lt;li&gt;API formats response&lt;/li&gt;
&lt;li&gt;Frontend renders data&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Now imagine an agent layer that:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Interprets natural language queries&lt;/li&gt;
&lt;li&gt;Translates them to API calls&lt;/li&gt;
&lt;li&gt;Aggregates data from multiple sources&lt;/li&gt;
&lt;li&gt;Formats output based on user context&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;You've just replaced half your backend boilerplate with a reasoning layer.&lt;/p&gt;

&lt;p&gt;This is already happening. Perplexity and OpenAI's search prototypes aren't just "better Google"—they're &lt;strong&gt;API orchestration engines disguised as search&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;You ask: "What's the cheapest flight to Tokyo next week?"&lt;/p&gt;

&lt;p&gt;Behind the scenes:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Searches multiple airline APIs&lt;/li&gt;
&lt;li&gt;Cross-references with hotel availability&lt;/li&gt;
&lt;li&gt;Checks visa requirements&lt;/li&gt;
&lt;li&gt;Factors in your calendar (if integrated)&lt;/li&gt;
&lt;li&gt;Returns a synthesized answer with booking links&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;That's not search. That's a &lt;strong&gt;distributed system with a conversational interface&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Enshittification Problem
&lt;/h2&gt;

&lt;p&gt;Here's the dark side: as agents get better at &lt;em&gt;looking&lt;/em&gt; useful, they're also getting better at producing garbage.&lt;/p&gt;

&lt;p&gt;The "enshittification of documentation" is real. AI-generated tutorials are flooding search results, written by bots optimizing for SEO, not accuracy.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real example from Reddit:&lt;/strong&gt; A developer spent 2 hours debugging a Next.js issue using a top-ranked tutorial. Turns out, the tutorial was AI-generated, referenced outdated APIs, and had never been tested.&lt;/p&gt;

&lt;p&gt;The problem compounds:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;AI generates plausible-sounding content&lt;/li&gt;
&lt;li&gt;Google ranks it highly (good formatting, keywords, etc.)&lt;/li&gt;
&lt;li&gt;Humans read it, assume it's correct&lt;/li&gt;
&lt;li&gt;Other AIs scrape it as "training data"&lt;/li&gt;
&lt;li&gt;The cycle repeats&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;We're training future models on &lt;strong&gt;synthetic garbage generated by previous models&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;This is the "Dead Internet Theory" coming true—not through malice, but through &lt;strong&gt;incentive misalignment&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Developers Should Actually Care About
&lt;/h2&gt;

&lt;p&gt;Forget the hype. Here's what matters:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. &lt;strong&gt;Security Is the New Bottleneck&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;AI-powered social engineering is terrifyingly good. That 300% spike in phishing attacks? Just the beginning.&lt;/p&gt;

&lt;p&gt;If you're building agent systems, &lt;strong&gt;authentication and authorization&lt;/strong&gt; are your #1 priority. An agent with access to your email, calendar, and payment info is a single phishing attack away from disaster.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. &lt;strong&gt;Energy Costs Are Real&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Training Grok 3-class models consumes absurd amounts of power. The environmental impact is non-trivial.&lt;/p&gt;

&lt;p&gt;If you're deploying agents at scale, &lt;strong&gt;inference costs&lt;/strong&gt; will eat your margins. Optimize for efficiency, not capability.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. &lt;strong&gt;Job Market Is Polarizing&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;The "learn to code and get a junior dev job" pipeline is broken. Entry-level roles are shrinking because agents handle the grunt work.&lt;/p&gt;

&lt;p&gt;But specialized roles—AI/ML engineers, security architects, infrastructure specialists—are booming.&lt;/p&gt;

&lt;p&gt;The future isn't "everyone gets replaced." It's &lt;strong&gt;"generalists get squeezed, specialists get leverage."&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The Uncomfortable Truth
&lt;/h2&gt;

&lt;p&gt;AI agents aren't replacing knowledge workers. They're &lt;strong&gt;amplifying the gap&lt;/strong&gt; between those who know how to use them and those who don't.&lt;/p&gt;

&lt;p&gt;A skilled developer with an AI assistant can outproduce a team of 5 juniors. But a junior developer relying on AI-generated code without understanding the fundamentals? They're producing technical debt at scale.&lt;/p&gt;

&lt;p&gt;The same pattern applies everywhere:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A marketer with AI tools can A/B test hundreds of variants instantly&lt;/li&gt;
&lt;li&gt;A designer can prototype in minutes instead of hours&lt;/li&gt;
&lt;li&gt;A researcher can synthesize thousands of papers overnight&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But &lt;strong&gt;only if they know what good looks like&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Agents don't replace expertise. They &lt;strong&gt;multiply it&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where This Goes Next
&lt;/h2&gt;

&lt;p&gt;The next 12 months will separate the real innovations from the vaporware.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What will survive:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Specialized agents for deep verticals (legal, medical, financial analysis)&lt;/li&gt;
&lt;li&gt;Infrastructure-level agent systems (API orchestration, data aggregation)&lt;/li&gt;
&lt;li&gt;Security-first agent frameworks (zero-trust, sandboxed execution)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;What will fade:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;General-purpose "do everything" agents&lt;/li&gt;
&lt;li&gt;Consumer-facing scheduling/email bots (too much liability, too little margin)&lt;/li&gt;
&lt;li&gt;AI-first startups with no moat beyond a GPT wrapper&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Final Take
&lt;/h2&gt;

&lt;p&gt;AI agents aren't magic. They're &lt;strong&gt;probabilistic reasoning systems with API access&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;That's simultaneously less impressive than the hype suggests and more useful than the skeptics admit.&lt;/p&gt;

&lt;p&gt;The winners won't be the companies with the best demos. They'll be the ones solving &lt;strong&gt;specific, high-value problems&lt;/strong&gt; where automation was previously impossible.&lt;/p&gt;

&lt;p&gt;And developers? Your job isn't to compete with agents. It's to &lt;strong&gt;decide what they should automate, audit what they produce, and fix what they break&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;That's not going away anytime soon.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;Want to stay ahead of the AI agent curve?&lt;/strong&gt; Follow along as I break down the tools, frameworks, and strategies that actually matter. No hype, no bullshit—just practical insights for developers building in the agentic era.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://reboundbytes.top/blog/ai-agents-productivity-theater" rel="noopener noreferrer"&gt;Rebound Bytes&lt;/a&gt;. No fluff, just code.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>aiagents</category>
      <category>automation</category>
      <category>developertools</category>
      <category>productivity</category>
    </item>
    <item>
      <title>Waymo's Open Secret: The 'Robot' Taxi Is Just Outsourcing in a Trench Coat</title>
      <dc:creator>Tahmid Khan A</dc:creator>
      <pubDate>Mon, 09 Feb 2026 03:25:51 +0000</pubDate>
      <link>https://forem.com/tahmid_khana_alim/waymos-open-secret-the-robot-taxi-is-just-outsourcing-in-a-trench-coat-3534</link>
      <guid>https://forem.com/tahmid_khana_alim/waymos-open-secret-the-robot-taxi-is-just-outsourcing-in-a-trench-coat-3534</guid>
      <description>&lt;p&gt;There is a recurring theme in the history of Artificial Intelligence: if you pull back the curtain on a "magical" new technology, you inevitably find a human being sweating over a keyboard.&lt;/p&gt;

&lt;p&gt;First, it was Amazon's "Just Walk Out" technology, which turned out to be 1,000 people in India watching security cameras and manually adding items to your cart.&lt;/p&gt;

&lt;p&gt;Now, it's Waymo.&lt;/p&gt;

&lt;p&gt;New reports (and admissions from Waymo itself) have confirmed what skeptics have suspected for years: those "fully autonomous" robotaxis navigating San Francisco's complex streets aren't quite as autonomous as the marketing suggests. A significant portion of their "intelligence" is actually beam-streamed from remote centers where human operators—often in lower-cost labor markets like the Philippines—are guiding the cars through tricky situations.&lt;/p&gt;

&lt;p&gt;The stock market reacted with a shrug (Alphabet/GOOG is down ~2.5% today, but that's just noise). But for the developer community and the broader public, this is yet another crack in the "AI is Magic" narrative.&lt;/p&gt;

&lt;h2&gt;
  
  
  The "Human in the Loop" Reality
&lt;/h2&gt;

&lt;p&gt;To be technically fair to Waymo: these remote operators aren't "driving" the cars in the video game sense. They aren't holding a Logitech steering wheel and hitting the gas.&lt;/p&gt;

&lt;p&gt;The official term is "Remote Assistance" (RA). When the car gets confused—say, by a construction cone that looks like a person, or a police officer using hand signals—it stops and pings home. A human looks at the camera feed and clicks a path: &lt;em&gt;"Go around the cone on the left."&lt;/em&gt; The car then executes the maneuver itself.&lt;/p&gt;

&lt;p&gt;Waymo argues this is a safety feature. And it is.&lt;/p&gt;

&lt;p&gt;But it also fundamentally changes the economics of the business.&lt;/p&gt;

&lt;p&gt;If every "autonomous" ride requires 5 minutes of human attention to navigate a 20-minute trip, you haven't invented a robot taxi. You've invented a &lt;strong&gt;very expensive remote-controlled car.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The "Mechanical Turk" Economy
&lt;/h2&gt;

&lt;p&gt;This feeds into a darker narrative about the current AI boom: are we actually solving hard technical problems, or are we just finding new ways to hide low-wage labor behind an API?&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Amazon Just Walk Out:&lt;/strong&gt; "Computer Vision" = Humans watching video.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Customer Service AI:&lt;/strong&gt; "Chatbots" = Humans editing responses.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Data Labeling:&lt;/strong&gt; "Self-Supervised Learning" = Millions of workers in the Global South drawing boxes around stop signs for pennies.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;We are building a Rube Goldberg machine of labor arbitrage. We sell the service at a premium ("High Tech AI"), pay the labor at a discount ("Remote Gig Work"), and pocket the difference while telling investors we've solved General Intelligence.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why This Matters for Developers
&lt;/h2&gt;

&lt;p&gt;For those of us building software, this is a cautionary tale about &lt;strong&gt;marketing vs. architecture&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;We are being sold tools and platforms that promise full automation. "Autonomous Coding Agents" that will build your app. "Self-Driving" databases. "Zero-Touch" infrastructure.&lt;/p&gt;

&lt;p&gt;But in production, "autonomous" almost always means "autonomous until it breaks."&lt;/p&gt;

&lt;p&gt;The Waymo news is a reminder that &lt;strong&gt;edge cases are infinite&lt;/strong&gt;. You can train a model on a trillion miles of driving data, and it will still freeze when it sees a woman in a chicken costume chasing a duck across a crosswalk (a real thing that happens in San Francisco).&lt;/p&gt;

&lt;p&gt;Humans are the exception handlers. And until AI can handle the chicken-duck scenario without pinging Manila, we aren't replacing humans. We're just moving them to a call center.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Trust Erosion
&lt;/h2&gt;

&lt;p&gt;The real damage here isn't to Waymo's technology—which is objectively impressive—but to public trust.&lt;/p&gt;

&lt;p&gt;We were told the cars drive themselves. They don't.&lt;br&gt;
We were told the stores run themselves. They didn't.&lt;br&gt;
We were told the code writes itself. It doesn't (as my last post on AI productivity showed).&lt;/p&gt;

&lt;p&gt;We are entering the "Trough of Disillusionment" for the 2020s AI hype cycle. The magic is wearing off, and we're starting to see the wires.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Waymo is still the leader in this space. Their cars &lt;em&gt;mostly&lt;/em&gt; drive themselves. But the dream of zero-marginal-cost transportation—where a robot taxi costs pennies because no human needs to get paid—is dead for the foreseeable future.&lt;/p&gt;

&lt;p&gt;As long as there's a human in the loop, the cost floor is set by human wages. And as long as we pretend otherwise, we're just lying to ourselves.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Stock Watch:&lt;/strong&gt; Alphabet (GOOG) sits at &lt;strong&gt;$323.10&lt;/strong&gt;, down slightly. The market knows that even with human helpers, Waymo is miles ahead of Tesla. But "miles ahead" doesn't mean "finished."&lt;/p&gt;

&lt;p&gt;&lt;em&gt;What do you think? Is "Remote Assistance" cheating, or just smart engineering? Let me know.&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://reboundbytes.top/blog/waymo-remote-operators-scandal" rel="noopener noreferrer"&gt;Rebound Bytes&lt;/a&gt;. No fluff, just code.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>waymo</category>
      <category>automation</category>
      <category>technews</category>
    </item>
  </channel>
</rss>
