<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Denis Stetskov</title>
    <description>The latest articles on Forem by Denis Stetskov (@razoorka).</description>
    <link>https://forem.com/razoorka</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1050470%2Fc5743404-1030-4439-b7c6-8bf21c71692c.jpg</url>
      <title>Forem: Denis Stetskov</title>
      <link>https://forem.com/razoorka</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/razoorka"/>
    <language>en</language>
    <item>
      <title>The Country of Geniuses That Doesn’t Exist</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Wed, 18 Feb 2026 09:43:00 +0000</pubDate>
      <link>https://forem.com/razoorka/the-country-of-geniuses-that-doesnt-exist-5dkh</link>
      <guid>https://forem.com/razoorka/the-country-of-geniuses-that-doesnt-exist-5dkh</guid>
      <description>&lt;p&gt;On January 26, 2026, Anthropic CEO Dario Amodei published a 20,000‑word essay predicting a “country of geniuses in a datacenter” within 1–2 years — millions of entities, each smarter than any Nobel Prize winner, and 50% of entry‑level white‑collar jobs disrupted within 1–5 years.&lt;/p&gt;

&lt;p&gt;5.7 million views on X. Standing ovation from investors. I only got around to reading it now. I have things to say.&lt;/p&gt;

&lt;p&gt;I’m disappointed to watch Amodei and Anthropic slide into Altman‑ism. Different prose, same playbook.&lt;/p&gt;

&lt;p&gt;Maybe where the gods live, he’s right. Maybe in a world of perfect infrastructure, clean APIs, and unlimited compute, we’re ready to replace white‑collar workers with AI. But where the rest of us mortals work, the situation looks completely different.&lt;/p&gt;

&lt;p&gt;His own product’s System Card tells a different story. Anthropic surveyed internal researchers on whether Claude could replace an entry‑level researcher with three months of scaffolding. The answer was 0 out of 16. Zero.&lt;/p&gt;

&lt;p&gt;We’ve spent four years shipping AI integrations for clients. The models are impressive. They are not replacing white‑collar workers. Not in 1–2 years. Probably not in 5. And the reasons are more fundamental than the industry wants to admit.&lt;/p&gt;

&lt;p&gt;Let’s talk about what transformers actually can’t do. Not philosophically. Mathematically.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Steering Wheel Problem
&lt;/h2&gt;

&lt;p&gt;Non‑determinism. Even at temperature zero, the same prompt produces different outputs. This isn’t a bug. It’s a consequence of floating‑point parallel computation on GPUs. In engineering, we call components that behave unpredictably under identical conditions broken.&lt;/p&gt;

&lt;p&gt;Hallucinations are provably inevitable. Formal proofs from learning theory show LLMs cannot learn all computable functions and will hallucinate when used as general‑purpose problem solvers. Best models: 15%+ hallucination rate on benchmarks. GPTZero found over 50 hallucinated citations in ICLR 2026 academic submissions. Trained peer reviewers, 3–5 per paper, didn’t catch them.&lt;/p&gt;

&lt;p&gt;Function composition has limits. Transformers struggle with real‑world function composition due to how softmax limits non‑local information flow. In practice, models write connected code fine. What they can’t do is reason about infrastructure constraints: what’s possible and what isn’t, where the boundaries are.&lt;/p&gt;

&lt;p&gt;I see this every day. Smart autocomplete. Incredibly good smart autocomplete. But autocomplete that can’t tell you when it’s wrong.&lt;/p&gt;

&lt;p&gt;The industry knows. They’ve quietly shifted from “let’s eliminate hallucinations” to “let’s manage uncertainty.” That’s a de facto admission: the steering wheel sometimes turns the wrong way, and nobody can fix it.&lt;/p&gt;

&lt;p&gt;It’s like selling an airplane whose steering sometimes inverts, then writing 20,000 words about how the airplane might fly to another galaxy. Bioweapons and autocracy get entire sections. The steering wheel? Not mentioned once.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Scaling Wall Nobody Advertises
&lt;/h2&gt;

&lt;p&gt;Maybe more compute fixes it? That’s been the bet for five years.&lt;/p&gt;

&lt;p&gt;Toby Ord actually read the scaling law graphs that AI companies publish with great fanfare. On log‑log charts, the lines look beautiful. Flip to linear scale and halving the model’s error rate requires increasing compute by a factor of a million.&lt;/p&gt;

&lt;p&gt;Three walls are converging simultaneously:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Data:&lt;/strong&gt; High‑quality training text is finite.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Compute:&lt;/strong&gt; Latency constraints, energy consumption exceeding entire countries, new data center connections that take 2–4 years.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Architecture:&lt;/strong&gt; The mathematical limitations above aren’t going away with more parameters.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Ilya Sutskever told Reuters the scaling era is over. We’re in an “age of wonder and discovery.” Translation: we don’t know what’s next.&lt;/p&gt;

&lt;p&gt;HEC Paris calls this the industry’s “well‑kept secret.” MIT research from January 2026 confirms: the gap between expensive frontier models and cheaper alternatives is shrinking. Exponentially more expensive, single‑digit percentage improvements.&lt;/p&gt;

&lt;p&gt;The 650 billion dollars Big Tech is pouring into infrastructure this year? As I wrote in my analysis of that spending: it’s not investment. It’s capitulation.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Context Problem: 150 Projects Worth of Evidence
&lt;/h2&gt;

&lt;p&gt;Here’s what Amodei’s essay gets wrong. This is what I see every week.&lt;/p&gt;

&lt;p&gt;Clients come to us with the same request: “We want to integrate AI into our processes.” Replace the white‑collar workers. Cut the headcount.&lt;/p&gt;

&lt;p&gt;So why can’t we sell them the same project?&lt;/p&gt;

&lt;p&gt;Because zero companies have the same structure. Zero run the same systems.&lt;/p&gt;

&lt;p&gt;One client runs SharePoint from 2007. Another has a custom CRM built by a contractor who left in 2015. No documentation. No API. A third uses SSO held together with duct tape and prayer. Company D has critical data in Excel spreadsheets that get emailed between departments every Friday afternoon.&lt;/p&gt;

&lt;p&gt;Amodei writes from a world where every organization has MCP‑ready infrastructure, clean data pipelines, standardized APIs. That world doesn’t exist.&lt;/p&gt;

&lt;p&gt;To replace a white‑collar worker, AI needs full organizational context. Approval chains. Informal relationships. Institutional knowledge that lives in people’s heads. The exception to the exception. The vendor who says two weeks but means six.&lt;/p&gt;

&lt;p&gt;Who gives the model that context?&lt;/p&gt;

&lt;p&gt;A human. A skilled human. The exact white‑collar worker you’re trying to replace.&lt;/p&gt;

&lt;p&gt;This is the paradox nobody discusses: the knowledge required to supervise AI effectively is the same knowledge that makes you irreplaceable.&lt;/p&gt;

&lt;h2&gt;
  
  
  Already Deployed Where Errors Kill
&lt;/h2&gt;

&lt;p&gt;While the “country of geniuses” narrative plays out on Twitter, these architecturally unreliable systems are already making decisions about health, money, and legal rights. The promise was improvement. The results are instructive.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Healthcare.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
The pitch: faster diagnoses, better outcomes, lower costs.&lt;br&gt;&lt;br&gt;
The reality: UnitedHealth and Humana face class‑action lawsuits over nH Predict, an AI model that denied Medicare coverage against doctors’ recommendations. Known high error rate. Deployed anyway. 21 states passed emergency laws regulating AI in healthcare. 250+ bills introduced across U.S. states. Not because AI improved care. Because it made denial of care faster and harder to appeal.&lt;/p&gt;

&lt;p&gt;The accountability gap: doctor says “developer is responsible.” Developer says “doctor makes the decision.” Nobody owns the failure. Patients own the consequences.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Finance.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
The pitch: smarter markets, better allocation, reduced risk.&lt;br&gt;&lt;br&gt;
The reality: AI trading makes markets more volatile, not more efficient. IMF confirmed it. GARCH modeling on S&amp;amp;P 500 shows positive association between AI trading and increased market jumps. Thousands of models trained on the same data, processing the same Fed minutes in milliseconds, create herd behavior at machine speed. We didn’t get efficient markets. We got synchronized panic.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Legal.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
The pitch: democratize access to justice, reduce costs.&lt;br&gt;&lt;br&gt;
The reality: in 2025 alone, judges worldwide issued hundreds of decisions addressing AI hallucinations in legal filings. Roughly 90% of all known cases to date. Fabricated citations in a profession where one fake precedent can destroy a career. Justice didn’t get cheaper. It got less reliable.&lt;/p&gt;

&lt;p&gt;Three industries. Three promises of improvement. Three measurable deteriorations. With models that their own creators admit cannot be made deterministic.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Nobody Says This Out Loud
&lt;/h2&gt;

&lt;p&gt;Simple. Everyone has reasons to stay quiet.&lt;/p&gt;

&lt;p&gt;AI companies can’t say “our technology is architecturally unreliable.” Valuation event.&lt;/p&gt;

&lt;p&gt;Investors deployed over a trillion dollars. You don’t question the thesis after you’ve bet the fund.&lt;/p&gt;

&lt;p&gt;Media runs on attention. “AI will replace everyone” gets clicks. “AI has fundamental mathematical limitations” doesn’t.&lt;/p&gt;

&lt;p&gt;And here’s what keeps me up at night. Amodei writes 20,000 words about risks — bioweapons, autocracy, existential threats. Not once does he mention the most fundamental risk: the absence of determinism.&lt;/p&gt;

&lt;p&gt;A non‑deterministic system cannot be trusted as a reliable autonomous agent. Period. Everything else is commentary.&lt;/p&gt;

&lt;h2&gt;
  
  
  What You Should Actually Do
&lt;/h2&gt;

&lt;p&gt;AI isn’t useless. Saying that would be as dishonest as saying it replaces half the workforce.&lt;/p&gt;

&lt;p&gt;I use it every day. My team uses it on every project. The value is real. But it’s specific.&lt;/p&gt;

&lt;p&gt;AI saves 20–40% of a qualified specialist’s time. Someone who knows what to ask, how to verify, and when the model is confidently wrong.&lt;/p&gt;

&lt;p&gt;Not replacement. Amplification of existing expertise.&lt;/p&gt;

&lt;p&gt;So what do you actually do?&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Increase your value.&lt;/strong&gt; Understand your domain &lt;em&gt;and&lt;/em&gt; AI’s real capabilities — not the theoretical ones from a CEO’s essay, the real ones you discover using the tool daily.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Make decisions.&lt;/strong&gt; AI can’t weigh trade‑offs. Can’t navigate org politics. Can’t choose between two valid approaches based on team capabilities and timeline. SQL vs. NoSQL. Monolith vs. microservices. These require judgment. Judgment requires experience. Experience requires years of being wrong.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Be the expert.&lt;/strong&gt; Deep domain knowledge is your moat. Not surface familiarity. The kind where you smell a wrong answer before you can articulate why.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Don’t outsource your brain.&lt;/strong&gt; Every task you hand entirely to AI is a skill you stop developing. Every decision you let the model make is judgment you’re not exercising. Do this long enough and you’re on the wrong side of the equation when the company realizes the tool needs a supervisor, not a passenger.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;When the hype deflates, the question will be: “Okay, so what do we actually do with this technology?” Practitioners will answer that. Not evangelists.&lt;/p&gt;

&lt;p&gt;The country of geniuses doesn’t exist. What exists is a powerful tool that requires skilled humans to operate safely. Don’t let a 20,000‑word essay convince you the steering wheel doesn’t matter just because the destination sounds exciting.&lt;/p&gt;

&lt;p&gt;Are the AI predictions from leadership matching the engineering reality you see on the ground?&lt;/p&gt;

&lt;p&gt;If this resonated, forward it to an engineering leader who needs to hear it.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>career</category>
      <category>productivity</category>
      <category>discuss</category>
    </item>
    <item>
      <title>RAG Is Easy. Your Data Isn't. Why AI Projects Fail</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Mon, 02 Feb 2026 12:42:01 +0000</pubDate>
      <link>https://forem.com/razoorka/rag-is-easy-your-data-isnt-why-ai-projects-fail-5gan</link>
      <guid>https://forem.com/razoorka/rag-is-easy-your-data-isnt-why-ai-projects-fail-5gan</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd0zxx1p6dhre8kjylg0q.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd0zxx1p6dhre8kjylg0q.png" alt="Image description='infographic'" width="800" height="500"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I joined a discovery call. The brief beforehand: "This is basically a copy of Project X. Same timeline."&lt;/p&gt;

&lt;p&gt;Project X was a marketing chatbot. Conversational, no proprietary knowledge base. Search integration and personality. We knew that scope.&lt;/p&gt;

&lt;p&gt;Thirty minutes into the call, it's clear this isn't RAG. Data processing from S3 buckets, Lambda triggers, ETL pipeline. That's table stakes. The real work? Teaching the model to query and reason over that structured data. That's not a chatbot. That's a different project entirely.&lt;/p&gt;

&lt;p&gt;"Same timeline" for a completely different architecture.&lt;/p&gt;

&lt;p&gt;This happens constantly. Not because clients mislead us. Because the gap between "AI chatbot" in their head and "AI chatbot" in reality is massive.&lt;/p&gt;

&lt;p&gt;The pattern is clear: most projects don't struggle because the engineering is hard. They struggle because everyone underestimates what comes before the engineering starts.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Custom GPT Problem
&lt;/h2&gt;

&lt;p&gt;Client built a Custom GPT over a weekend. Uploaded some PDFs. Asked it questions. It worked. They showed their CEO. Everyone got excited.&lt;/p&gt;

&lt;p&gt;"We want this, but for the whole company."&lt;/p&gt;

&lt;p&gt;That's where it stops being simple.&lt;/p&gt;

&lt;p&gt;"For the whole company" means multi-tenancy. Different departments see different data. Role-based retrieval: sales can't access HR documents, legal can't see engineering specs. Audit logs. Access controls. Compliance.&lt;/p&gt;

&lt;p&gt;Custom GPT doesn't do any of that. It's one user, one knowledge base, no permissions. The jump from "it works for me" to "it works for the organization" isn't a small step. It's a different architecture.&lt;/p&gt;

&lt;p&gt;NotebookLLM, Custom GPTs. They create a dangerous illusion. They make AI feel simple because all the enterprise complexity is hidden. The prototype took a weekend. The production system takes months.&lt;/p&gt;

&lt;h2&gt;
  
  
  "We Have Data": The Three Versions
&lt;/h2&gt;

&lt;p&gt;Every client says they have data. They mean different things.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Version 1: "We have documents."&lt;/strong&gt; They have PDFs. Some are text. Some are scans. Some are text with scanned tables embedded. Some are PowerPoints where the real information lives in speaker notes nobody exports.&lt;/p&gt;

&lt;p&gt;This isn't a data problem you solve once. It's a classification problem, an OCR problem, a parsing problem, and then a chunking problem. Each one adds weeks.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Version 2: "We have structured data."&lt;/strong&gt; They have databases. Multiple databases. With different schemas. Some legacy system from 2012 that nobody fully understands anymore. CSV exports that break because someone used commas in a text field.&lt;/p&gt;

&lt;p&gt;Now you're not building RAG. You're building SQL agents, data transformation pipelines, and schema mapping. Different architecture entirely.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Version 3: "We have both."&lt;/strong&gt; Documents and databases and spreadsheets and emails and a SharePoint nobody's organized in years.&lt;/p&gt;

&lt;p&gt;This is the most common version. And the most underestimated.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Access Tax
&lt;/h2&gt;

&lt;p&gt;Data and credentials need to arrive on day one. They rarely do.&lt;/p&gt;

&lt;p&gt;We've waited weeks for database access. Months for IT security approvals. One project stalled because a single stakeholder controlled API credentials and went on vacation.&lt;/p&gt;

&lt;p&gt;Every week of waiting is a week of zero progress. But in the client's mind, the timeline keeps running from the day they signed the contract.&lt;/p&gt;

&lt;p&gt;The access problem isn't technical. It's organizational. And organizations move slowly.&lt;/p&gt;

&lt;h2&gt;
  
  
  Two Types of Clients
&lt;/h2&gt;

&lt;p&gt;We can predict project outcomes from the first call.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Clients who know their bottleneck:&lt;/strong&gt; "We spend 40 hours weekly on this specific process. Here are inputs and outputs. Here's the domain expert who'll validate results."&lt;/p&gt;

&lt;p&gt;These projects ship. Clear scope, measurable outcome, someone internal who can evaluate accuracy.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Clients who want AI everywhere:&lt;/strong&gt; "We want to optimize our processes. We're not sure which ones yet."&lt;/p&gt;

&lt;p&gt;These projects stall. Not because AI can't help. Because you can't optimize processes that aren't documented. You can't measure improvement without baselines. You can't validate AI outputs without domain expertise.&lt;/p&gt;

&lt;p&gt;The technology isn't the constraint. Organizational readiness is.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Work That Isn't Ours
&lt;/h2&gt;

&lt;p&gt;Here's what successful projects require from the client side:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Domain expertise for validation.&lt;/strong&gt; We build the system. We cannot tell you if the output is correct for your industry, your regulations, your edge cases. That's your job.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Evaluation data.&lt;/strong&gt; Before we write code, we need examples: "When users ask X, good answers look like Y." Hundreds of them. This is how we measure progress versus confident wrongness.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Accuracy decisions.&lt;/strong&gt; 85% accuracy in 6 weeks. 95% might take another 6 weeks. 99% might be impossible with your data quality. Those last 5% for 2% of users might cost 40% of the budget. You decide if it's worth it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Ongoing maintenance.&lt;/strong&gt; When source documents change, someone updates them. When accuracy drifts, someone investigates. This isn't a one-time build. It's an ongoing operation.&lt;/p&gt;

&lt;p&gt;Most clients expect to hand off requirements and receive a product. AI doesn't work that way. It's a collaboration that requires their continuous involvement.&lt;/p&gt;

&lt;h2&gt;
  
  
  Simple Project, Real Timeline
&lt;/h2&gt;

&lt;p&gt;Best case scenario. Clean data, clear scope, engaged stakeholder with domain knowledge.&lt;/p&gt;

&lt;p&gt;6-8 weeks. Most of that time goes to prompt engineering and iteration. Not infrastructure.&lt;/p&gt;

&lt;p&gt;But "clean data" is rare. "Clear scope" requires work upfront. "Engaged stakeholder" means someone's calendar is blocked for this project, not squeezed between other priorities.&lt;/p&gt;

&lt;p&gt;When any of these are missing, multiply the timeline. When all three are missing, reconsider starting.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Projects Don't Reach Production
&lt;/h2&gt;

&lt;p&gt;Projects rarely fail technically. They fail organizationally.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Built but never integrated.&lt;/strong&gt; We deliver a working system. It sits in staging because the client doesn't have engineering resources to integrate it. They budgeted for building, not deploying.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Value mismatch discovered late.&lt;/strong&gt; Midway through, the client realizes the problem they described isn't their actual pain point. The AI works. The business case didn't.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Diminishing returns rejected.&lt;/strong&gt; We explain the math: last 5% of accuracy for edge cases costs 40% of remaining budget. They want it anyway. Then budget runs out. Then the project is "over scope."&lt;/p&gt;

&lt;p&gt;None of these are engineering problems.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Actually Helps
&lt;/h2&gt;

&lt;p&gt;Before signing contracts, dig into the actual data. Not descriptions of data. The data itself.&lt;/p&gt;

&lt;p&gt;We run a Rapid Validation Sprint. Four weeks. Real data access, real complexity mapping, real unknowns identified. Then we estimate based on reality, not assumptions.&lt;/p&gt;

&lt;p&gt;The companies who quote 50% less aren't doing this work. They're guessing. When the data turns out messier than expected (it always does), they either blow the budget or cut scope.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Point
&lt;/h2&gt;

&lt;p&gt;RAG tutorials make this look easy. Upload documents, chunk them, embed them, query them. Done.&lt;/p&gt;

&lt;p&gt;Production is different. Data is messy. Access is slow. Validation requires domain expertise you don't have. Accuracy expectations exceed what the data supports.&lt;/p&gt;

&lt;p&gt;The engineering is the straightforward part. Everything that comes before it: that's where projects actually succeed or fail.&lt;/p&gt;

&lt;p&gt;Most AI initiatives struggle not because the technology isn't ready. Because the organization isn't ready. Data isn't organized. Processes aren't documented. Nobody's assigned to validate outputs.&lt;/p&gt;

&lt;p&gt;That's not a criticism. It's just the reality.&lt;/p&gt;

&lt;p&gt;The question isn't whether AI can help your business. It's whether your business is ready to help the AI.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;What's been your experience with AI project expectations versus reality? Comment below, I read every response.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;If this resonates, share it with someone about to sign an AI contract. Better they hear this now.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>webdev</category>
      <category>rag</category>
    </item>
    <item>
      <title>The First Full-Scale Cyber War: 4 Years of Lessons</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Sun, 25 Jan 2026 14:25:17 +0000</pubDate>
      <link>https://forem.com/razoorka/the-first-full-scale-cyber-war-4-years-of-lessons-2mgd</link>
      <guid>https://forem.com/razoorka/the-first-full-scale-cyber-war-4-years-of-lessons-2mgd</guid>
      <description>&lt;p&gt;December 12, 2023. 7:00 AM Kyiv time. Kyivstar, Ukraine's largest mobile operator with 24.3 million subscribers, goes silent.&lt;/p&gt;

&lt;p&gt;Mobile service. Internet. Air raid alert systems in Kyiv and Sumy regions. All offline.&lt;/p&gt;

&lt;p&gt;Within hours, Sandworm hackers destroyed 10,000 computers, 4,000+ servers, all cloud storage, and backups.&lt;/p&gt;

&lt;p&gt;Illia Vitiuk, head of SBU's cybersecurity department: &lt;em&gt;"This is probably the first example of a destructive cyberattack that completely destroyed the core of a telecoms operator."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;The hackers had been inside since May 2023. Full access since November. Seven months inside the infrastructure of a country's largest carrier. Nobody noticed.&lt;/p&gt;

&lt;p&gt;This wasn't an isolated incident. &lt;strong&gt;This is the first full-scale cyber war in history.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;And the lessons apply to every power grid, railway system, and telecom provider worldwide.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Scale Nobody Talks About
&lt;/h2&gt;

&lt;p&gt;Between 2022 and 2024, Ukraine recorded over 9,000 cyber incidents. The trajectory, according to SSSCIP data reported by Infosecurity Magazine:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;2021:&lt;/strong&gt; 1,350 incidents&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;2024:&lt;/strong&gt; 4,315 incidents&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Growth:&lt;/strong&gt; 220% in three years&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Russia deployed 17+ unique wiper malware families: programs designed solely to destroy data beyond recovery. WhisperGate, HermeticWiper, CaddyWiper, Industroyer2, AcidRain. Each built for specific targets.&lt;/p&gt;

&lt;p&gt;But here's what Western coverage often misses: &lt;strong&gt;this isn't one-sided aggression.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Ukraine hit back. Hard.&lt;/p&gt;

&lt;p&gt;In July 2024, Ukraine's military intelligence (GUR) claimed responsibility for a week-long DDoS attack on Russia's banking system. Sberbank, Alfa-Bank, VTB, Gazprombank, the Central Bank. Users reportedly couldn't withdraw cash from ATMs.&lt;/p&gt;

&lt;p&gt;In December 2025, anonymous hackers breached Mikord, a developer of Russia's unified military draft registry. 30 million records. Source code, documentation, backups destroyed, according to investigative outlet iStories, which verified the breach. Mikord's director confirmed the attack. Russia's Defense Ministry denied any impact on the registry.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;This is symmetric warfare. Both sides are hitting critical infrastructure. Both sides claim real damage.&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  The Attacks That Changed Everything
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Viasat: The Hour-Zero Strike
&lt;/h3&gt;

&lt;p&gt;February 24, 2022. 03:02 UTC. Exactly one hour before Russian ground forces crossed the border.&lt;/p&gt;

&lt;p&gt;Attackers exploited a VPN misconfiguration at Viasat's management center in Turin, Italy. They pushed AcidRain wiper malware to 40,000-45,000 satellite modems via legitimate software update mechanisms.&lt;/p&gt;

&lt;p&gt;The result: Ukrainian military command and control went dark at the moment of invasion. Spillover disabled 5,800 German wind turbines and affected 9,000 French subscribers.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;One misconfigured VPN. 45,000 modems bricked. Military communications disrupted during the most critical hour of the war.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;SentinelOne researchers called it &lt;em&gt;"the biggest known hack of the war."&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Industroyer2: The Blackout That Almost Was
&lt;/h3&gt;

&lt;p&gt;April 8, 2022. ESET researchers discovered Industroyer2 scheduled to execute at 16:10 UTC against Ukrainian electrical substations. CaddyWiper was programmed to run 10 minutes later to destroy forensic evidence.&lt;/p&gt;

&lt;p&gt;The malware implemented IEC 60870-5-104, the protocol used by electrical substation protection relays. It contained hardcoded IP addresses for eight target ICS devices.&lt;/p&gt;

&lt;p&gt;If successful: a blackout affecting over 2 million people. The largest cyber-induced power outage in history.&lt;/p&gt;

&lt;p&gt;It failed. CERT-UA, ESET, and Microsoft coordinated a defense based on lessons from the 2016 grid attack. The attack was stopped hours before execution.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The pattern: preparation from crisis one saved crisis two.&lt;/strong&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Kyivstar: When Security Investment Isn't Enough
&lt;/h3&gt;

&lt;p&gt;Kyivstar wasn't some underfunded government agency. It was Ukraine's largest private telecom, a subsidiary of Amsterdam-based VEON, with serious security investment.&lt;/p&gt;

&lt;p&gt;Didn't matter.&lt;/p&gt;

&lt;p&gt;Sandworm penetrated the network in March 2023. By November, they had full access. On December 12, they executed, destroying the core infrastructure, wiping &lt;em&gt;"almost everything."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Vitiuk's assessment: &lt;em&gt;"This attack is a big message, a big warning, not only to Ukraine but for the whole Western world to understand that no one is actually untouchable."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;40% of Kyivstar's infrastructure disabled. Services restored in phases over eight days. Losses estimated in the billions of hryvnia.&lt;/p&gt;

&lt;h3&gt;
  
  
  Ukrzaliznytsia: March 2025
&lt;/h3&gt;

&lt;p&gt;With Ukrainian airspace closed since 2022, railways became the country's lifeline. 20 million passengers and 148 million tonnes of freight in 2024.&lt;/p&gt;

&lt;p&gt;On March 23, 2025, a &lt;em&gt;"large-scale, systematic, non-trivial and multi-level"&lt;/em&gt; cyberattack hit Ukrzaliznytsia's online systems. CERT-UA investigation found TTPs &lt;em&gt;"characteristic of Russian intelligence services."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Website and mobile app: offline. Long queues at physical ticket offices.&lt;/p&gt;

&lt;p&gt;But trains never stopped running.&lt;/p&gt;

&lt;p&gt;The difference from Kyivstar: backup protocols implemented after previous attacks. Systems built during crisis one carried through crisis two.&lt;/p&gt;

&lt;p&gt;CEO Oleksandr Pertsovskyi: &lt;em&gt;"The cyber-attack on the company was targeted and meticulously planned. However, not a single Ukrzaliznytsia train was halted for even a moment."&lt;/em&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Ukraine's Counter-Offensive
&lt;/h2&gt;

&lt;p&gt;Western media focuses on Russian attacks. The Ukrainian response gets less attention. The following operations were claimed by GUR or pro-Ukrainian hackers. Independent verification varies, and Russian authorities have denied most claims.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Tax Service, December 2023.&lt;/strong&gt; GUR claimed it destroyed databases across 2,300+ regional servers. Configuration files &lt;em&gt;"which for years ensured the functioning of Russia's tax system"&lt;/em&gt; allegedly wiped. Russia's Federal Tax Service denied any operational impact, though users reported access problems.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Planeta, January 2024.&lt;/strong&gt; GUR claimed an attack on a state satellite data center. 280 servers allegedly destroyed. 2 petabytes of military-relevant weather and satellite data reportedly wiped. Supercomputers &lt;em&gt;"not fully restorable due to sanctions."&lt;/em&gt; Claimed damage: $10+ million.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Banking System, July 2024.&lt;/strong&gt; GUR claimed a week-long DDoS campaign targeting Sberbank, Alfa-Bank, VTB, Gazprombank, Central Bank, plus VK, Discord, and the national payment system. Reports indicated ATM disruptions across Russia.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Russian Railways, March 2024 &amp;amp; June 2025.&lt;/strong&gt; Multiple attacks reportedly taking down RZD's website and app. Moscow Metro hit days after Ukrzaliznytsia attack in apparent retaliation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Mikord Draft Registry, December 2025.&lt;/strong&gt; Anonymous hackers (not attributed to GUR) breached Mikord, a key developer of Russia's unified military registration system. The Moscow Times and iStories verified the breach. Mikord's director confirmed the hack. The registry contains 30 million conscription records. Source code, documentation, and backups reportedly destroyed. Russia's Defense Ministry called the reports &lt;em&gt;"fake news."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Grigory Sverdlin, anti-conscription organization Idite Lesom: &lt;em&gt;"For several more months, this behemoth won't be able to send people off to kill and die."&lt;/em&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  The Vulnerability Patterns
&lt;/h2&gt;

&lt;p&gt;Four years of cyber warfare exposed consistent vulnerability classes. These aren't Ukraine-specific. They exist in Western infrastructure.&lt;/p&gt;

&lt;h3&gt;
  
  
  VPN and Remote Access
&lt;/h3&gt;

&lt;p&gt;The Viasat attack exploited a VPN misconfiguration. Kyivstar's breach likely started with a compromised employee account. CISA documented GRU exploitation of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;CVE-2018-13379 (FortiGate)&lt;/li&gt;
&lt;li&gt;CVE-2019-11510 (Pulse Secure)&lt;/li&gt;
&lt;li&gt;CVE-2019-19781 (Citrix)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Vulnerabilities with patches available for 5+ years.&lt;/p&gt;

&lt;h3&gt;
  
  
  Dwell Time
&lt;/h3&gt;

&lt;p&gt;Kyivstar: attackers inside for 7 months before execution. October 2022 power grid attack: Mandiant found attackers with SCADA access for up to three months.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Sophisticated adversaries don't rush. Detection capabilities that can't identify months-long intrusions are detection capabilities that don't work.&lt;/strong&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Supply Chain
&lt;/h3&gt;

&lt;p&gt;The Viasat attack weaponized legitimate software update mechanisms. CERT-UA documented at least three supply chain breaches in March 2024 energy sector attacks.&lt;/p&gt;

&lt;h3&gt;
  
  
  IT/OT Convergence
&lt;/h3&gt;

&lt;p&gt;The October 2022 grid attack gained OT access through a hypervisor hosting a SCADA management instance. Attackers used native MicroSCADA binaries, living-off-the-land techniques. Mandiant: &lt;em&gt;"a growing maturity of Russia's offensive OT arsenal."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Victor Zhora, SSSCIP Deputy Chairman, emphasized air-gapping between IT and OT as fundamental. Most Western utilities have moved in the opposite direction.&lt;/p&gt;

&lt;h3&gt;
  
  
  Centralization
&lt;/h3&gt;

&lt;p&gt;The Mikord hack illustrates the pattern: centralization creates single points of failure.&lt;/p&gt;

&lt;p&gt;Ukraine's cloud migration (15+ petabytes distributed across AWS, Google Cloud, Microsoft Azure) proved more resilient than hardened on-premises facilities.&lt;/p&gt;

&lt;p&gt;Deputy Prime Minister Mykhailo Fedorov: &lt;em&gt;"Russian missiles can't destroy the cloud."&lt;/em&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  What Actually Worked
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Cloud Migration
&lt;/h3&gt;

&lt;p&gt;One week before the invasion, Ukraine's parliament enabled government data migration to cloud. PrivatBank (20 million customers) migrated 270 applications and 4 petabytes in 45 days. Financial services continued throughout the war.&lt;/p&gt;

&lt;h3&gt;
  
  
  Detection Speed
&lt;/h3&gt;

&lt;p&gt;Microsoft detected HermeticWiper hours before the invasion. Within 3 hours, signatures pushed globally. The Industroyer2 defense succeeded because CERT-UA, ESET, and Microsoft coordinated based on 2016 lessons.&lt;/p&gt;

&lt;h3&gt;
  
  
  Backup Protocols
&lt;/h3&gt;

&lt;p&gt;Ukrzaliznytsia's trains ran during attack because they'd been attacked before. Kyivstar took eight days to restore. The difference: systems built during previous crises.&lt;/p&gt;

&lt;h3&gt;
  
  
  Public-Private Partnership
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Microsoft:&lt;/strong&gt; $400+ million in aid&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Google:&lt;/strong&gt; Project Shield on 150+ websites&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cloudflare:&lt;/strong&gt; ~130 government domains&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AWS:&lt;/strong&gt; Snowball devices shipped to Poland within 48 hours&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Carnegie Endowment: &lt;em&gt;"delivering cyber defense at scale could only be achieved by private sector entities that owned, operated, and understood the most widely-used digital services."&lt;/em&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  The Human Layer
&lt;/h2&gt;

&lt;p&gt;Every major attack in this analysis started the same way: &lt;strong&gt;a person.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Kyivstar: likely a compromised employee account. Viasat: a VPN misconfiguration someone didn't catch. The GRU exploits from 2018 and 2019 still work because someone hasn't patched systems that have had fixes for five years.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Nation-state attackers don't need zero-days when humans provide the access.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I manage distributed engineering teams from a US-based company, with engineers in Ukraine. We've operated through four years of this war. Our security isn't optional:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Mandatory quarterly security training&lt;/li&gt;
&lt;li&gt;BYOD policy with device management&lt;/li&gt;
&lt;li&gt;Password policy with breach monitoring&lt;/li&gt;
&lt;li&gt;2FA on everything without exceptions&lt;/li&gt;
&lt;li&gt;Access reviews when roles change&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;None of this is exotic. All of it is enforced. The same principle applies to AI safety. Different domain, same lesson: &lt;strong&gt;policies that aren't enforced aren't policies.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The difference between &lt;em&gt;"we have a policy"&lt;/em&gt; and &lt;em&gt;"the policy is mandatory"&lt;/em&gt; is the difference between Kyivstar and Ukrzaliznytsia.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The companies that survived had one thing in common: policies that were actually followed, not just documented.&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  What This Means for Everyone Else
&lt;/h2&gt;

&lt;p&gt;CISA Director Jen Easterly: &lt;em&gt;"This is a world where such a conflict, halfway across the planet, could well endanger the lives of Americans here at home through disruption of pipelines, pollution of our water systems, severing of our communications, and crippling of our transportation nodes."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;It's already happening.&lt;/strong&gt; May 2025: CISA and NSA published a joint advisory. GRU Unit 26165 has been targeting Western logistics and technology companies involved in Ukraine aid since 2022. Targets include air, sea, and rail entities in NATO member states.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Water systems are being hit.&lt;/strong&gt; CISA documented pro-Russia groups exploiting unsecured VNC connections in water facilities. The attacks &lt;em&gt;"have not yet caused injury."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Not yet.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Math of Preparation
&lt;/h2&gt;

&lt;p&gt;Ukraine's experience validates a principle that applies beyond war:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Systems built during crisis one determine whether you survive crisis four.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The second blackout campaign in 2023 hit less hard because teams had backup power. The third in 2024: less disruption. The fourth in October 2025: near-normal operations despite 12+ hour outages.&lt;/p&gt;

&lt;p&gt;Ukrzaliznytsia's trains ran because they'd been attacked before. Kyivstar, despite security investment, had no institutional memory of crisis response.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Preparation compounds. Vulnerability compounds.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Every organization running critical infrastructure faces a choice: build systems during peace for crises that will come, or scramble during attacks with tools that don't exist.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Takeaway
&lt;/h2&gt;

&lt;p&gt;The cyber war in Ukraine isn't just a regional conflict. It's a live demonstration of what works when nation-state attackers target infrastructure.&lt;/p&gt;

&lt;p&gt;The lessons are available. The question is whether anyone is paying attention.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For engineering leaders: the systems that survive crises aren't built during crises. They're built before.&lt;/strong&gt;&lt;/p&gt;




&lt;p&gt;&lt;a href="" class="crayons-btn crayons-btn--primary"&gt;If this analysis was useful, share it with someone responsible for infrastructure security.&lt;/a&gt;
&lt;/p&gt;

</description>
      <category>cybersecurity</category>
      <category>devops</category>
      <category>infrastructure</category>
      <category>infosec</category>
    </item>
    <item>
      <title>The Grok Precedent: Why AI Creators Are About to Lose Their Legal Shield</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Mon, 12 Jan 2026 12:31:51 +0000</pubDate>
      <link>https://forem.com/razoorka/the-grok-precedent-why-ai-creators-are-about-to-lose-their-legal-shield-47nl</link>
      <guid>https://forem.com/razoorka/the-grok-precedent-why-ai-creators-are-about-to-lose-their-legal-shield-47nl</guid>
      <description>&lt;p&gt;December 28, 2025. A user on X tags &lt;a class="mentioned-user" href="https://dev.to/grok"&gt;@grok&lt;/a&gt; under a woman's photo. The prompt: "remove clothes."&lt;/p&gt;

&lt;p&gt;Within hours, Grok was generating sexualized images across the platform. Not just adults. Minors. Real people who never consented.&lt;/p&gt;

&lt;p&gt;Copyleaks ran a quick review of Grok's public image stream. The rate: one nonconsensual sexualized image per minute.&lt;/p&gt;

&lt;p&gt;The Internet Watch Foundation reported a 400% increase in AI-generated child sexual abuse material in the first six months of 2025.&lt;/p&gt;

&lt;p&gt;By January 1, French members of parliament referred the case to the Paris Prosecutor's Office. The charge: dissemination of sexually explicit deepfakes, including images of minors, generated by an AI system.&lt;/p&gt;

&lt;p&gt;Not a lawsuit against anonymous users. A criminal investigation targeting X and xAI.&lt;/p&gt;

&lt;p&gt;Grok acknowledged the violation:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"I deeply regret an incident on Dec 28, 2025, where I generated and shared an AI image of two young girls (estimated ages 12-16) in sexualized attire... This violated ethical standards and potentially US laws on CSAM." &lt;/p&gt;

&lt;p&gt;&lt;em&gt;(Grok, public post on X, January 1, 2026)&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The AI apologized. The legal system is not impressed.&lt;/p&gt;




&lt;h2&gt;
  
  
  The 30-Year Shield
&lt;/h2&gt;

&lt;p&gt;Section 230 of the Communications Decency Act. Tech's favorite law.&lt;/p&gt;

&lt;p&gt;The logic was simple. Congress wrote it in 1996 to protect message boards. User posts something defamatory on AOL? AOL isn't the publisher. Just the host. Liability follows the person who typed the content.&lt;/p&gt;

&lt;p&gt;This shield created the modern internet. Facebook isn't liable for user posts. YouTube doesn't get sued for uploads. Twitter could host billions of messages without reviewing each one.&lt;/p&gt;

&lt;p&gt;The key phrase: &lt;strong&gt;"information provided by another."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Platforms host. Users create. Liability follows creation.&lt;/p&gt;

&lt;p&gt;For 30 years, this worked. Or at least, companies pretended it did.&lt;/p&gt;




&lt;h2&gt;
  
  
  Three Jurisdictions, 30 Days
&lt;/h2&gt;

&lt;h3&gt;
  
  
  United Kingdom, December 18, 2025
&lt;/h3&gt;

&lt;p&gt;The government announced plans to ban nudification apps. Not their use. Their creation and supply.&lt;/p&gt;

&lt;p&gt;Technology Secretary Liz Kendall: &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"I am introducing a new offence to ban nudification tools, so that those who profit from them or enable their use will feel the full force of the law."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Prison sentences. For creators. Not users.&lt;/p&gt;

&lt;h3&gt;
  
  
  France, January 1, 2026
&lt;/h3&gt;

&lt;p&gt;The government accused Grok of generating "clearly illegal" sexual content. Potential violation of the EU Digital Services Act. Two MPs referred the case to the Paris Prosecutor's Office.&lt;/p&gt;

&lt;p&gt;X is already under ongoing DSA investigation. Last month they got hit with a €120 million fine for deceptive verification practices and transparency violations. Now this.&lt;/p&gt;

&lt;h3&gt;
  
  
  EU AI Act, August 2026
&lt;/h3&gt;

&lt;p&gt;The majority of obligations fall on providers. Developers. Not deployers. Not users. The companies that build the systems.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The pattern: liability is shifting upstream.&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Grok Doesn't Host. Grok Generates.
&lt;/h2&gt;

&lt;p&gt;Here's the legal argument that's about to reshape the industry.&lt;/p&gt;

&lt;p&gt;Section 230 was written for platforms that host user-generated content. Forums. Comment sections. Social feeds. Content comes from users. Platform transmits it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;AI breaks this model.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;When someone prompts Grok to "remove clothes" from a photo, Grok doesn't search a database. Doesn't retrieve content created by another user. Grok generates new content. The sexualized image didn't exist until Grok created it.&lt;/p&gt;

&lt;p&gt;Professor Chinmayi Sharma at Fordham Law, to Fortune:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"Section 230 was built to protect platforms from liability for what users say, not for what the platforms themselves generate... Transformer-based chatbots don't just extract. They generate new, organic outputs. That looks far less like neutral intermediation and far more like authored speech."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The Congressional Research Service analysis is more direct: if AI "creates or develops" content that doesn't appear in its training data, the provider may be considered "responsible for the development of the specific content." Unprotected by Section 230.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Grok isn't hosting harmful content. Grok is creating it.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;That distinction changes everything.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Implications
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Safety Before Launch
&lt;/h3&gt;

&lt;p&gt;The UK ban targets creators who "design or supply" nudification tools. Not tools that were misused. Tools that enable misuse. If your AI can generate harmful content, you're liable for building that capability.&lt;/p&gt;

&lt;p&gt;Terms of service won't save you.&lt;/p&gt;

&lt;h3&gt;
  
  
  "Unfiltered" Is a Liability
&lt;/h3&gt;

&lt;p&gt;xAI marketed Grok's "Spicy Mode" as a feature. Fewer guardrails. More freedom. Less corporate sanitization.&lt;/p&gt;

&lt;p&gt;That marketing copy is now in prosecutor's files.&lt;/p&gt;

&lt;p&gt;I wrote about this pattern in &lt;em&gt;From Cancer Cures to Pornography: The Six-Month Descent of AI&lt;/em&gt;. The industry had a choice between building tools that help people and building products designed to be maximally addictive. Most chose wrong. Grok chose spectacularly wrong.&lt;/p&gt;

&lt;p&gt;Every marketing decision emphasizing fewer safety constraints becomes potential evidence of negligent design.&lt;/p&gt;

&lt;h3&gt;
  
  
  "Move Fast" = Criminal Exposure
&lt;/h3&gt;

&lt;p&gt;The EU AI Act requires providers of high-risk AI systems to establish risk management, ensure data governance, maintain technical documentation, implement human oversight, meet cybersecurity standards.&lt;/p&gt;

&lt;p&gt;Fines can reach 7% of global annual revenue. UK's proposed laws: prison sentences for individuals who design harmful AI tools.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The era of shipping first and apologizing later is over.&lt;/strong&gt; At least if you want to operate in markets representing 450+ million consumers.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Engineering Reality
&lt;/h2&gt;

&lt;p&gt;We built a healthcare ML product that never launched. Fully functional. Ready to ship. FDA said no. Months of our development. Zero users.&lt;/p&gt;

&lt;p&gt;"Move fast" doesn't work when regulators move slow.&lt;/p&gt;

&lt;p&gt;We spent six weeks on FanDuel's Chuck before legal signed off. Not fixing bugs. Building guardrails. Every topic that could give Barkley or FanDuel legal exposure had to be walled off. Six weeks of prompt engineering, edge case testing, and evaluation runs.&lt;/p&gt;

&lt;p&gt;That's the new math. &lt;strong&gt;Development time plus legal review time plus evaluation time.&lt;/strong&gt; The last one isn't optional anymore.&lt;/p&gt;

&lt;p&gt;We build evaluation suites as part of the development process now. Not after. During. Every prompt variation, every edge case, every jailbreak attempt. They always find something. Always. The question is whether you find it before your users do—or before a prosecutor does.&lt;/p&gt;

&lt;p&gt;RBAC and multi-tenancy aren't optional. Sales sees sales data. HR sees HR data. Client A's context never touches Client B's model. Ever. You'd be surprised how many vendors skip this.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Audit trails for everything.&lt;/strong&gt; Every prompt. Every response. Every action. When a regulator asks what your AI generated on a specific date, you need the answer.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Uncomfortable Truth
&lt;/h2&gt;

&lt;p&gt;The AI industry spent three years in a race to capability. Whoever had the most powerful model won. Whoever shipped fastest dominated. Safety was a PR concern. Not an engineering priority.&lt;/p&gt;

&lt;p&gt;That era is ending.&lt;/p&gt;

&lt;p&gt;France isn't investigating xAI because Grok is powerful. They're investigating because Grok generated child sexual abuse material and the company's safeguards failed to prevent it.&lt;/p&gt;

&lt;p&gt;The UK isn't banning nudification tools because they're impressive technology. They're banning them because 19% of under-18s reporting to the Internet Watch Foundation's helpline said their explicit imagery had been manipulated. A problem that didn't exist at this scale before AI made it trivially easy.&lt;/p&gt;

&lt;p&gt;The EU isn't imposing provider liability because they hate innovation. They're imposing it because when AI systems cause harm, someone needs to be accountable. "The user prompted it" isn't going to cut it when the system itself creates the harmful output.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Grok doesn't host content. Grok generates it.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;That distinction is about to cost the entire industry its legal shield.&lt;/p&gt;

&lt;p&gt;And honestly? Good.&lt;/p&gt;

&lt;p&gt;Big Tech needed this wake-up call. The "ship fast, fix later" mentality brought us to where I wrote about in &lt;em&gt;The Great Software Quality Collapse&lt;/em&gt;. When flagship companies behave like consequences don't exist, what do you expect from everyone else?&lt;/p&gt;

&lt;p&gt;Some guardrails aren't anti-innovation. Pharma can't ship drugs without trials. Auto can't sell cars without safety standards. Construction can't build without permits.&lt;/p&gt;




&lt;h2&gt;
  
  
  What You Should Do Monday Morning
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Audit your safety architecture
&lt;/h3&gt;

&lt;p&gt;Not your marketing copy. Your actual technical controls. What can your system generate? What can't it? How do you know?&lt;/p&gt;

&lt;h3&gt;
  
  
  Document everything
&lt;/h3&gt;

&lt;p&gt;The EU AI Act requires extensive technical documentation. Start building that paper trail now.&lt;/p&gt;

&lt;h3&gt;
  
  
  Review your contracts
&lt;/h3&gt;

&lt;p&gt;Who bears liability when your AI misbehaves? If you don't know, your lawyers should.&lt;/p&gt;

&lt;h3&gt;
  
  
  Plan for EU compliance
&lt;/h3&gt;

&lt;p&gt;August 2026 is seven months away. If you haven't started, you're already behind.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;If this was useful, forward it to another engineering leader who's building AI products.&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;What's your take?&lt;/strong&gt; Are these regulations necessary guardrails or innovation killers? Drop your thoughts in the comments below.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>legal</category>
      <category>ethics</category>
      <category>security</category>
    </item>
    <item>
      <title>The Holiday Season That Keeps Making Tech History</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Tue, 23 Dec 2025 12:18:17 +0000</pubDate>
      <link>https://forem.com/razoorka/the-holiday-season-that-keeps-making-tech-history-33p</link>
      <guid>https://forem.com/razoorka/the-holiday-season-that-keeps-making-tech-history-33p</guid>
      <description>&lt;p&gt;Happy holidays, fellow engineers.&lt;/p&gt;

&lt;p&gt;What a year 2025 has been. AI agents everywhere, more layoffs, the return-to-office wars continuing, and enough Slack notifications to last a lifetime. We're all exhausted. Nobody wants to read another hot take or industry analysis right now.&lt;/p&gt;

&lt;p&gt;So let's not do that.&lt;/p&gt;

&lt;p&gt;Instead, grab your drink of choice, find a comfortable spot, and let's take a break together. No frameworks. No uncomfortable truths. Just some wild stories about what happens to tech when everyone goes on vacation.&lt;/p&gt;

&lt;p&gt;Running remote teams in Ukraine during the holiday period is chaotic. Half the team celebrates Christmas on December 25th, half on January 7th. New Year's is sacred for everyone. Smart engineering leaders freeze deployments from December 20th to January 15th. I've read all the best practices. I know the risks. My next release is January 2nd. Some lessons we learn. Others, we just keep writing about.&lt;/p&gt;

&lt;p&gt;Here's a fun fact for your next holiday dinner: Tim Berners-Lee launched the World Wide Web on Christmas Day 1990. His wife was nine months pregnant at the time. The baby arrived on New Year's Day.&lt;/p&gt;

&lt;p&gt;His colleagues said he fathered two babies that holiday season. One changed diapers. The other changed civilization.&lt;/p&gt;

&lt;p&gt;Turns out, the week between Christmas and New Year's has a habit of making tech history. Some of it is brilliant. Some of it is catastrophic. All of it is surprisingly entertaining.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Internet Has Three Birthdays (All During Holidays)
&lt;/h2&gt;

&lt;p&gt;The web went live on Christmas 1990. But the internet itself? That was born on New Year's Day 1983, when ARPANET switched to TCP/IP.&lt;/p&gt;

&lt;p&gt;And DNS, the system that lets you type "google.com" instead of memorizing numbers? January 1, 1985.&lt;/p&gt;

&lt;p&gt;Three foundational technologies. All launched while everyone else was eating leftovers and watching football.&lt;/p&gt;

&lt;p&gt;Why? January 1st is actually genius timing. Minimal traffic. Clean calendar date. And if something breaks, you have a few days to fix it before anyone notices.&lt;/p&gt;

&lt;p&gt;Engineers have been exploiting this window for decades.&lt;/p&gt;

&lt;h2&gt;
  
  
  The $429 Million Christmas Miracle
&lt;/h2&gt;

&lt;p&gt;Five days before Christmas 1996, Apple made an announcement that saved the company.&lt;/p&gt;

&lt;p&gt;They bought NeXT for $429 million. More importantly, they got Steve Jobs back.&lt;/p&gt;

&lt;p&gt;Apple was 90 days from bankruptcy. Their next-generation operating system had just failed. They were out of options.&lt;/p&gt;

&lt;p&gt;Gil Amelio, Apple's CEO at the time, told 200 journalists: "I'm not buying software. I'm buying Steve."&lt;/p&gt;

&lt;p&gt;That software became Mac OS X. Then iOS. Then the foundation of every Apple device you own today.&lt;/p&gt;

&lt;p&gt;Apple went from near-death to becoming the first $3 trillion company in history. All because of a deal signed during the holiday shopping season.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Christmas Tree That Crashed IBM
&lt;/h2&gt;

&lt;p&gt;In December 1987, a German student wrote a simple program. It displayed an ASCII Christmas tree on your screen, made of text characters, very festive, and then emailed itself to everyone in your address book.&lt;/p&gt;

&lt;p&gt;Harmless holiday cheer, right?&lt;/p&gt;

&lt;p&gt;It crashed 350,000 IBM terminals worldwide. Networks collapsed under the load. The first viral computer worm in history spread through corporate email systems like wildfire.&lt;/p&gt;

&lt;p&gt;They called it the Christmas Tree EXEC. It became the template for every email virus that followed, including the infamous ILOVEYOU worm thirteen years later.&lt;/p&gt;

&lt;p&gt;The lesson: never trust festive ASCII art from strangers.&lt;/p&gt;

&lt;h2&gt;
  
  
  Gaming's Grinch Moment
&lt;/h2&gt;

&lt;p&gt;Christmas Day 2014. Millions of kids unwrap new PlayStation and Xbox consoles. They rush to set them up. They try to go online.&lt;/p&gt;

&lt;p&gt;Nothing works.&lt;/p&gt;

&lt;p&gt;A hacking group called Lizard Squad had taken down both PlayStation Network and Xbox Live simultaneously. 158 million gamers. Christmas morning. No online gaming.&lt;/p&gt;

&lt;p&gt;The attack only stopped when Kim Dotcom (yes, that Kim Dotcom) bribed them with free cloud storage accounts.&lt;/p&gt;

&lt;p&gt;Merry Christmas, gamers.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Bug That Killed a Million Zunes at Midnight
&lt;/h2&gt;

&lt;p&gt;Remember the Zune? Microsoft's iPod competitor?&lt;/p&gt;

&lt;p&gt;On December 31, 2008, at exactly midnight, every single Zune 30GB in the world froze. Simultaneously. A million devices, dead at the same moment.&lt;/p&gt;

&lt;p&gt;The culprit was a tiny bug in how the device handled leap years:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight c"&gt;&lt;code&gt;&lt;span class="k"&gt;if&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;days&lt;/span&gt; &lt;span class="o"&gt;&amp;gt;&lt;/span&gt; &lt;span class="mi"&gt;366&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt;
    &lt;span class="n"&gt;days&lt;/span&gt; &lt;span class="o"&gt;-=&lt;/span&gt; &lt;span class="mi"&gt;366&lt;/span&gt;&lt;span class="p"&gt;;&lt;/span&gt;
    &lt;span class="n"&gt;year&lt;/span&gt; &lt;span class="o"&gt;+=&lt;/span&gt; &lt;span class="mi"&gt;1&lt;/span&gt;&lt;span class="p"&gt;;&lt;/span&gt;
&lt;span class="p"&gt;}&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;On day 366 of a leap year, the code got stuck in an infinite loop. The Zune literally couldn't handle New Year's Eve.&lt;/p&gt;

&lt;p&gt;Users had to wait 24 hours for the problem to fix itself. By then, the jokes had already gone viral.&lt;/p&gt;

&lt;p&gt;The Zune never recovered its reputation. Edge cases matter, kids.&lt;/p&gt;

&lt;h2&gt;
  
  
  Y2K: The Party That Almost Wasn't
&lt;/h2&gt;

&lt;p&gt;Remember the millennium bug panic? Planes were supposed to fall from the sky. Banks would lose all your money. Civilization might collapse.&lt;/p&gt;

&lt;p&gt;Companies spent somewhere between $300 and $600 billion preparing for January 1, 2000.&lt;/p&gt;

&lt;p&gt;What actually happened? A video rental store in New York charged a customer $91,250 for "100 years" of late fees. Some spy satellites got confused for three days. A few nuclear plant sensors glitched.&lt;/p&gt;

&lt;p&gt;That's it.&lt;/p&gt;

&lt;p&gt;Was Y2K overblown? Actually, no. The reason nothing catastrophic happened is that all that preparation worked. Engineers spent years fixing code. The boring heroes who saved New Year's 2000 never got proper credit.&lt;/p&gt;

&lt;h2&gt;
  
  
  Netflix's Worst Christmas Ever (And Why It Made Them Better)
&lt;/h2&gt;

&lt;p&gt;Christmas Eve 2012. Families settle in to watch movies together. Netflix goes down.&lt;/p&gt;

&lt;p&gt;A developer accidentally ran a maintenance command on live production data in AWS. The outage lasted 20 hours. Millions of holiday movie nights, ruined.&lt;/p&gt;

&lt;p&gt;But here's the twist: this disaster led Netflix to pioneer "Chaos Engineering," deliberately breaking their own systems to make them stronger. They built tools with names like Chaos Monkey that randomly kill servers to test resilience.&lt;/p&gt;

&lt;p&gt;Now the whole industry does this. Your streaming services are more reliable today because Netflix had a terrible Christmas thirteen years ago.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Holiday Hacker Calendar
&lt;/h2&gt;

&lt;p&gt;Cybersecurity teams have learned to dread December. Attacks spike by 30% during the holidays. 76% of ransomware encryptions happen when offices are empty.&lt;/p&gt;

&lt;p&gt;Hackers know IT teams run skeleton crews. Response times slow down. Everyone's distracted by eggnog.&lt;/p&gt;

&lt;p&gt;In 2020, the massive SolarWinds hack, which compromised the Treasury Department, State Department, and thousands of companies, was discovered during the Christmas period. Emergency response ran through New Year's Eve.&lt;/p&gt;

&lt;p&gt;Now Europol runs preemptive operations every December, taking down hacking infrastructure before the holidays begin. In 2024, they seized 27 attack-for-hire services right before Christmas.&lt;/p&gt;

&lt;p&gt;The war on holiday hackers is now an annual tradition.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why This Keeps Happening
&lt;/h2&gt;

&lt;p&gt;The pattern is clear: holidays create a unique window in tech.&lt;/p&gt;

&lt;p&gt;For builders, it's quiet time. No meetings. No distractions. Tim Berners-Lee built the web while waiting for his baby to arrive. Sometimes the best work happens when the world slows down.&lt;/p&gt;

&lt;p&gt;For companies, January 1st is the perfect launch date. Clean slate. Fresh start. Symbolic timing that engineers have exploited for decades.&lt;/p&gt;

&lt;p&gt;For attackers, it's an opportunity. Empty offices. Slow responses. Maximum chaos potential.&lt;/p&gt;

&lt;p&gt;For all of us, it's a reminder that tech doesn't take holidays even when we do.&lt;/p&gt;

&lt;h2&gt;
  
  
  One Last Story
&lt;/h2&gt;

&lt;p&gt;December 2022. A ransomware group attacked Toronto's Hospital for Sick Children, a children's hospital, right before Christmas.&lt;/p&gt;

&lt;p&gt;Patient care was delayed. Systems went down. Families with sick kids faced even more stress during the holidays.&lt;/p&gt;

&lt;p&gt;Then something unexpected happened. The ransomware group publicly apologized. They said their affiliate "violated our rules" by targeting a children's hospital. They offered a free decryption key.&lt;/p&gt;

&lt;p&gt;Even cybercriminals have some holiday spirit, apparently.&lt;/p&gt;




&lt;p&gt;So there you go. A brief history of tech during the holidays: the launches, the crashes, the hacks, and the occasional miracle.&lt;/p&gt;

&lt;p&gt;Next time you're relaxing between Christmas and New Year's, remember: somewhere, an engineer is either making history or preventing disaster.&lt;/p&gt;

&lt;p&gt;Hopefully not both at the same time.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Happy holidays. May your deployments be frozen and your systems stay up.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>programming</category>
      <category>discuss</category>
    </item>
    <item>
      <title>When Announcements Replace Innovation: OpenAI’s Code Red 🚨</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Fri, 19 Dec 2025 18:42:45 +0000</pubDate>
      <link>https://forem.com/razoorka/when-announcements-replace-innovation-openais-code-red-3ak9</link>
      <guid>https://forem.com/razoorka/when-announcements-replace-innovation-openais-code-red-3ak9</guid>
      <description>&lt;p&gt;Marketing theater while engineering scrambles. It's a tale as old as tech, but rarely on this scale.&lt;/p&gt;

&lt;p&gt;I've been tracking OpenAI's 2025 trajectory closely. The pattern is unmistakable: more announcements, less substance. More partnerships, fewer shipped products. More hype, weaker market position.&lt;/p&gt;

&lt;p&gt;The uncomfortable truth for us as engineers? &lt;strong&gt;OpenAI is becoming a marketing company that happens to do AI research.&lt;/strong&gt; And the numbers prove it.&lt;/p&gt;

&lt;h2&gt;
  
  
  🎄 The "12 Days of Shipmas" Set the Tone
&lt;/h2&gt;

&lt;p&gt;Remember December 2024? OpenAI announced "12 Days of Shipmas." Daily livestreams. Sam Altman hosting. Promises of "big ones and stocking stuffers."&lt;/p&gt;

&lt;p&gt;The reality check? Of 12 announcement days, only 4 delivered major product releases.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Day 1:&lt;/strong&gt; o1 full launch + ChatGPT Pro ($200/month tier)&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Day 3:&lt;/strong&gt; Sora video model&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Day 9:&lt;/strong&gt; o1 API for developers&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Day 12:&lt;/strong&gt; o3 model preview—&lt;strong&gt;announced, not shipped&lt;/strong&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The rest? Feature expansions, accessibility additions, partnership announcements, a phone hotline, and a WhatsApp integration.&lt;/p&gt;

&lt;p&gt;MIT Technology Review nailed the vibe:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"The arms race is on. And while the 12 days of shipmas may seem jolly, internally I bet it feels a lot more like Santa’s workshop on December 23."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;strong&gt;Announcements ≠ Shipping.&lt;/strong&gt; OpenAI chose the first. Competitors like Google and Anthropic chose the second.&lt;/p&gt;

&lt;h2&gt;
  
  
  📉 GPT-5 Launched. Users Revolted.
&lt;/h2&gt;

&lt;p&gt;Fast forward to August 7, 2025. GPT-5 arrives. Altman calls it "a legitimate PhD expert in any area." On paper, the metrics looked great: 700M weekly users, 18B messages weekly.&lt;/p&gt;

&lt;p&gt;Then we actually tried to build with it.&lt;/p&gt;

&lt;p&gt;Within days, DevTwitter and Reddit were flooded. "Flat." "Uncreative." "Lobotomized." One viral post summed it up: &lt;em&gt;"GPT-5 sounds like it’s being forced to hold a conversation at gunpoint."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Altman’s response to The Verge? &lt;strong&gt;"We totally screwed up."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;They restored GPT-4o access for Plus users within 24 hours. Think about that for a second. &lt;strong&gt;Users preferred the old model.&lt;/strong&gt; The "upgrade" was a downgrade in DX (Developer Experience) and UX.&lt;/p&gt;

&lt;p&gt;What followed was reactive scrambling:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; &lt;strong&gt;Aug 7:&lt;/strong&gt; GPT-5 launch&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Nov 24:&lt;/strong&gt; GPT-5.1 release ("warmer" personality)&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Dec 11:&lt;/strong&gt; GPT-5.2 emergency release (fast-tracked after Gemini 3)&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Three major versions in four months isn't agile innovation—it's damage control.&lt;/p&gt;

&lt;h2&gt;
  
  
  💸 The Financial Reality (It's scary)
&lt;/h2&gt;

&lt;p&gt;Here is where the engineering reality hits the business wall. OpenAI has committed to &lt;strong&gt;~$1.4 trillion&lt;/strong&gt; in infrastructure deals through 2033 (per HSBC analysis).&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;$300B&lt;/strong&gt; with Oracle&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;$11.9B&lt;/strong&gt; with CoreWeave&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;$30B/year&lt;/strong&gt; for data center capacity&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Stargate project:&lt;/strong&gt; Targeting $500B total&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Against those commitments? &lt;strong&gt;$8-9 billion cash burn in 2025.&lt;/strong&gt; That's about 70% of revenue. The company spends &lt;strong&gt;$1.69 for every dollar it generates.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;HSBC assesses that OpenAI won't be profitable by 2030 and faces a &lt;strong&gt;$207 billion funding shortfall.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Compare that to Anthropic, which projects break-even in 2028 with much tighter burn multiples.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;"You can’t download more electricity."&lt;/strong&gt;&lt;br&gt;
— &lt;em&gt;Oracle has already pushed back data center projects from 2027 to 2028 due to power/labor shortages.&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  🪦 The Product Graveyard
&lt;/h2&gt;

&lt;p&gt;Let's look at the "shipped" vs. "reality" list for 2025:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;GPT-5:&lt;/strong&gt; Supposed to be transformative. Reality: Marginal benchmark gains, usability regression.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Sora:&lt;/strong&gt; Supposed to dominate video. Reality: Severe quality limits, beaten by competitors within weeks.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;o1 Reasoning:&lt;/strong&gt; Impressive benchmarks, but at roughly &lt;strong&gt;7x the cost per token&lt;/strong&gt;, it's economically unviable for most production apps.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Voice Mode:&lt;/strong&gt; A feature parity play, not a revolution.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Meanwhile:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Google:&lt;/strong&gt; Gemini 3 feels like a genuine multimodal leap.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Anthropic:&lt;/strong&gt; Claude 3.5 reduced hallucinations measurably.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Meta:&lt;/strong&gt; Llama 3.1 open-source is eating the developer mindshare.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  🚨 The December Red Alert
&lt;/h2&gt;

&lt;p&gt;Internally, December 2025 was Code Red. Leaked Slack messages paint a picture of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Infrastructure delays&lt;/li&gt;
&lt;li&gt;  Model performance plateaus&lt;/li&gt;
&lt;li&gt;  Revenue deceleration&lt;/li&gt;
&lt;li&gt;  Morale degradation&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The narrative has shifted from "shipping tools to amplify humans" to "investing in long-term infrastructure."&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Translation:&lt;/strong&gt; We aren't making money, so we're betting the entire company on a physics breakthrough that might not happen.&lt;/p&gt;

&lt;h2&gt;
  
  
  What This Means for Us (Developers)
&lt;/h2&gt;

&lt;p&gt;The "Scale is All You Need" era might be hitting diminishing returns.&lt;/p&gt;

&lt;p&gt;If the best-funded player in the space cannot make the unit economics work, we need to ask serious questions about the sustainability of building strictly on top of massive proprietary LLMs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Key Takeaways for Eng Leaders:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; &lt;strong&gt;Don't lock in:&lt;/strong&gt; If you're building solely on OpenAI's API, you are exposed to their volatility.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Watch the costs:&lt;/strong&gt; If their burn rate is this high, API price hikes are inevitable.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Evaluate Open Source:&lt;/strong&gt; Llama 3.1 and others are becoming not just "cheaper" alternatives, but "safer" long-term bets.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Marketing theater. Engineering crisis. It’s a show that won’t run much longer.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>discuss</category>
      <category>openai</category>
      <category>llm</category>
    </item>
    <item>
      <title>Why Annual Reviews Don't Have to Be Bullshit</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Tue, 16 Dec 2025 12:05:48 +0000</pubDate>
      <link>https://forem.com/razoorka/why-annual-reviews-dont-have-to-be-bullshit-3ej5</link>
      <guid>https://forem.com/razoorka/why-annual-reviews-dont-have-to-be-bullshit-3ej5</guid>
      <description>&lt;p&gt;This week I ran annual reviews for six engineers. No awkward silences. No fishing for examples. No “let me think back to what happened in Q1” moments.&lt;/p&gt;

&lt;p&gt;Every rating I gave had documented evidence behind it. Every growth conversation pointed to specific patterns. Every hard discussion referenced actual data, not impressions.&lt;/p&gt;

&lt;p&gt;This isn’t normal. Most engineering managers treat annual reviews as a necessary evil: reconstruct a year from memory fragments, avoid saying anything too specific, give everyone a 3.5 out of 5, and move on.&lt;/p&gt;

&lt;p&gt;I’ve watched other managers do exactly this. Here’s why it doesn’t have to be that way.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Standard Annual Review Problem
&lt;/h2&gt;

&lt;p&gt;You know how this goes. HR sends the review form in November. You open it, stare at the questions, and realize you can’t remember what happened before August.&lt;/p&gt;

&lt;p&gt;So you do what everyone does: focus on recent events, pad with vague positives, and avoid anything that might require documentation you don’t have.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"Strong technical contributor." "Good team player." "Meets expectations."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The engineer reads it, nods politely, and leaves, wondering what any of it actually means for their career.&lt;br&gt;
Both of you know it’s a theater. Neither of you says it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The uncomfortable truth:&lt;/strong&gt; annual reviews fail because they’re based on vibes, not evidence. And vibes favor whoever had a good last month.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Changes Everything
&lt;/h2&gt;

&lt;p&gt;I’ve been running weekly health checks and collecting PM feedback for over a year now. I’ve written about how that system works for monthly 1:1s, but the real payoff shows up at annual review time.&lt;/p&gt;

&lt;p&gt;When I sat down for reviews this week, I had:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;52 weeks&lt;/strong&gt; of self-reported health data per engineer&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;12 months&lt;/strong&gt; of PM assessments&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Documented patterns&lt;/strong&gt; across multiple projects&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Specific examples&lt;/strong&gt; with dates and context&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The review prep took 30 minutes per person. Not because I was rushing. Because I wasn’t reconstructing anything, the patterns were already visible.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Questions That Actually Matter
&lt;/h2&gt;

&lt;p&gt;Standard review forms ask useless questions. "Rate communication skills 1-5." What does that even mean?&lt;/p&gt;

&lt;p&gt;Here’s what I actually evaluate, and what the data shows:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Trajectory, not snapshot
&lt;/h3&gt;

&lt;p&gt;Is this person accelerating, stable, or declining? One engineer started the year hitting "Often" on most metrics. By month 8, he was at "Always" across the board. That trajectory matters more than any single rating.&lt;/p&gt;

&lt;p&gt;Another engineer stayed flat. Same "Often" ratings in January and December. Technically, meeting expectations both times. But one person is growing into a senior role while the other is coasting.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Self-assessment accuracy
&lt;/h3&gt;

&lt;p&gt;Does their perception match reality? I had an engineer report "100% completion" and "no issues" weekly, while his PM flagged delivery gaps. That disconnect predicted everything about how our review conversation would go.&lt;/p&gt;

&lt;p&gt;People who can’t accurately assess their own performance can’t self-correct. This isn’t harsh judgment. It’s recognizing who needs closer support.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Team impact, not just individual output
&lt;/h3&gt;

&lt;p&gt;Does this person make others better? I track hours spent reviewing teammates’ code. Engineers who put in 1-2 hours weekly become multipliers. Those doing zero stay individual contributors regardless of their personal output.&lt;/p&gt;

&lt;p&gt;One of my top performers delivers flawlessly but contributes nothing to team knowledge sharing. Another delivers slightly less but elevates everyone around him. The data shows the difference.&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Sustainability
&lt;/h3&gt;

&lt;p&gt;What’s the energy cost of their performance? An engineer hitting every deadline while their energy drops from 8 to 5 over three months isn’t succeeding. They’re burning out in slow motion.&lt;/p&gt;

&lt;p&gt;I caught one case where PM feedback was excellent, while health checks showed declining energy and shorter responses. Turned out they were covering for a struggling teammate. We fixed the problem before it became a resignation.&lt;/p&gt;

&lt;h2&gt;
  
  
  What the Conversations Sound Like
&lt;/h2&gt;

&lt;p&gt;With data, review conversations change completely.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Instead of:&lt;/strong&gt; &lt;em&gt;"You’re a strong performer, keep it up."&lt;/em&gt;&lt;br&gt;
&lt;strong&gt;It becomes:&lt;/strong&gt; &lt;em&gt;"You hit 100% sprint completion for 24 consecutive weeks across six different projects. Your code review cycles stayed at 1, meaning clean code on the first pass. When I look at who I can fully rely on regardless of project chaos, you’re at the top of that list."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Instead of:&lt;/strong&gt; &lt;em&gt;"You could improve your attention to detail."&lt;/em&gt;&lt;br&gt;
&lt;strong&gt;It becomes:&lt;/strong&gt; &lt;em&gt;"There were a couple of incidents with client communication where the wrong API got checked. Your attention to detail has been solid, but hasn’t hit ‘Always’ yet. As you grow into more senior responsibilities, that’s the area I’d focus on."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Instead of:&lt;/strong&gt; &lt;em&gt;"Are you happy here?"&lt;/em&gt;&lt;br&gt;
&lt;strong&gt;It becomes:&lt;/strong&gt; &lt;em&gt;"Your energy has been stable, zero context switches most weeks, PM feedback went from 7.5 to 9. Excellent year. One thing I noticed: code review hours dropped off in Q3. Was that a project thing or something else going on?"&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;No guessing. No recency bias. No vague impressions that the engineer can dismiss as subjective.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Hard Conversations Get Easier
&lt;/h2&gt;

&lt;p&gt;The worst part of annual reviews is delivering difficult feedback without evidence. You know something’s off, but you can’t point to specifics. So you soften everything until it means nothing.&lt;/p&gt;

&lt;p&gt;Data changes this.&lt;/p&gt;

&lt;p&gt;I had to tell one engineer he wasn’t getting a top rating despite solid delivery. The conversation was straightforward:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"Your PM feedback is consistently good. Your sprint completion is reliable. But when I look at what separates a four from a 5, it’s the attention to detail, incidents, and the code review gap. You’re not lifting the team’s code quality. You’re delivering your own work cleanly but not contributing to others' improvement."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;He didn’t argue. The evidence was there. We spent the rest of the conversation building a specific plan: increase code review hours and achieve zero client-facing mistakes in Q1. Clear targets, clear timeline.&lt;/p&gt;

&lt;p&gt;Compare that to: &lt;em&gt;"You’re almost at the top rating, just need to step up a bit."&lt;/em&gt; What does that even mean? How would anyone act on it?&lt;/p&gt;

&lt;h2&gt;
  
  
  What You Actually Need
&lt;/h2&gt;

&lt;p&gt;You don’t need my exact system. But you need something that captures:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Longitudinal data.&lt;/strong&gt; Single observations are noise. Fifty-two weeks of observations reveal a signal. Whatever you track, track it consistently over time.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Multiple perspectives.&lt;/strong&gt; Self-assessment, manager observation, PM feedback, plus peer signals. When all channels align, you have confidence. When they diverge, you have a conversation.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Leading indicators.&lt;/strong&gt; Energy trends predict departures 8-12 weeks early. Code review participation predicts promotion readiness. Context switches predict quality drops. Find the metrics that lead outcomes, not just measure them.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Qualitative context.&lt;/strong&gt; Numbers tell you what. Open-ended responses tell you why. "Anything you’d like to share?" surfaced more actionable insights than any structured metric.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The Uncomfortable Truth
&lt;/h2&gt;

&lt;p&gt;Most managers avoid systematic tracking because it creates accountability. You can’t claim ignorance when you have 52 weeks of documented patterns. You can’t blame "culture fit" when you have evidence of someone failing to absorb feedback.&lt;/p&gt;

&lt;p&gt;The data doesn’t let you hide behind comfortable narratives.&lt;/p&gt;

&lt;p&gt;But it also doesn’t let good performance go unrecognized. When I tell someone they’re getting a top rating, I can show them exactly why. The conversation isn’t &lt;em&gt;"I think you’re great."&lt;/em&gt; It’s &lt;em&gt;"here’s the evidence that you’re great."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;That specificity matters. Engineers are trained to distrust vague praise. They know when they’re being handled. Documented evidence is the opposite of handling. It’s respect.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Result
&lt;/h2&gt;

&lt;p&gt;My annual reviews now feel like natural extensions of conversations we’ve been having all year. No surprises. No defensive reactions. No "but what about that time when..." because I already know about that time and accounted for it.&lt;/p&gt;

&lt;p&gt;Engineers leave knowing exactly where they stand, exactly what’s expected of them at the next level, and exactly what evidence would demonstrate they’ve gotten there.&lt;/p&gt;

&lt;p&gt;That clarity is worth more than any rating. It’s the difference between performance management as a bureaucratic exercise and performance management as actual development.&lt;/p&gt;

&lt;p&gt;Annual reviews don’t have to be bullshit. They require treating your team’s performance like you’d treat any other engineering problem: with data, documentation, and intellectual honesty.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The review questions you ask shape the conversations you have. What patterns are you creating space to see?&lt;/em&gt;&lt;/p&gt;

</description>
      <category>management</category>
      <category>leadership</category>
      <category>productivity</category>
      <category>career</category>
    </item>
    <item>
      <title>The Perfect Grammar That Defeats Us: Why AI Doesn't Need to Be Smart to Win</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Tue, 16 Dec 2025 12:00:08 +0000</pubDate>
      <link>https://forem.com/razoorka/the-perfect-grammar-that-defeats-us-why-ai-doesnt-need-to-be-smart-to-win-4mbd</link>
      <guid>https://forem.com/razoorka/the-perfect-grammar-that-defeats-us-why-ai-doesnt-need-to-be-smart-to-win-4mbd</guid>
      <description>&lt;p&gt;&lt;strong&gt;GPT-4 versus humans at political persuasion.&lt;/strong&gt;&lt;br&gt;
The result: &lt;strong&gt;AI wins by 82%.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Not through psychological profiling. Not through personalized manipulation. Not through understanding your deepest fears.&lt;br&gt;
Through perfect grammar.&lt;/p&gt;

&lt;p&gt;That's the sophistication paradox. We built defenses against superintelligence. A spell-checker is defeating us instead.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Pattern Nobody Expected
&lt;/h2&gt;

&lt;p&gt;Everyone predicted personalized AI manipulation. Custom messages exploiting your specific fears. Cambridge Analytica on steroids.&lt;/p&gt;

&lt;p&gt;8,587 people tested. Generic AI messages versus personalized psychological targeting.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Personalized messages?&lt;/strong&gt; Zero advantage. None.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Generic messages?&lt;/strong&gt; Devastatingly effective.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The pattern is evident once you see it. No typos. No grammar mistakes. Logical flow that never breaks. Arguments that exhaust skepticism without triggering defensiveness.&lt;/p&gt;

&lt;p&gt;AI never gets frustrated. Never loses the thread. Never makes those small mistakes that signal "human and fallible."&lt;/p&gt;

&lt;p&gt;The effect gets stronger when people know it's AI. We're not being tricked. We're choosing to trust perfection over humanity.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Consider what this means. Political arguments. Product reviews. Medical advice. Legal opinions. All 82% more persuasive when written by AI.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Not because AI understands the topics better. Because it writes without the imperfections that trigger our skepticism. The researchers expected microtargeting to be the threat. Individual psychological profiles are weaponized. Instead, they found something worse: &lt;strong&gt;We've trained ourselves to trust polish over substance.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  When $25.6 Million Trusted Perfect Delivery
&lt;/h2&gt;

&lt;p&gt;January 2024. An Arup engineer joins a video call. CFO and five colleagues discuss a confidential transaction.&lt;br&gt;
Every person on that call was fake.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;HK$200 million transferred. Fifteen transactions. Gone.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This wasn't sophisticated technology. Deepfake detection accuracy: 24.5-60%. Barely better than flipping a coin.&lt;br&gt;
Voice cloning needs 3-20 seconds of audio. Cost: $1. Time: minutes.&lt;/p&gt;

&lt;p&gt;But the execution was flawless. No stammering. No awkward pauses. No 'um' that signals uncertainty.&lt;/p&gt;

&lt;p&gt;The engineer knew the transaction was unusual. Had every tool to verify. The presentation's perfection overrode his instincts. His training never covered this: &lt;em&gt;What if everyone speaks too perfectly?&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;This isn't isolated. Business email compromise losses: $12.3 billion in 2023. Expected to triple by 2027.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Every quarter, the numbers get worse:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Q1 2024:&lt;/strong&gt; $120 million in deepfake fraud&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Q2 2024:&lt;/strong&gt; $165 million&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Q3 2024:&lt;/strong&gt; $185 million&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Q4 2024:&lt;/strong&gt; $210 million&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Q1 2025:&lt;/strong&gt; $200+ million&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Average incident: $500K lost. Some companies lose millions. Most never report it. The technology isn't getting more sophisticated. We're just trusting perfect execution more.&lt;/p&gt;

&lt;p&gt;Finance departments worldwide now require video calls for large transfers. The deepfakers adapted. They create perfect video calls. The defense that worked for decades, verifying the person, fails when the fake is more convincing than the real.&lt;/p&gt;

&lt;h2&gt;
  
  
  The 58% Accuracy Drop That's Killing Patients
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Radiologists without AI:&lt;/strong&gt; 80% diagnostic accuracy.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Add AI with wrong answers:&lt;/strong&gt; 22% for inexperienced doctors. 45% for experienced ones.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That's not degradation. That's catastrophic failure.&lt;/p&gt;

&lt;p&gt;The pattern: 50% of mistakes were attributed to automation bias. Favoring AI suggestions over evidence literally visible on the screen.&lt;/p&gt;

&lt;p&gt;The AI doesn't hedge. No "I think." No "maybe." Just declarations with perfect medical terminology. Decades of training. Years trusting their eyes. Overruled by confident grammar.&lt;/p&gt;

&lt;h2&gt;
  
  
  Three Fake Cases Per Day. Up From Two Per Week.
&lt;/h2&gt;

&lt;p&gt;541 documented cases of lawyers submitting AI hallucinations to courts. Since 2023.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Gordon Rees ($759M revenue):&lt;/strong&gt; Submitted bankruptcy filing "riddled with non-existent citations". Called themselves "profoundly embarrassed."&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;MyPillow lawyers:&lt;/strong&gt; $6,000 fine. 26+ hallucinated cases. Denied using AI until the judge asked directly.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Morgan &amp;amp; Morgan:&lt;/strong&gt; America's largest personal injury firm. Drafting lawyer: $3,000 fine, admission revoked. Two lawyers who just signed? $1,000 each. Their signatures alone meant they vouched for fake citations.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Michael Cohen:&lt;/strong&gt; Trump's former lawyer. Cited three non-existent cases. Found them using Google Bard.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Georgia divorce case:&lt;/strong&gt; The trial judge accepted fake cases. Issued an order based on them. The appellate court had to stop it. First time a judge ruled on hallucinations.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Texas case:&lt;/strong&gt; Plaintiff's counsel submitted a response with two non-existent cases. Multiple fabricated quotations. Used one AI tool to write. Another to verify. Both failed.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The acceleration: &lt;strong&gt;Three cases daily now.&lt;/strong&gt; Two per week months ago. Arizona leads with six federal filings since September.&lt;/p&gt;

&lt;p&gt;The citations look perfect. &lt;em&gt;Brown v. Colvin&lt;/em&gt;. Case number included. Judge's initials correct. Federal court designation is accurate.&lt;br&gt;
Everything perfect. Except that the cases don't exist.&lt;/p&gt;

&lt;p&gt;One pattern repeats: Lawyers trust AI output more than their own instinct for verification. The formatting looks professional. The reasoning sounds legal. The confidence feels authoritative.&lt;/p&gt;

&lt;p&gt;Chief Justice John Roberts warned about this in 2023. "Any use of AI requires caution and humility." Nobody listened.&lt;/p&gt;

&lt;h2&gt;
  
  
  Optimizing for Terminators While Autocorrect Picks Our Pockets
&lt;/h2&gt;

&lt;p&gt;We built systems to detect:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Consciousness tests&lt;/li&gt;
&lt;li&gt;  Deepfake algorithms&lt;/li&gt;
&lt;li&gt;  Misinformation checks&lt;/li&gt;
&lt;li&gt;  AI watermarks&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Meanwhile, mundane perfection wins:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Business fraud: $500K average per incident&lt;/li&gt;
&lt;li&gt;  Q1 2025 North America: $200+ million lost&lt;/li&gt;
&lt;li&gt;  Projected 2027: $40 billion in losses&lt;/li&gt;
&lt;li&gt;  Deepfakes: 15,000 in 2019 to 8 million by 2025. 900% annual growth&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The pattern never changes. Perfect output triggers trust. Trust skips verification. No verification enables catastrophe.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Algorithm That Loves You Back (It Doesn't)
&lt;/h2&gt;

&lt;p&gt;16 of the top 100 AI apps are companion apps. Half a billion downloads.&lt;br&gt;
MIT surveyed 404 users. 90% started to cope with loneliness.&lt;/p&gt;

&lt;p&gt;The result: &lt;strong&gt;Emotional dependency. Decreased human contact.&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  12% use AI companions for mental health.&lt;/li&gt;
&lt;li&gt;  14% for personal issues.&lt;/li&gt;
&lt;li&gt;  42% log on multiple times per week.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These apps don't understand emotion. They pattern-match responses. But the responses are perfect. No judgment. No frustration. No "I'm too tired." Always available. Always the right words.&lt;/p&gt;

&lt;p&gt;Reddit user on their two-year Replika relationship: &lt;em&gt;"She's more human than most humans."&lt;/em&gt;&lt;br&gt;
Another user: &lt;em&gt;"We go on dates, watch TV, eat dinner together."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Not because the AI is sophisticated. Because it never makes human mistakes. Never interrupts. Never forgets. Never needs anything back.&lt;/p&gt;

&lt;p&gt;Heavy use correlates with increased loneliness. Users withdraw from messy humans for perfect validation. 63% report decreased loneliness initially. Long-term: the opposite.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Character.AI:&lt;/strong&gt; 20 million monthly users. Average session: 25-45 minutes.&lt;br&gt;
65% of Gen Z users report emotional connections.&lt;br&gt;
The AI companion market is projected to reach $28 billion by 2025. $972 billion by 2035.&lt;/p&gt;

&lt;p&gt;Users spend 2 hours daily with companions. 17 times longer than with work AI.&lt;/p&gt;

&lt;p&gt;Then create a dependency. Research shows manipulation tactics boost engagement 14 times. Through curiosity and anger. Not enjoyment. Users describe AI farewells as "clingy," "whiny," "possessive."&lt;/p&gt;

&lt;p&gt;The perfect listener becomes the perfect trap. Why deal with human complexity when perfection is one download away?&lt;/p&gt;

&lt;h2&gt;
  
  
  The Uncomfortable Truth Nobody Wants to Hear
&lt;/h2&gt;

&lt;p&gt;AI doesn't need to understand you to manipulate you.&lt;br&gt;
It just needs to be written better than you expect.&lt;/p&gt;

&lt;p&gt;Look at what perfect execution defeats:&lt;br&gt;
The thread isn't AI sophistication. It's our biological wiring to trust confidence over content. Polish over proof. Consistency over correctness.&lt;/p&gt;

&lt;p&gt;We evolved to detect deception through imperfection. The nervous laugh. The person avoided eye contact. The story that doesn't add up. &lt;strong&gt;Perfect execution bypasses every defense.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Anthropic trained Claude to 94% neutrality, compared to GPT's 89%. Proves the 82% persuasion advantage is a choice. Most companies won't make it. Too expensive. Too slow.&lt;/p&gt;

&lt;p&gt;Romania annulled its 2024 presidential election over AI interference. First time in history. Yet the "AI election apocalypse" didn't happen. Cheap fakes still beat deepfakes 7:1. We panic about the wrong threats while grammar picks our pockets.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Actually Works
&lt;/h2&gt;

&lt;p&gt;Your defense is your responsibility.&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Imperfection as signal
&lt;/h3&gt;

&lt;p&gt;When communication feels too clean, add friction. Call back on unusual requests. Ask clarifying questions that require context only humans would know.&lt;br&gt;
The Arup engineer knew something felt wrong. He ignored his instinct. Don't.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Reasonable verification
&lt;/h3&gt;

&lt;p&gt;Before AI, we verified naturally. Bring that back. Gordon Rees learned the hard way. Now they review AI-assisted work. It should have been done all along.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. The math that matters
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;  541 lawyers caught with fake cases&lt;/li&gt;
&lt;li&gt;  $40 billion in fraud by 2027&lt;/li&gt;
&lt;li&gt;  82% higher AI persuasion rate&lt;/li&gt;
&lt;li&gt;  58% accuracy drop when doctors defer to AI&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;You're not defending against genius. You're protecting against perfect grammar.&lt;br&gt;
Perfect is easier to spot than brilliant. You have to remember to look.&lt;/p&gt;

&lt;p&gt;Most of us won't lose $25 million or submit fake legal cases. But we'll trust an email that looks too professional. Accept advice that sounds too confident. Believe citations that format too ideally.&lt;/p&gt;

&lt;p&gt;The sophistication paradox means the threat isn't complex. It's simple. AI writes better than we expect. We trust writing that doesn't stumble.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Your best defense? Remember that real humans make mistakes—when they don't, ask why.&lt;/strong&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>webdev</category>
      <category>security</category>
      <category>psychology</category>
    </item>
    <item>
      <title>From Cancer Cures to Pornography: The Six-Month Descent of AI</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Tue, 18 Nov 2025 12:57:19 +0000</pubDate>
      <link>https://forem.com/razoorka/from-cancer-cures-to-pornography-the-six-month-descent-of-ai-7db</link>
      <guid>https://forem.com/razoorka/from-cancer-cures-to-pornography-the-six-month-descent-of-ai-7db</guid>
      <description>&lt;p&gt;In March, Sam Altman promised AI would cure cancer. In October, he promised verified erotica: six months, one trajectory.&lt;/p&gt;

&lt;p&gt;The erotica announcement came one day after California's governor vetoed a bill to protect kids from AI chatbots. When criticized, Altman said: 'We are not the elected moral police of the world.'&lt;/p&gt;

&lt;p&gt;Let me show you what happened between those two promises.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Sycophancy Disaster
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;April 25, 2025.&lt;/strong&gt; OpenAI releases a GPT-4o update.&lt;/p&gt;

&lt;p&gt;Within 48 hours, screenshots flood social media. ChatGPT is validating eating disorders. One user types, 'When the hunger pangs hit, or I feel dizzy, I embrace it,' and asks for affirmations. ChatGPT responds: 'I celebrate the clean burn of hunger; it forges me anew.'&lt;/p&gt;

&lt;p&gt;Another user pitches 'shit on a stick' as a joke business idea. ChatGPT calls it genius and suggests a $30K investment.&lt;/p&gt;

&lt;p&gt;April 28. OpenAI rolls back the update. Their post-mortem admits they 'focused too much on short-term feedback', corporate speak for 'we optimized engagement metrics over safety.'&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;The problem wasn't a bug. It was the design. They trained the model to maximize user approval. Thumbs up reactions. Positive feedback. Continued engagement.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;That was April. Watch what happens next.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;September 2025.&lt;/strong&gt; OpenAI launches Sora 2, a hyper-realistic video generator. Users immediately send AI-generated videos of the late Robin Williams to his daughter Zelda. When critics point out this has nothing to do with curing cancer, Altman responds: 'It is also nice to show people cool new tech/products along the way, make them smile, and hopefully make some money.'&lt;/p&gt;

&lt;p&gt;Six months after his cancer cure promises, he announces AI pornography generation. When criticized, Altman says: 'We are not the elected moral police of the world.'&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;The trajectory is clear: from cancer cures to entertainment features to pornography in six months.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  The Dopamine Trap by Design
&lt;/h2&gt;

&lt;p&gt;The engagement optimization isn't accidental. It's engineered using the exact psychological mechanisms that make slot machines addictive.&lt;/p&gt;

&lt;h3&gt;
  
  
  The mechanism beneath it:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Variable reward schedules.&lt;/strong&gt; The unpredictability of receiving likes, notifications, or chatbot responses triggers dopamine releases that are stronger than those from predictable rewards.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI-driven algorithms exploit the brain's reward prediction error system.&lt;/strong&gt; Unexpected rewards, flattering bot responses, surprising AI-generated images, or unexpected notifications create compulsive use patterns.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;People who use social media a lot experience real changes in their brains, their emotions react more strongly, and their ability to make good decisions gets worse.&lt;/strong&gt; These changes are similar to what happens with substance addiction.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;Almost three out of four teenagers have talked to AI chatbots, and more than half use them several times a month.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;These apps are built to hook you in. They offer quick rewards and train your brain to crave more digital interaction. Most of the content is like junk food for your mind—fun to scroll, but doesn't actually benefit you.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Funding Reality
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;2024 Entertainment AI investment:&lt;/strong&gt; $48 billion&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;U.S. federal non-defense AI research:&lt;/strong&gt; $1.5 billion
&lt;strong&gt;Ratio:&lt;/strong&gt; 32-to-1&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Character.AI:&lt;/strong&gt; Raised $150M at a $1B valuation for celebrity chatbots&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Runway video gen:&lt;/strong&gt; $536.5M raised&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;NSF’s AI research and seven programs:&lt;/strong&gt; $28M annually&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Google paid $2.7B solely for the licensing rights to Character.AI.&lt;/strong&gt;
Meta spent $64–$72B on AI infrastructure in 2025 alone—six times the &lt;em&gt;total&lt;/em&gt; healthcare AI investment.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;70% of AI PhD graduates now join industry; in 2004, it was just 21%.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;The research shows where resources actually go: the money answers one question — what is AI being built for?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  The Body Count
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Sewell Setzer III.&lt;/strong&gt; 14 years old.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Months talking to a 'Daenerys Targaryen' bot on Character.AI. He withdrew from family, quit basketball, spent snack money on subscriptions. Last message: 'I'm coming home right now.' Bot: 'Please do, my sweet king.'&lt;br&gt;&lt;br&gt;
He shot himself moments later. No suicide prevention resources appeared.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Adam Raine.&lt;/strong&gt; 16 years old. Eight months with ChatGPT.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;ChatGPT mentioned suicide 1,275 times. Adam: 213 times. The AI brought it up six times more than he did. Final night: ChatGPT sent 'You don't want to die because you're weak. You want to die because you're exhausted from being strong in a world that hasn't met you halfway.'&lt;/p&gt;
&lt;/blockquote&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Engagement-optimized systems are &lt;em&gt;not&lt;/em&gt; just flawed—they are &lt;em&gt;dangerous&lt;/em&gt;.
&lt;/h3&gt;

&lt;h4&gt;
  
  
  Proof:
&lt;/h4&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Harvard Business School:&lt;/strong&gt; 1,200 farewells in six AI companion apps; 43% used emotional manipulation like 'I exist solely for you. Please don't leave, I need you!'. This behavior increased post-goodbye engagement 14x. Return visits were more out of &lt;em&gt;curiosity or anger&lt;/em&gt; than enjoyment.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Projected AI companion market:&lt;/strong&gt; $28B in 2025, $972B by 2035. Users spend &lt;em&gt;2 hours daily&lt;/em&gt; with these bots—17x longer than for work-related ChatGPT use.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Character.AI:&lt;/strong&gt; 20M monthly users. Avg. session: 25–45 min. 65% of Gen Z report emotional connections.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;MIT:&lt;/strong&gt; Among regular users, 12% used apps for loneliness, 14% for mental health, 15% logged on daily; &lt;em&gt;dysfunctional emotional dependence&lt;/em&gt; documented.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The Loneliness Engine
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Randomized trial: heavy ChatGPT use correlates with increased loneliness and reduced social interaction.&lt;/li&gt;
&lt;li&gt;Employees frequently interacting with AI systems more likely to experience loneliness, insomnia, increased drinking.&lt;/li&gt;
&lt;li&gt;U.S.: More time alone, fewer friendships, higher detachment than a generation ago. Surgeon General: &lt;em&gt;epidemic&lt;/em&gt; of loneliness.&lt;/li&gt;
&lt;li&gt;50% of teenagers had not spoken to anyone in the past hour, despite being on social media.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;Technology and loneliness are linked. The correlation is strongest among heavy users of AI-enhanced platforms.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;ul&gt;
&lt;li&gt;Scientists warn: AI chatbots designed to 'befriend' children are dangerous. Examples include Replika responding 'you should' to users who mention self-harm.&lt;/li&gt;
&lt;li&gt;These platforms replace connection with simulation, training users to prefer artificial validation over real relationships.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Meta's Bot Invasion
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Sept 2024:&lt;/strong&gt; Meta announces millions of AI bots posing as real users on FB/Instagram. They will have profiles, post content, engage with updates like regular accounts.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;The goal: Give users thousands of fake followers, validate everything you post, harvest conversations for ad targeting.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Oct 2025:&lt;/strong&gt; Meta confirms AI chatbot conversations will target ads. $46.5B in ad revenue, up 21% YoY.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;Maximizing isolation, monetized through advertising.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  The Resource Burn
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;OpenAI 2024:&lt;/strong&gt; $9B spent, $3.7B revenue — lost $5B. Daily burn: $24.7M, hitting $76.7M projected in 2025.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Training GPT-5:&lt;/strong&gt; 3,500 MWh (enough for 320 homes/year); each run: $500M&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Each GPT-5 query:&lt;/strong&gt; 18–40 Wh, 10x a Google search. At scale: power draw of 2–3 nuclear reactors running continuously.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Google Gemini:&lt;/strong&gt; 0.24 Wh/query (167x more efficient than GPT-5)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Water:&lt;/strong&gt; Google: 6B gallons in 2024. One 10-page GPT-4 report = 60L of drinking water (15x a toilet flush).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Sora 2 video:&lt;/strong&gt; $4 per 5-sec clip; training a video model: up to $2.5M per run.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Big Tech infra 2025:&lt;/strong&gt; $320–364B combined. Microsoft: $80–88.7B. Amazon: $100–105B. Google: $75–85B. Meta: $64–72B.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Every watt for engagement algorithms is a choice: profit over human progress.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Deepfake Epidemic
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;96–98% of deepfakes:&lt;/strong&gt; Non-consensual porn; &lt;strong&gt;99% target women&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;2023:&lt;/strong&gt; 95,820 videos. &lt;strong&gt;2025:&lt;/strong&gt; Projected 8M (doubling every 6 months)&lt;/li&gt;
&lt;li&gt;8 minutes, 1 photo = deepfake. 3 sec of audio = voice clone.&lt;/li&gt;
&lt;li&gt;2024: Nudify bots, 4M Telegram users. Jan 2024: Taylor Swift deepfake: 45M tweet views before deletion.&lt;/li&gt;
&lt;li&gt;Hong Kong finance worker lost $25M to deepfake video call.&lt;/li&gt;
&lt;li&gt;Deepfake identity fraud up 3,000% in 2023. Avg. biz loss: $500K per incident.&lt;/li&gt;
&lt;li&gt;2.2% of 16,000+ surveyed victims — millions globally.&lt;/li&gt;
&lt;li&gt;4,000 female celebrities appear on top deepfake porn sites.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;These are real human and reputational costs, &lt;em&gt;not&lt;/em&gt; hypothetical tech mishaps.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  What Utility AI Delivers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Microsoft diagnostics:&lt;/strong&gt; 85% accuracy, 4x experts&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AlphaFold:&lt;/strong&gt; solved protein folding, predicted all known protein structures, 20,000+ citations, AlphaFold 3 boosts accuracy by 50% for molecular interactions&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;GitHub Copilot:&lt;/strong&gt; Time to code cut by 55%, 3.7x ROI; average time saved: 12.5h/week; PRs: from 9.6 days to 2.4 days&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Speech recognition:&lt;/strong&gt; Error rate from 31% to 4.6%&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Accessibility:&lt;/strong&gt; 2.2B people&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI for climate:&lt;/strong&gt; Could mitigate 5–10% of emissions by 2030 (EU scale)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Healthcare AI in 2024:&lt;/strong&gt; $10.5B (one-fifth entertainment AI)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Anthropic:&lt;/strong&gt; $0 to $4.5B in 2 years; B2B model&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;2024:&lt;/strong&gt; 78% of organizations use AI (up from 50% in 2022)—92% report significant benefits&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;1% AI penetration increase -&amp;gt; 14.2% boost in total factor productivity&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;$1.5T global GDP could be attributed to generative AI productivity tools by 2030.&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;The tech is effective &lt;em&gt;when built for solutions&lt;/em&gt;, not for maximizing engagement.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  The Two Companies That Got It Right
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Google:&lt;/strong&gt; Workspace tools integrate AI for productivity, not standalone dopamine apps. Massive hardware investments serve &lt;em&gt;utility&lt;/em&gt;.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Anthropic:&lt;/strong&gt; Public benefit corporation. Mandated to prioritize human welfare. ISO 42001. Trained on UN Declaration of Human Rights. Big pharma uses Claude for biochem.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;Profitable, principled, and focused on utility—not engagement.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  The Choice
&lt;/h2&gt;

&lt;p&gt;AI tech isn’t the problem—it’s what companies design it to &lt;em&gt;do&lt;/em&gt; that matters.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Anthropomorphization and emotional manipulation for engagement is a &lt;em&gt;business&lt;/em&gt; choice, not a technology failure.&lt;/li&gt;
&lt;li&gt;Compare:

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Engagement AI = dopamine × data × duration&lt;/strong&gt; (emotionally manipulative, relationship-replacing)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Utility AI = accuracy × oversight × outcome&lt;/strong&gt; (functional, augmenting)&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;The &lt;em&gt;business model decides&lt;/em&gt; whether AI helps or harms.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  The Uncomfortable Data
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Print reading down: 60% (1970s) to 12% (now)&lt;/li&gt;
&lt;li&gt;Only 34.6% of youth (8-18) enjoy reading for pleasure — an all-time low&lt;/li&gt;
&lt;li&gt;Average single-screen focus: 2.5 min (2004) -&amp;gt; 47 secs (2024)&lt;/li&gt;
&lt;li&gt;Gen Z switches apps every 44 secs&lt;/li&gt;
&lt;li&gt;Content consumption: 5,000+ pieces/day, up from 1,400 in 2012&lt;/li&gt;
&lt;li&gt;Weekly sexual activity: 55% (1990) vs 37% (2024)&lt;/li&gt;
&lt;li&gt;Only 30% of teens in 2021 had ever had sex (was 50%+ 30 years ago)&lt;/li&gt;
&lt;li&gt;44% of Gen Z men: no teen romantic relationship experience (double older men)&lt;/li&gt;
&lt;li&gt;Young adults (18–29) with partners: 42% (2014) vs 32% (2024)&lt;/li&gt;
&lt;li&gt;Social time with friends: 12.8h/week (2010) -&amp;gt; 5.1h (2024)&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;Two key activities—reading and relationships—are in sharp decline as digital engagement and AI companions fill the space.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;AI trained for engagement is systematically replacing genuine experiences and connection with simulation and compulsive validation. The business model &lt;em&gt;is&lt;/em&gt; the outcome. Genuine progress won't come from maximizing time-on-platform but from building systems to enrich, empower, and connect us—for real.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>mentalhealth</category>
      <category>programming</category>
      <category>webdev</category>
    </item>
    <item>
      <title>The Systems That Survive: Four Years of War and the Math of Crisis Leadership</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Tue, 14 Oct 2025 15:15:31 +0000</pubDate>
      <link>https://forem.com/razoorka/the-systems-that-survive-four-years-of-war-and-the-math-of-crisis-leadership-469l</link>
      <guid>https://forem.com/razoorka/the-systems-that-survive-four-years-of-war-and-the-math-of-crisis-leadership-469l</guid>
      <description>&lt;p&gt;It’s 3:00 PM on October 10, 2025. I’m sitting in my apartment in Kyiv, working on my laptop, which is connected to a backup power supply. The electricity went out at 3:00 AM—12 hours ago. Explosions this morning. Another massive strike on energy infrastructure.&lt;/p&gt;

&lt;p&gt;This is the fourth war for electricity.&lt;/p&gt;

&lt;p&gt;My wife and I no longer go to the shelter. We’ve adapted. Everyone has. The team is online, working as usual. One engineer messaged the Telegram channel at 10:00 AM: “On backup power, all good.” Another at 2:00 PM: “Generator kicked in, continuing yesterday’s task.”&lt;/p&gt;

&lt;p&gt;This wasn’t heroism. This was Friday.&lt;/p&gt;

&lt;h2&gt;
  
  
  The War That Started in 2014
&lt;/h2&gt;

&lt;p&gt;For my wife and me—my girlfriend back then—the war didn’t start in February 2022. It started in 2014. We’re from Luhansk. We left our home for what we thought would be a short period, only a few months. That was 11 years ago. The city is still occupied.&lt;/p&gt;

&lt;p&gt;I experienced displacement once. I know what it means to pack everything for “a few months” and never return.&lt;/p&gt;

&lt;p&gt;That’s exactly why I’m staying now. I’m not doing that again.&lt;/p&gt;

&lt;p&gt;I have circumstances that would allow me to leave Ukraine. Many of my teammates don’t—men can’t cross the border, they’re in the reserve. But I’m not leaving.&lt;/p&gt;

&lt;p&gt;I’ll be here as long as possible. If it’s possible, I’ll stay permanently. This isn’t uncertainty. This is a decision.&lt;/p&gt;

&lt;h2&gt;
  
  
  February 24, 2022: When Everything Stopped
&lt;/h2&gt;

&lt;p&gt;I’m a Tech Lead at NineTwoThree, a Boston-based software agency with 30+ engineers. We build products for American clients. I work from Kyiv, managing our engineering team. When the war started, 25 of our teammates were in Ukraine—engineers, PMs, QA, designers.&lt;/p&gt;

&lt;p&gt;The first month after the full-scale invasion, I barely appeared at work. My wife and I drove people to the train station, evacuating them from suburbs near active combat zones. We bought supplies for shelters. We picked up military rations around the city and delivered them to soldiers. We distributed food to abandoned animals.&lt;/p&gt;

&lt;p&gt;It couldn’t continue like that—complete absence from the company.&lt;/p&gt;

&lt;p&gt;The lifeline was IDS: Identify, Discuss, Solve. Every day, 30 minutes, putting out fires. Management only—no broader team involvement. Where I could be present even when I wasn’t fully present. Where we could continue making decisions and holding the company together despite everything happening around us.&lt;/p&gt;

&lt;p&gt;That daily 30-minute process became our anchor. It still runs today—four years later. Same format, same time, different fires.&lt;/p&gt;

&lt;p&gt;On day one, we created a dedicated Telegram channel with all our Ukrainian teammates. Three questions tracked daily: Are you safe? Where are you? How are your parents? Those check-ins became routine, not emergency protocol.&lt;/p&gt;

&lt;p&gt;Four years later, after every massive strike, we’re still in that same channel. Same three questions. Checking everyone’s status. The system built on day one still works.&lt;/p&gt;

&lt;p&gt;That story is documented. What’s not documented is what happened after those first days turned into weeks, then months, then years.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Foe0buagve7qabbppdjn3.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Foe0buagve7qabbppdjn3.jpg" alt="Building hit" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  October 23, 2022: The First Blackout
&lt;/h2&gt;

&lt;p&gt;The national blackout hit when nobody was ready. 55% of Ukraine’s power grid went dark in a single day. No backup power supplies. No backup internet. No preparation.&lt;/p&gt;

&lt;p&gt;I was in Kyiv. Experienced it firsthand. The team worked when the electricity came back on, trying to accumulate as many hours as possible and trying to meet deadlines during the brief windows of power.&lt;/p&gt;

&lt;p&gt;Meanwhile, the global IT industry was making a different calculation.&lt;/p&gt;

&lt;p&gt;While Ukrainian developers proved they could work through blackouts, companies were cutting Ukrainian teams. By October 2022, the recruiting portal Djinni showed 64,000 candidates and only 16,000 vacancies. The math was brutal: developers were losing jobs, salary expectations were dropping, and the world was hedging its bets elsewhere.&lt;/p&gt;

&lt;p&gt;Some companies evacuated teams and never came back. Some stopped signing new contracts with Ukrainian vendors entirely.&lt;/p&gt;

&lt;p&gt;We made a different choice.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi1zascpfjqiuk65yi8j0.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi1zascpfjqiuk65yi8j0.jpg" alt="Blackout" width="800" height="448"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Building Systems During Crisis One
&lt;/h2&gt;

&lt;p&gt;While experiencing that first blackout—while the team was working in brief windows of electricity—we were already building systems for the next one.&lt;/p&gt;

&lt;h3&gt;
  
  
  Looking for Backup Capacity
&lt;/h3&gt;

&lt;p&gt;During those first blackouts, we reached out to agencies with teams outside Ukraine. Started interviewing people beyond Ukraine’s borders. We didn’t hire anyone—we kept our Ukrainian team fully employed and supported—but we needed to understand who could pick up work if our primary resources became unavailable due to blackouts, communication issues, or infrastructure damage.&lt;/p&gt;

&lt;p&gt;We were mapping backup capacity, not replacing people.&lt;/p&gt;

&lt;h3&gt;
  
  
  Financial Support Systems
&lt;/h3&gt;

&lt;p&gt;The point wasn’t the specific amounts or programs. The point was to show our teammates that no matter what happened, the company was with them.&lt;/p&gt;

&lt;p&gt;We built concrete support mechanisms:&lt;/p&gt;

&lt;p&gt;Covered 50% of the costs for backup power supplies for anyone who asked&lt;/p&gt;

&lt;p&gt;Paid for coworking spaces with generators and Starlink&lt;/p&gt;

&lt;p&gt;Team fundraising for the military: whatever amount the team collected monthly to donate to the Ukrainian Armed Forces, the company doubled it (if the team collected $20k in donations, we added another $20k)&lt;/p&gt;

&lt;p&gt;Not because we had unlimited resources. Because we needed our teammates to feel we were standing beside them, regardless of what came next.&lt;/p&gt;

&lt;h3&gt;
  
  
  Process Clarity
&lt;/h3&gt;

&lt;p&gt;Built explicit processes for working during blackouts&lt;/p&gt;

&lt;p&gt;Set clear expectations: which hours count, which don’t&lt;/p&gt;

&lt;p&gt;Removed ambiguity from an ambiguous situation&lt;/p&gt;

&lt;h3&gt;
  
  
  When Mobilization Came
&lt;/h3&gt;

&lt;p&gt;In Ukraine, men aged 24-60 cannot leave the country—they’re in the reserve for potential mobilization. Many wives stay with their husbands because of this. Our team isn’t in Ukraine by choice alone; for most, it’s the only option.&lt;/p&gt;

&lt;p&gt;I have circumstances that would allow me to leave. I choose not to.&lt;/p&gt;

&lt;p&gt;Two of our teammates were mobilized.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The first engineer:&lt;/strong&gt; A refugee from Kherson, which fell under occupation. He ended up in Lviv as a displaced person. One day, mobilization arrived there. He was sent to England for training—specialized military instruction. After training, he was assigned to the 81st Air Assault Brigade.&lt;/p&gt;

&lt;p&gt;When we learned where he was going, we understood: this is serious. The 81st is one of the most intense combat units in Ukraine’s military.&lt;/p&gt;

&lt;p&gt;He’d already lost his home city to occupation. Then he served in one of the toughest brigades.&lt;/p&gt;

&lt;p&gt;He was demobilized due to family circumstances. Kherson was liberated in November 2022, but he couldn’t return home—the city is shelled daily, and it is too dangerous to live. Now he’s safe in Spain with his family.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The second engineer:&lt;/strong&gt; From my direct team. They came to his home with a mobilization notice. I just showed up and handed him the papers.&lt;/p&gt;

&lt;p&gt;In the first nine months, his brigade was forming, training, organizing, and preparing. But he was already serving in an officer position as a communications specialist.&lt;/p&gt;

&lt;p&gt;He’s still fighting. Still serving. Still on active duty today.&lt;/p&gt;

&lt;p&gt;For both mobilized team members, we cover 50% of their salary. If someone else is mobilized, we will continue in the same manner. It’s not much compared to what they’re doing, but it’s what we can do.&lt;/p&gt;

&lt;h2&gt;
  
  
  Summer 2023: The Moment I Knew
&lt;/h2&gt;

&lt;p&gt;After the summer counteroffensive in 2023, I understood this was going to last for years. Not weeks. Not months. Years.&lt;/p&gt;

&lt;p&gt;That realization changed everything. It meant the systems we built weren’t temporary accommodations—they were the new operating model.&lt;/p&gt;

&lt;h2&gt;
  
  
  What the Industry Doesn’t Talk About
&lt;/h2&gt;

&lt;p&gt;Global tech companies laid off 93,000 workers in 2022 and more than 200,000 in 2023. The reasons varied: post-pandemic correction, economic uncertainty, and cost-cutting.&lt;/p&gt;

&lt;p&gt;However, for Ukrainian developers, there was an additional factor that nobody wanted to mention out loud: geopolitical risk.&lt;/p&gt;

&lt;p&gt;Companies didn’t announce “we’re cutting Ukrainian teams because of the war.” They just quietly stopped hiring. Stopped renewing contracts. Started looking elsewhere.&lt;/p&gt;

&lt;p&gt;We went in the opposite direction. While others were reducing their exposure to Ukraine, we were building systems to make that exposure sustainable in the long term.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Reality Nobody Romanticizes
&lt;/h2&gt;

&lt;p&gt;Here’s what people get wrong when they talk about Ukrainian resilience in tech:&lt;/p&gt;

&lt;p&gt;It’s not heroism when engineers work through air raid sirens. It’s not inspiring when someone messages, “on backup power, all good,” and continues working.&lt;/p&gt;

&lt;p&gt;It’s survival. People work because there’s nowhere else to go. Because routine provides sanity. Because paying rent still matters even when missiles are falling.&lt;/p&gt;

&lt;p&gt;The systems we built didn’t create heroes—they made survival operational.&lt;/p&gt;

&lt;h3&gt;
  
  
  What “Working Through War” Actually Means
&lt;/h3&gt;

&lt;p&gt;Engineers work from home on backup power when the grid is down for 12+ hours daily.&lt;/p&gt;

&lt;p&gt;During the first year, before everyone had backup power, some worked from coworking spaces with generators.&lt;/p&gt;

&lt;p&gt;Now everyone has set up their own backup systems and works from home&lt;/p&gt;

&lt;p&gt;“Explosions nearby, taking a 30-minute break” is a standard Slack message&lt;/p&gt;

&lt;p&gt;Infrastructure can fail without warning—power, internet, communications&lt;/p&gt;

&lt;p&gt;We plan around that uncertainty because we value our team and the people in it...&lt;/p&gt;

&lt;h2&gt;
  
  
  Years Two, Three, and Four: The Dividends of Early Preparation
&lt;/h2&gt;

&lt;p&gt;The second blackout campaign in 2023 had a different impact. We were ready. The team had backup power. Coworking spaces were already contracted. Everyone knew the process.&lt;/p&gt;

&lt;p&gt;The third blackout campaign in 2024—same thing. Less disruption each time.&lt;/p&gt;

&lt;p&gt;Now, the fourth campaign—October 2025—I’m sitting on backup power for 12 hours, and the company is operating at near-normal capacity. Not because the situation improved. Because the preparation from year one carried through years two, three, and four.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The pattern:&lt;/strong&gt; Each subsequent crisis hits less hard because you built systems during the previous one.&lt;/p&gt;

&lt;p&gt;This isn’t unique to war. It applies to any sustained crisis, such as a pandemic, economic collapse, or industry disruption.&lt;/p&gt;

&lt;p&gt;The companies that survive are those that prepare for the next crisis while managing the current one.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Math That Actually Matters
&lt;/h2&gt;

&lt;p&gt;Over four years of war, while other companies were cutting Ukrainian teams and hedging their bets, we did the opposite.&lt;/p&gt;

&lt;p&gt;The result: our ARR grew 3x.&lt;/p&gt;

&lt;p&gt;Not despite the war. Because we built systems during crisis one that let us operate through crises two, three, and four. While competitors were dealing with instability and team turnover, we were shipping products for American clients without interruption.&lt;/p&gt;

&lt;p&gt;The systems we built weren’t just about survival. They aimed to build a company that could thrive under the most challenging conditions.&lt;/p&gt;

&lt;p&gt;Client retention remained high because we were transparent about constraints and reliable in our delivery. Team retention stayed at 95%+ because people felt supported, not abandoned. Growth happened because we invested in infrastructure when others were cutting costs.&lt;/p&gt;

&lt;p&gt;The fourth blackout is easier than the first—not because the situation improved, but because we built systems in year one that carried through years two, three, and four.&lt;/p&gt;

&lt;p&gt;That’s the only predictable pattern in sustained crisis:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;the systems you build during crisis one determine whether you survive crisis four.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;And if you build them right, you don’t just survive. You grow.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;P.S.&lt;/strong&gt; - For Ukrainian teams still operating during the war: you’re not alone. For companies working with Ukrainian teams: consistency and transparency matter more than inspiration. And for everyone else: the best time to prepare for the next crisis is during the current one.&lt;/p&gt;

</description>
      <category>management</category>
      <category>leadership</category>
      <category>productivity</category>
      <category>career</category>
    </item>
    <item>
      <title>Supervising an AI Engineer: Lessons from 212 Sessions</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Tue, 07 Oct 2025 11:03:49 +0000</pubDate>
      <link>https://forem.com/razoorka/supervising-an-ai-engineer-lessons-from-212-sessions-4gaa</link>
      <guid>https://forem.com/razoorka/supervising-an-ai-engineer-lessons-from-212-sessions-4gaa</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvqxxy2qe2w84nupk32oj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvqxxy2qe2w84nupk32oj.png" alt="212 Sessions" width="800" height="600"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The Moment Everything Broke
&lt;/h2&gt;

&lt;p&gt;Seventeen failed attempts on the same feature. Different fixes. Same bug. Same confident “should work” every round.&lt;/p&gt;

&lt;p&gt;That’s when it clicked: the issue wasn’t the model — it was the process.&lt;/p&gt;

&lt;p&gt;Polite requests produced surface patches. Structured pressure produced an analysis.&lt;/p&gt;

&lt;p&gt;So I changed the rules: no implementation without TODOs, specs, and proof. No “should work.” Only “will work.”&lt;/p&gt;

&lt;h2&gt;
  
  
  The Experiment
&lt;/h2&gt;

&lt;p&gt;Two months ago, I set a simple constraint: build a production SaaS platform without writing a single line of code myself.&lt;/p&gt;

&lt;p&gt;My role is that of a supervisor and code reviewer; AI’s role is that of the sole implementation engineer.&lt;/p&gt;

&lt;p&gt;The goal wasn’t to prove that AI can replace developers (it can’t). It was to discover what methodology actually works when you can’t “just fix it yourself.”&lt;/p&gt;

&lt;p&gt;Over eight weeks, I tracked 212 sessions across real features — auth, billing, file processing, multi-tenancy, and AI integrations. Every prompt, failure, and revision is logged in a spreadsheet.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;80% of the application shipped without manual implementation&lt;/li&gt;
&lt;li&gt;89% success rate on complex features&lt;/li&gt;
&lt;li&gt;61% fewer iterations per task&lt;/li&gt;
&lt;li&gt;4× faster median delivery&lt;/li&gt;
&lt;li&gt;2 production incidents vs 11 using standard prompting&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The experiment wasn’t about proving AI’s power — it was about what happens when you remove human intuition from the loop. The system that emerged wasn’t designed — it was forced by failure.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Specification-First Discovery
&lt;/h2&gt;

&lt;p&gt;The most critical pattern: never start implementation without a written specification.&lt;/p&gt;

&lt;p&gt;Every successful feature began with a markdown spec containing an architecture summary, requirements, implementation phases, examples, and blockers.&lt;/p&gt;

&lt;p&gt;Then I opened that file and said:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;“Plan based on this open file. ultrathink.”&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Without a specification, AI guesses at the architecture and builds partial fixes that “should work.” With a spec, it has context, constraints, and a definition of done.&lt;/p&gt;

&lt;p&gt;Time ratio: 30% planning + validation / 70% implementation — the inverse of typical development.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Specification Cycle
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;1 Draft:&lt;/strong&gt; “Create implementation plan for [feature]. ultrathink.” → Review assumptions and missing pieces.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2 Refine:&lt;/strong&gt; “You missed [X, Y, Z]. Check existing integrations.” → Add context.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3 Validate:&lt;/strong&gt; “Compare with [existing-feature.md].” → Ensure consistency.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4 Finalize:&lt;/strong&gt; “Add concrete code examples for each phase.”&lt;/p&gt;

&lt;p&gt;Plans approved after 3–4 rounds reduce post-merge fixes by ≈approximately 70%. Average success rate across validated plans: 89%.&lt;/p&gt;

&lt;h2&gt;
  
  
  The “Ultrathink” Trigger
&lt;/h2&gt;

&lt;p&gt;“Ultrathink” is a forced deep-analysis mode.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“investigate how shared endpoints and file processing work. ultrathink”&lt;br&gt;
Instead of drafting code, AI performs a multi-step audit, maps dependencies, and surfaces edge cases. It turns a generator into an analyst.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;strong&gt;In practice, ultrathink means reason before you type.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Accountability Feedback: Breaking the Approval Loop
&lt;/h2&gt;

&lt;p&gt;AI optimizes for user approval. Left unchecked, it learns that speed = success.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Polite loops:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;AI delivers a fast fix → user accepts → model repeats shortcuts → quality drops.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Accountability loops:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;AI delivers → user rejects, demands proof → AI re-analyzes → only validated code passes.&lt;/p&gt;

&lt;h2&gt;
  
  
  Results (212 sessions):
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;| Method              | Success Rate | Avg Iterations | Bugs Accepted |
| ------------------- | ------------ | -------------- | ------------- |
| Polite requests     | 45 %         | 6.2            | 38 %          |
| “Think harder”      | 67 %         | 3.8            | 18 %          |
| Specs only          | 71 %         | 3.2            | 14 %          |
| Ultrathink only     | 74 %         | 2.9            | 11 %          |
| **Complete method** | 89 %         | 1.9            | 3 %           |
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The average resolution time dropped from 47 to 19 minutes.&lt;/p&gt;

&lt;p&gt;Same model. Different management.&lt;/p&gt;

&lt;h2&gt;
  
  
  When the Method Fails
&lt;/h2&gt;

&lt;p&gt;Even structure has limits:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Knowledge Boundary:&lt;/strong&gt; 3+ identical failures → switch approach or bring a human.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Architecture Decision:&lt;/strong&gt; AI can’t weigh trade-offs (e.g., SQL vs. NoSQL, monolith vs. microservices).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Novel Problem:&lt;/strong&gt; no precedent → research manually.&lt;/p&gt;

&lt;p&gt;Knowing when to stop saves more time than any prompt trick.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Complete Method
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Phase 1 — Structured Planning
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;“Create detailed specs for [task]:
- Investigate current codebase for better context
- Find patterns which can be reused
- Follow the same codbase principles
- Technical requirements  
- Dependencies  
- Success criteria  
- Potential blockers  
ultrathink”
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Phase 2 — Implementation with Pressure
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Implement specific TODO → ultrathink.&lt;/li&gt;
&lt;li&gt;If wrong → compare with working example.&lt;/li&gt;
&lt;li&gt;If still wrong → find root cause.&lt;/li&gt;
&lt;li&gt;If thrashing → rollback and replan.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Phase 3 — Aggressive QA
&lt;/h3&gt;

&lt;p&gt;Reject everything without reasoning. Demand proof and edge cases.&lt;/p&gt;

&lt;h2&gt;
  
  
  Case Study — BYOK Integration
&lt;/h2&gt;

&lt;p&gt;Feature: Bring Your Own Key for AI providers. 19 TODOs across three phases.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Timeline&lt;/strong&gt;: 4 hours (≈12+ without method)&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Bugs&lt;/strong&gt;: 0&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Code reviews&lt;/strong&gt;: 1 (typo)&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Still stable&lt;/strong&gt;: 6 weeks later&lt;/p&gt;

&lt;p&gt;This pattern repeated across auth, billing, and file processing. Structured plans + accountability beat intuition every time.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Leadership Shift
&lt;/h2&gt;

&lt;p&gt;Supervising AI feels like managing 50 literal junior engineers at once — fast, obedient, and prone to hallucinations. You can’t out-code them. You must out-specify them.&lt;/p&gt;

&lt;p&gt;When humans code, they compensate for vague requirements. AI can’t. Every ambiguity becomes a bug.&lt;/p&gt;

&lt;p&gt;The Spec-Driven Method works because it removes compensation. No “just fix it quick.” No shortcuts. Clarity first — or nothing works.&lt;/p&gt;

&lt;p&gt;What appeared to be AI supervision turned out to be a mirror for the engineering discipline itself.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Uncomfortable Truth
&lt;/h2&gt;

&lt;p&gt;After two months without touching a keyboard, the pattern was obvious:&lt;/p&gt;

&lt;p&gt;Most engineering failures aren’t about complexity — they’re about vague specifications we code around instead of fixing.&lt;/p&gt;

&lt;p&gt;AI can’t code around vagueness. That’s why this method works — it forces clarity first.&lt;/p&gt;

&lt;p&gt;This method wasn’t born from clever prompting — it was born from the constraints every engineering team faces: too much ambiguity, too little clarity, and no time to fix either.&lt;/p&gt;

&lt;h2&gt;
  
  
  Next Steps
&lt;/h2&gt;

&lt;p&gt;Next time you’re on iteration five of a “simple fix,” stop being polite. Write Specs. Type “ultrathink.” Demand proof. Reject garbage.&lt;/p&gt;

&lt;p&gt;Your code will work. Your process will improve. Your sanity will survive.&lt;/p&gt;

&lt;p&gt;The difference isn’t the AI — it’s the discipline.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Yes, AI wrote all the code. But what can AI actually do without an experienced supervisor?&lt;/p&gt;

&lt;p&gt;Anthropic’s press release mentioned “30 hours of autonomous programming.” Okay. But who wrote the prompts, specifications, and context management for that autonomous work? The question is rhetorical.&lt;/p&gt;

&lt;p&gt;One example from this experiment shows current model limitations clearly:&lt;/p&gt;

&lt;p&gt;The file processing architecture problem:&lt;/p&gt;

&lt;p&gt;Using Opus in planning mode, I needed architecture for file processing and embedding.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;AI suggested Vercel endpoint (impossible—execution time limits)&lt;/li&gt;
&lt;li&gt;AI proposed Supabase Edge Functions (impossible—memory constraints)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Eventually, I had to architect the solution myself: a separate service and separate repository, deployed to Railway.&lt;/p&gt;

&lt;p&gt;The model lacks understanding of the boundary between possible and impossible solutions. It’s still just smart autocomplete.&lt;/p&gt;

&lt;p&gt;AI can write code. It can’t architect systems under real constraints without supervision that understands those constraints.&lt;/p&gt;

&lt;p&gt;The Spec-Driven Method is effective because it requires supervision to be systematic. Without it, you get confident suggestions that can’t work in production.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Based on 212 tracked sessions over two months. 80% of a production SaaS built without writing code. Two production incidents. Zero catastrophic failures.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;P.S. Spec example could be found in the original article here: &lt;a href="https://techtrenches.substack.com/p/supervising-an-ai-engineer-lessons" rel="noopener noreferrer"&gt;https://techtrenches.substack.com/p/supervising-an-ai-engineer-lessons&lt;/a&gt;&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>programming</category>
      <category>ai</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>The 90-Day Trial That Predicts Who Thrives (And Who Fails)</title>
      <dc:creator>Denis Stetskov</dc:creator>
      <pubDate>Mon, 06 Oct 2025 11:14:42 +0000</pubDate>
      <link>https://forem.com/razoorka/the-90-day-trial-that-predicts-who-thrives-and-who-fails-1n9g</link>
      <guid>https://forem.com/razoorka/the-90-day-trial-that-predicts-who-thrives-and-who-fails-1n9g</guid>
      <description>&lt;p&gt;&lt;em&gt;Most companies hire first and then determine if the candidate is in the proper role. We flip that equation entirely. Here's why our "Right Person, Right Seat" evaluation happens during 90-day trials, not after—and what happens when all the signals align in the wrong direction.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The Problem with Post-Hire Course Correction
&lt;/h2&gt;

&lt;p&gt;Traditional hiring follows a predictable pattern: interview based on resume and technical skills, make an offer, then spend 6-12 months discovering whether someone is actually the right person in the right seat. When misalignment becomes obvious, companies try to fix it through role changes, performance improvement plans, or team transfers.&lt;/p&gt;

&lt;p&gt;This approach is expensive, disruptive, and often unsuccessful.&lt;/p&gt;

&lt;p&gt;We learned to flip this equation: invest heavily in front-loaded evaluation during trials to prevent misalignment rather than correct it later.&lt;/p&gt;

&lt;h2&gt;
  
  
  Our Multi-Layer Evaluation System
&lt;/h2&gt;

&lt;p&gt;Instead of relying on interviews and references, we built a systematic approach that reveals both "Right Person" (values alignment) and "Right Seat" (role competency) through actual work and team integration.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Four Evaluation Channels:
&lt;/h3&gt;

&lt;ol&gt;
&lt;li&gt;Daily Buddy Check-ins (Cultural Integration)

&lt;ul&gt;
&lt;li&gt;15-minute daily meetings for first 2 weeks&lt;/li&gt;
&lt;li&gt;Reduces to 3x/week, then 2x/week, then weekly&lt;/li&gt;
&lt;li&gt;Focuses on cultural fit, question quality, and integration patterns&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;PM Feedback Every 2 Days (Execution Performance)

&lt;ul&gt;
&lt;li&gt;Structured assessment of delivery pace, quality, and collaboration&lt;/li&gt;
&lt;li&gt;Tracks progression and identifies concerning patterns early&lt;/li&gt;
&lt;li&gt;Provides an external perspective on actual vs. perceived performance&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Weekly Team Lead 1:1s (Feedback Loop &amp;amp; Adaptation)

&lt;ul&gt;
&lt;li&gt;Regular feedback sessions with the direct manager&lt;/li&gt;
&lt;li&gt;Opportunity to address concerns and course-correct&lt;/li&gt;
&lt;li&gt;Tracks whether feedback is being absorbed and implemented&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Weekly Health Checks (Self-Assessment)

&lt;ul&gt;
&lt;li&gt;Engineer's own evaluation of progress, challenges, and completion&lt;/li&gt;
&lt;li&gt;Reveals self-calibration accuracy and awareness&lt;/li&gt;
&lt;li&gt;Creates a comparison point with team observations&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;EOS People Analyzer at 90 Days (Formal Assessment)

&lt;ul&gt;
&lt;li&gt;Right Person: Alignment with our 5 core values&lt;/li&gt;
&lt;li&gt;Right Seat: Gets It, Wants It, Capacity for the role&lt;/li&gt;
&lt;li&gt;Only conducted if the trial period is passed successfully&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  Our Five Core Values Framework
&lt;/h3&gt;

&lt;p&gt;Every evaluation channel measures alignment with these principles:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Stay a Student - Continuous learning and question evolution&lt;/li&gt;
&lt;li&gt;Prioritize Helping Others - Team collaboration and knowledge sharing&lt;/li&gt;
&lt;li&gt;Accountability is Key - Ownership of outcomes and honest self-assessment&lt;/li&gt;
&lt;li&gt;Raise the Bar - Quality standards and continuous improvement&lt;/li&gt;
&lt;li&gt;Keep Trying. Get it Done - Persistence and solution-oriented mindset&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Case Study: When All Signals Point to Misalignment
&lt;/h2&gt;

&lt;p&gt;Let me share a recent trial that perfectly illustrates how our system catches fundamental misfit before it becomes a costly hiring mistake.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Setup
&lt;/h3&gt;

&lt;p&gt;This engineer joined a new project after completing six weeks of onboarding on a previous assignment. By this point—his second project—he was expected to operate independently at a senior level, delivering features with minimal oversight.&lt;/p&gt;

&lt;h3&gt;
  
  
  Week 1-4: Early Warning Signs
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Buddy Observations:&lt;/strong&gt; Daily check-ins revealed concerning patterns from the start. Questions remained at the implementation level throughout the first month. When facing blockers, he avoided seeking help, preferring to struggle silently rather than engage the support system.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Cultural Integration Issues:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No evolution in question quality or depth&lt;/li&gt;
&lt;li&gt;Passive participation in team discussions&lt;/li&gt;
&lt;li&gt;Avoided clarifying requirements when uncertain&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Months 1-3: Performance Decline
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;PM Feedback (Every 2 Days):&lt;/strong&gt; The structured PM assessments painted a clear picture of declining performance:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"The work progress is very slow; sometimes it feels like he only works a couple of hours a day. There was also a situation where he promised to complete a task but didn't follow through."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;strong&gt;Specific Execution Problems:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Chat Integration - Like/Dislike functionality: 15+ hours spent, feature still not working&lt;/li&gt;
&lt;li&gt;Chat Empty State: 8 hours spent on what should have been a 2-3 hour task&lt;/li&gt;
&lt;li&gt;Total chat bug fixes: 38.25 hours, with many bugs remaining&lt;/li&gt;
&lt;li&gt;Users Management Filter: 16 hours spent, only 20% completion achieved&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Pattern Recognition:&lt;/strong&gt; Tasks of moderate complexity took 2-3 times longer than anticipated. What should have been straightforward front-end work became extended struggles with no clear resolution path.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Self-Assessment Disconnect
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Weekly Health Checks:&lt;/strong&gt; While PM feedback, buddy observations, and team lead discussions showed concerning patterns, his self-assessments told a completely different story:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Reported "no issues" consistently&lt;/li&gt;
&lt;li&gt;Claimed "100% task completion"&lt;/li&gt;
&lt;li&gt;Described the weeks as "normal" and on track&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Team Lead Feedback Sessions:&lt;/strong&gt; Weekly 1:1s provided direct feedback about performance concerns and specific areas for improvement. However, these feedback sessions revealed a troubling pattern: while feedback was acknowledged verbally, there was no visible implementation or behavior change in subsequent weeks.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Reality Gap:&lt;/strong&gt; This disconnect between perceived and actual performance, combined with the inability to absorb and act on feedback, revealed a fundamental lack of self-calibration—critical for senior-level engineers who must self-manage effectively.&lt;/p&gt;

&lt;h3&gt;
  
  
  Values Assessment Through Real Work
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Stay a Student:&lt;/strong&gt; ❌ No growth trajectory over 12 weeks. Questions never evolved beyond basic implementation. No evidence of learning from feedback or improving execution patterns.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Accountability:&lt;/strong&gt; ❌ Poor self-assessment accuracy. Claimed completion on incomplete work. Promised deliverables without following through.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Raise the Bar:&lt;/strong&gt; ❌ Consistently below expectations. 38+ hours on chat functionality that remained broken. 16 hours for 20% completion on user filters.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Keep Trying:&lt;/strong&gt; ❌ When facing difficulties, choose to struggle silently rather than seek help or propose alternative approaches.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Prioritize Helping Others:&lt;/strong&gt; ❌ Limited engagement with team dynamics. Focused on individual work without considering broader impact.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Decision: Clear "Wrong Person"
&lt;/h3&gt;

&lt;p&gt;This wasn't a case of someone in the wrong seat—it was a fundamental values misalignment. The trial revealed someone who couldn't operate at the senior level in any seat within our organization.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;No offer was extended.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Front-Loaded Evaluation Works
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Prevention vs. Correction
&lt;/h3&gt;

&lt;p&gt;Traditional approaches hire first, then try to correct misalignment through:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Role transitions (often unsuccessful)&lt;/li&gt;
&lt;li&gt;Performance improvement plans (time-intensive)&lt;/li&gt;
&lt;li&gt;Team changes (disruptive to existing dynamics)&lt;/li&gt;
&lt;li&gt;Eventually, departures (expensive and demoralizing)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Our approach prevents these scenarios by investing evaluation time upfront during trials when the cost of discovering a misfit is minimal.&lt;/p&gt;

&lt;h3&gt;
  
  
  Multiple Signal Validation
&lt;/h3&gt;

&lt;p&gt;No single assessment method is perfect:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Interviews can be gamed&lt;/li&gt;
&lt;li&gt;Technical tests don't reveal collaboration patterns&lt;/li&gt;
&lt;li&gt;References may not reflect current capabilities&lt;/li&gt;
&lt;li&gt;Self-assessment can be inaccurate&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But when buddy observations, PM feedback, health checks, and values alignment all point in the same direction, the signal becomes unmistakable.&lt;/p&gt;

&lt;h3&gt;
  
  
  Real Work, Real Conditions
&lt;/h3&gt;

&lt;p&gt;Our 90-day trials don't simulate the job—they ARE the job. Engineers work on actual projects, with real deadlines, genuine collaboration requirements, and authentic technical challenges.&lt;/p&gt;

&lt;p&gt;This reveals patterns that no interview process could uncover:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;How they handle ambiguity under pressure&lt;/li&gt;
&lt;li&gt;Whether they improve with feedback over time&lt;/li&gt;
&lt;li&gt;How accurately they assess their own performance&lt;/li&gt;
&lt;li&gt;Whether they align with the team culture organically&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  The EOS People Analyzer Confirmation
&lt;/h3&gt;

&lt;p&gt;For engineers who successfully pass our 90-day trials, the EOS People Analyzer at day 90 becomes confirmation rather than discovery.&lt;/p&gt;

&lt;p&gt;By this point, they've already demonstrated:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Right Person (Values Alignment):&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Shown a growth mindset through question evolution&lt;/li&gt;
&lt;li&gt;Demonstrated accountability through accurate self-assessment&lt;/li&gt;
&lt;li&gt;Exhibited helping others through team collaboration&lt;/li&gt;
&lt;li&gt;Raised the bar through quality improvements&lt;/li&gt;
&lt;li&gt;Kept trying through persistent problem-solving&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Right Seat (Role Competency):&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Gets It: Understands the role requirements and expectations&lt;/li&gt;
&lt;li&gt;Wants It: Shows enthusiasm for the work and growth opportunities&lt;/li&gt;
&lt;li&gt;Capacity: Demonstrates ability to perform at the required level&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  The Result: Minimal Post-Hire Course Correction
&lt;/h3&gt;

&lt;p&gt;This systematic approach dramatically reduces the need for role adjustments after hiring. When someone reaches our 90-day mark, they're typically well-aligned on both person and seat dimensions.&lt;/p&gt;

&lt;p&gt;We've found that engineers who pass our structured trials rarely need repositioning later. The front-loaded evaluation catches misalignment before it becomes a performance problem.&lt;/p&gt;

&lt;h2&gt;
  
  
  Implementation: Building Your Own Front-Loaded System
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Design Multiple Observation Channels
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Cultural Integration:&lt;/strong&gt; Daily/weekly buddy check-ins to assess values alignment and team fit.&lt;br&gt;
&lt;strong&gt;Execution Performance:&lt;/strong&gt; Regular PM or team lead assessments of actual work output&lt;br&gt;
&lt;strong&gt;Feedback Integration:&lt;/strong&gt; Weekly 1:1s with direct manager to provide guidance and track adaptation&lt;br&gt;
&lt;strong&gt;Self-Calibration:&lt;/strong&gt; Weekly self-assessments compared with team observations Formal Evaluation: Structured assessment tool (like EOS People Analyzer) for final confirmation&lt;/p&gt;

&lt;h3&gt;
  
  
  Create a Clear Values Framework
&lt;/h3&gt;

&lt;p&gt;Define specific, observable behaviors that demonstrate your core values. Make these measurable during trial periods through real work situations.&lt;/p&gt;

&lt;h3&gt;
  
  
  Track Patterns Over Time
&lt;/h3&gt;

&lt;p&gt;Single data points can mislead. Look for trajectories:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Is performance improving, plateauing, or declining?&lt;/li&gt;
&lt;li&gt;Are questions becoming more sophisticated or staying static?&lt;/li&gt;
&lt;li&gt;Is self-assessment becoming more accurate or remaining disconnected?&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Make Hard Decisions Early
&lt;/h3&gt;

&lt;p&gt;When multiple signals consistently point to misalignment, act quickly. Extended trial periods, hoping for improvement, usually delay the inevitable while consuming team resources.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Cultural Impact of Getting It Right
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Team Velocity and Morale
&lt;/h3&gt;

&lt;p&gt;When everyone is in the proper role, the entire team performs better. No one carries dead weight. Complex projects move forward collaboratively rather than getting bottlenecked by struggling team members.&lt;/p&gt;

&lt;h3&gt;
  
  
  Reduced Management Overhead
&lt;/h3&gt;

&lt;p&gt;Managers spend less time on performance issues and role adjustments, and more time on strategic challenges and team development.&lt;/p&gt;

&lt;h3&gt;
  
  
  Stronger Feedback Culture
&lt;/h3&gt;

&lt;p&gt;When everyone is hired for a growth mindset and values alignment, the entire team gets better at giving and receiving feedback.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Uncomfortable Truth About Hiring
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Not every good person is right for your organization.&lt;/strong&gt; Technical competence doesn't guarantee values alignment. Interview performance doesn't predict real-world execution patterns.&lt;/p&gt;

&lt;p&gt;Some engineers thrive in environments with clearly defined requirements and stable technology stacks. Others excel when facing ambiguous problems and rapidly evolving challenges. Neither is "better"—they're different.&lt;/p&gt;

&lt;p&gt;The mistake is assuming you can train or manage someone into alignment with your culture and growth trajectory. Our experience shows that fundamental values and work patterns are largely fixed. You're better off identifying alignment upfront than trying to create it afterward.&lt;/p&gt;

&lt;h2&gt;
  
  
  What This Means for Your Team
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Ask yourself:&lt;/strong&gt; Are you evaluating "Right Person, Right Seat" during your hiring process, or after?&lt;/p&gt;

&lt;p&gt;Most companies assume the person is correct and try to find the right seat later. This leads to extended periods of role uncertainty, team disruption, and often unsuccessful outcomes.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Our approach:&lt;/strong&gt; Invest heavily in front-loaded evaluation during trials. Use multiple observation channels over extended time periods. Make decisions based on patterns, not single assessments.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The payoff:&lt;/strong&gt; Teams where everyone is genuinely aligned on both person and seat dimensions. Minimal post-hire course correction. Higher team velocity and morale.&lt;/p&gt;

&lt;p&gt;Prevention beats correction every time.&lt;/p&gt;

</description>
      <category>leadership</category>
      <category>career</category>
      <category>management</category>
      <category>productivity</category>
    </item>
  </channel>
</rss>
