<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Keith MacKay</title>
    <description>The latest articles on Forem by Keith MacKay (@keithjmackay).</description>
    <link>https://forem.com/keithjmackay</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1499463%2Fdc7d3636-8482-4d6e-b619-fb4367cf4dfd.jpg</url>
      <title>Forem: Keith MacKay</title>
      <link>https://forem.com/keithjmackay</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/keithjmackay"/>
    <language>en</language>
    <item>
      <title>Situational Leadership for AI: More Like a Capable Colleague than a Fancy Formula</title>
      <dc:creator>Keith MacKay</dc:creator>
      <pubDate>Mon, 11 May 2026 05:55:43 +0000</pubDate>
      <link>https://forem.com/keithjmackay/situational-leadership-for-ai-more-like-a-capable-colleague-than-a-fancy-formula-1hje</link>
      <guid>https://forem.com/keithjmackay/situational-leadership-for-ai-more-like-a-capable-colleague-than-a-fancy-formula-1hje</guid>
      <description>&lt;h1&gt;
  
  
  Situational Leadership for AI: More Like a Capable Colleague than a Fancy Formula
&lt;/h1&gt;

&lt;p&gt;&lt;em&gt;Treating LLMs like vending machines guarantees mediocre results—try situational leadership for better results&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;You wouldn’t hand a new hire a laptop, point vaguely at the codebase, and say “make it better.” You wouldn’t expect a brilliant data scientist to magically produce great marketing copy. You wouldn’t give the same instructions to a seasoned architect and a junior developer and expect identical results.&lt;/p&gt;

&lt;p&gt;And yet—this is exactly how many (most?) organizations interact with large language models.&lt;/p&gt;

&lt;p&gt;The executives who’ve spent decades learning to match tasks to talent, provide appropriate context, and adapt their leadership style to individual strengths suddenly forget everything when the “employee” runs on GPUs instead of caffeine. They pump generic prompts into their LLMs like quarters into a vending machine, then complain when the results disappoint.&lt;/p&gt;

&lt;p&gt;Here’s the uncomfortable truth: &lt;strong&gt;the skills that make you effective at managing capable knowledge workers are precisely the skills that make you effective at getting value from LLMs.&lt;/strong&gt; And if you’re bad at one, you’re probably bad at the other.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Situational Leadership Connection
&lt;/h2&gt;

&lt;p&gt;In the 1970s, Paul Hersey and Ken Blanchard developed Situational Leadership Theory—the radical notion that effective leaders (or parents) adapt their style to the person and task at hand. A directive approach works for someone new to a task. A supportive approach works for someone who knows what to do but lacks confidence. Delegation works for someone competent and committed.&lt;/p&gt;

&lt;p&gt;The model seems obvious in retrospect. Yet organizations still promote technical experts into management roles and watch them fail by treating every report identically—usually by treating everyone the way they themselves would want to be treated (we all see the world through our own lens of needs and desires and strengths and past experiences—the best managers come to understand the lens through which their reports are perceiving the world).&lt;/p&gt;

&lt;p&gt;Identical treatment for all gets shoddy results with teams. It also gives shoddy results with AI.&lt;/p&gt;

&lt;p&gt;Different LLMs have different capabilities, different training, different strengths, different weaknesses. Claude excels at nuanced reasoning and following complex instructions. GPT-4 has particular strengths in some business settings. Grok has some of the best information on trending cultural topics. Smaller models trade capability for speed. Specialized models dominate narrow domains. Treating them interchangeably—or worse, treating them like omniscient oracles—is the management equivalent of giving identical assignments to your entire team regardless of skill set.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The situational leader asks:&lt;/strong&gt; What does this specific person need from me to succeed at this specific task?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The effective AI user asks:&lt;/strong&gt; What does this specific model need from me to succeed at this specific task?&lt;/p&gt;

&lt;p&gt;The parallel is not metaphorical. It’s operational.&lt;/p&gt;

&lt;h2&gt;
  
  
  Context: the Briefing You Forgot to Give
&lt;/h2&gt;

&lt;p&gt;Picture a new employee’s first day. A good manager doesn’t just assign tasks—they provide context. Here’s what we’re building. Here’s why it matters. Here’s how your work connects to the larger mission. Here’s who does what and who can help with specific things as needed. Here are the constraints we’re operating under. Here are the decisions that have already been made and why. Who. What. When. Where. Why. How.&lt;/p&gt;

&lt;p&gt;Bad managers skip this. They assume context is obvious. They issue instructions and get frustrated when results miss the mark. “I asked for X, why did you give me Y?” Often it’s because they never explained what X meant in the organization, with its constraints, for its customers.&lt;/p&gt;

&lt;p&gt;LLMs suffer the same failure mode, magnified.&lt;/p&gt;

&lt;p&gt;When you prompt a model with “write me a marketing email,” you’ve given it essentially no context. What product? What audience? What tone? What’s the goal—awareness, conversion, retention? What’s worked before? What constraints exist? The model will generate something plausible and generic, because generic was all you gave it to work with.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Good managers are explicit about context. Effective AI users are explicit about context.&lt;/strong&gt; The quality of the output is directly proportional to the quality of the input—not the quantity of words, but the relevance of the information provided.&lt;/p&gt;

&lt;p&gt;This extends beyond single prompts. LLMs have session context—what’s been discussed, what’s been decided, what’s been tried. Just as a good manager maintains continuity across conversations with their team (”remember, we decided last week that...”), effective AI users maintain continuity across interactions.&lt;/p&gt;

&lt;p&gt;The model that helped you architect a system yesterday can help you implement it today—if you’ve preserved the context of those decisions. Wipe the slate clean and start over, and you’re back to onboarding a new hire who knows nothing about your project.&lt;/p&gt;

&lt;p&gt;Note that the level of granularity (what level of detail/specificity do you provide?) also matters. Provide too much, and you get what you ask for, with less opportunity for the AI to surprise you with some helpful additions. Provide too little, and the AI may drift away from your needs, creating something tangential but not quite right.&lt;/p&gt;

&lt;h2&gt;
  
  
  Task-Model Fit Is Just Role Clarity by Another Name
&lt;/h2&gt;

&lt;p&gt;Smart managers don’t assign accounting work to engineers or graphical work to people who think in spreadsheets. They match tasks to talents, projects to strengths, challenges to capabilities.&lt;/p&gt;

&lt;p&gt;This may strike you as more-basic-than-needed management advice. It’s also more rare than one might think for AI usage.&lt;/p&gt;

&lt;p&gt;Organizations often standardize on a single model for all use cases—the enterprise equivalent of hiring one type of person for every role. They use the same massive model to answer simple questions that a smaller model could handle faster and cheaper. They ask generalist models to perform specialist tasks. They wonder why results vary wildly.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Different models for different jobs:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Quick, factual queries&lt;/strong&gt; don’t need your most powerful (and expensive) model. A faster, smaller model often performs equivalently for routine tasks—just as you don’t need your most senior engineer to reset a password.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Complex reasoning&lt;/strong&gt; benefits from models trained to think step-by-step. Asking a model optimized for speed to navigate nuanced trade-offs is like asking your fastest coder to lead a design review. You’ll see this implemented in many models as a “thinking mode” of one sort or another.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Domain-specific work&lt;/strong&gt; often benefits from fine-tuned or specialized models. General practitioners are valuable; specialists exist for a reason. If you don’t have specialized models, ask your generalized model to adopt a persona that would be adept in the given domain, and you’ll find your tokens better spent -- the predictive model will provide answers weighted toward the desired domain, providing inherently better answers. Tune both the persona and the questions to the specific domain.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Creative tasks&lt;/strong&gt; may warrant different models than analytical tasks. The person who writes brilliant copy isn’t necessarily the person who debugs distributed systems.&lt;/p&gt;

&lt;p&gt;The effective AI user develops model literacy the way effective managers develop people literacy. They learn what each model does well, where each struggles, and how to route work accordingly.&lt;/p&gt;

&lt;h2&gt;
  
  
  Strengths and Weaknesses Aren’t Bugs—They’re Features
&lt;/h2&gt;

&lt;p&gt;Every employee has strengths to leverage and weaknesses to manage around. The brilliant architect who can’t stand presenting. The charismatic salesperson who crumbles under detail work. The reliable executor who struggles with ambiguity.&lt;/p&gt;

&lt;p&gt;Good managers realize that people are not uniformly excellent. They design teams and workflows that amplify strengths and compensate for weaknesses.&lt;/p&gt;

&lt;p&gt;LLMs are similar, but many organizations keep pretending they aren’t.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Current-generation models tend to struggle with:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Precise numerical reasoning (your accountant should check the math)&lt;/li&gt;
&lt;li&gt;Real-time information (they’re trained on historical data)&lt;/li&gt;
&lt;li&gt;Guaranteed factual accuracy (they can hallucinate convincingly)&lt;/li&gt;
&lt;li&gt;Following extremely long context with equal attention throughout&lt;/li&gt;
&lt;li&gt;Knowing what they don’t know&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Current-generation models tend to excel at:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Pattern recognition across large bodies of text&lt;/li&gt;
&lt;li&gt;Generating plausible first drafts quickly&lt;/li&gt;
&lt;li&gt;Explaining complex concepts at various levels&lt;/li&gt;
&lt;li&gt;Synthesizing information from multiple sources&lt;/li&gt;
&lt;li&gt;Brainstorming and ideation&lt;/li&gt;
&lt;li&gt;Reformatting and restructuring content&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The effective approach isn’t to wish away weaknesses—it’s to design workflows that play to strengths. Use the model for the first draft; have a human verify the facts. Use the model for synthesis; have a domain expert validate the conclusions. Use the model for ideation; have the team evaluate feasibility.&lt;/p&gt;

&lt;p&gt;This is management, not magic.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Feedback Loop That Everyone Ignores
&lt;/h2&gt;

&lt;p&gt;Managing knowledge workers isn’t a one-way broadcast. It’s a dialogue. You give direction. They execute. You provide feedback. They adjust. You refine requirements and improve processes. They deliver again, and may provide feedback of their own. Good managers create tight feedback loops that quickly converge on the right outcome.&lt;/p&gt;

&lt;p&gt;Most AI users treat prompts as one-shot commands, then complain when the first response isn’t perfect.&lt;/p&gt;

&lt;p&gt;Would you judge an employee’s capability based on their first draft of their first project on their first day? Then why judge a model based on a single cold-start response?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Effective AI interaction is iterative.&lt;/strong&gt; The first response is diagnostic—what did the model understand, what did it miss, where did it go wrong? The second prompt corrects course. The third refines. By the fourth or fifth exchange, you’ve co-created something neither you nor the model would have produced alone.&lt;/p&gt;

&lt;p&gt;This is how you work with smart humans. Also how you work with smart machines. Note that this is a reason that some recommend less detail up front for some tasks — the AI will generate more concepts for consideration that can expand your idea when it is allowed more wiggle-room. Different model strengths will affect this — ChatGPT has gotten better and better at providing exactly what you ask for, while Claude has gotten better and better at expanding on your inputs with additional ideas. Knowing this might cause you to choose one or the other for a specific task, or shape how you write your prompt.&lt;/p&gt;

&lt;p&gt;The executive who says “I tried ChatGPT once and it gave me garbage” is the executive who says “I told my team what I wanted once and they didn’t deliver exactly what I envisioned.” Both statements reveal more about the executive than about the team or tool.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Uncomfortable Implication
&lt;/h2&gt;

&lt;p&gt;Here’s where this gets interesting: if AI effectiveness requires the same skills as people management, then organizations’ AI results will correlate with their management culture.&lt;/p&gt;

&lt;p&gt;Companies are expecting young non-managers to become expert AI users without helping them understand how to manage. I believe this is the primary reason for the statistics showing much more productive use of AI by older users than by younger users...the inverse of historical technology usage patterns. Companies that are trimming hiring of entry-level folks to do more of the work with AI may have a hard time if they don’t create a path to management, with management training, for their smaller classes of entry-level employees. They will also need to build retention mechanisms to keep a higher percentage of entry-level folks longer...or they’ll wind up with hollowed-out organizations.&lt;/p&gt;

&lt;p&gt;Companies with clear communication norms and management training will get clearer AI outputs. Organizations that provide context will get contextually appropriate responses. Teams practiced in iterative refinement will iterate effectively with AI. Leaders who match tasks to capabilities will match tasks to models.&lt;/p&gt;

&lt;p&gt;And organizations with terrible management practices? They’ll get terrible AI results and blame the technology.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;AI is a mirror.&lt;/strong&gt; It reflects the quality of the instructions it receives, the clarity of the context provided, the thoughtfulness of the task assignment. If your AI outputs are mediocre, the first place to look is the inputs.&lt;/p&gt;

&lt;p&gt;This isn’t to say AI tools are perfect—they aren’t. Models have real limitations, and those limitations matter. But the gap between mediocre and excellent AI usage often has more to do with the human side than the machine side.&lt;/p&gt;

&lt;h2&gt;
  
  
  What This Means for Leaders
&lt;/h2&gt;

&lt;p&gt;If you’re a new manager struggling to get value from AI tools, you might actually be struggling with management fundamentals that AI is simply making visible. The fix isn’t better prompting tips—it’s developing the skills you’d need to manage capable knowledge workers effectively. If you’re using the AI to write code, a great management model to think about is that of the product owner. It’s a very special set of skills, as some might say. And NOT AT ALL the set of skills you have trained in as a traditional software engineer.&lt;/p&gt;

&lt;p&gt;If you’re a senior leader wondering why AI investments aren’t paying off, look at your organization’s management culture. Do your managers provide context effectively? Operate at the right level of granularity? Match tasks to capabilities? Create feedback loops? Adapt their approach to different situations and individuals?&lt;/p&gt;

&lt;p&gt;If not, AI tools will underperform—because they require the same inputs that humans require, just in different formats.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The organizations that will excel with AI are the ones that already excel at knowledge work management.&lt;/strong&gt; They understand that capable resources—human or artificial—need direction, not just commands. Context, not just tasks. Iteration, not just deadlines.&lt;/p&gt;

&lt;p&gt;The rest will keep pumping formulaic prompts into their systems and wondering why the results disappoint.&lt;/p&gt;

</description>
      <category>management</category>
      <category>leadership</category>
      <category>ai</category>
    </item>
    <item>
      <title>Software Moats in the Age of AI: What's Actually Defensible?</title>
      <dc:creator>Keith MacKay</dc:creator>
      <pubDate>Mon, 11 May 2026 05:25:37 +0000</pubDate>
      <link>https://forem.com/keithjmackay/software-moats-in-the-age-of-ai-whats-actually-defensible-nnn</link>
      <guid>https://forem.com/keithjmackay/software-moats-in-the-age-of-ai-whats-actually-defensible-nnn</guid>
      <description>&lt;h1&gt;
  
  
  Software Moats in the Age of AI: What's Actually Defensible?
&lt;/h1&gt;

&lt;p&gt;&lt;strong&gt;Why the "AI writes code now" narrative misses the point—and where competitive advantages stubbornly persist&lt;/strong&gt;&lt;/p&gt;




&lt;p&gt;Another pitch deck lands on your desk. Another startup promises to "displace custom software development." Another analyst proclaims coding is now a commodity. Another LinkedIn thought leader announces developers are obsolete.&lt;/p&gt;

&lt;p&gt;You've heard variations for eighteen months. And yet—enterprise software spending hasn't collapsed. Professional services firms keep hiring. COBOL programmers are getting paid (handsomely!).&lt;/p&gt;

&lt;p&gt;Where's the disconnect?&lt;/p&gt;

&lt;h2&gt;
  
  
  The Moat Everyone Forgot
&lt;/h2&gt;

&lt;p&gt;Traditional software moats centered on code: proprietary algorithms, accumulated functionality, engineering discipline, the sheer grind required to replicate something. Two years and $20 million to build something equivalent? That was your moat.&lt;/p&gt;

&lt;p&gt;AI compresses that timeline—for &lt;em&gt;greenfield&lt;/em&gt; development, building from scratch. But most enterprise software is brownfield: decades of accumulated code, undocumented business logic, integration points understood by the one developer that was deeply involved...or a black box written by retired engineers that nobody fully understands. AI hits some hard limits in brownfield. That's our focus here.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Context Problem (Temporary)
&lt;/h2&gt;

&lt;p&gt;AI assistants operate within fixed "context windows"—working memory of roughly 200,000 to 1 million tokens (a token is generally equivalent to about 3/4 of a word). This sounds generous until you realize enterprise systems span millions of lines across thousands of files. Context management is a huge issue at the moment -- EVERYTHING your agent needs as context for the current task (files, prompts, tool instructions, etc.) must be in its context window AT THE SAME TIME...otherwise, the agent doesn't know it exists. And even worse, elements at the beginning and end of the context window are given greater importance (like the recency and primacy effect in human cognition), so the muddled middle stuff can get de-emphasized. The longer a session runs, the worse performance can become because things get thrown away to make room or otherwise lost...a condition sometimes called "context rot". This can REALLY adversely affect performance.&lt;/p&gt;

&lt;p&gt;Now, ask an AI to modify a billing calculation touching seven services, three databases, and two external APIs. You'll get confident suggestions that cheerfully ignore most of the complexity that isn't explicitly called out. The AI doesn't know what it doesn't know—a trait it shares with us all, but it can execute a bad solution faster.&lt;/p&gt;

&lt;p&gt;This context challenge is a problem today, but solutions are coming. Recursive Language Models (RLMs) can now decompose arbitrarily large codebases, analyze pieces, and synthesize understanding across the whole. Within 18-36 months, "codebase too large" stops being a moat. (Note that RLM may sound like an object, but in reality it's a strategy, and one that can be used with any LLM)&lt;/p&gt;

&lt;p&gt;Now, even if the size constraint goes away, understanding &lt;em&gt;what&lt;/em&gt; code does differs from understanding &lt;em&gt;why&lt;/em&gt; it exists. The political archaeology of enterprise systems—why this exception, why that workaround—remains opaque to any model trained on code alone. Nobody documented why the foundation was intentionally built with one crooked wall in 1987.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Implication:&lt;/strong&gt; Codebase scale buys you 2-3 years. Institutional knowledge may buy you longer.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Languages AI Hasn't Learned (Yet)
&lt;/h2&gt;

&lt;p&gt;Large language models learned from public code. They excel at solving problems in Python and JavaScript. They may just hallucinate confidently in COBOL.&lt;/p&gt;

&lt;p&gt;COBOL still processes 95% of ATM transactions and 80% of in-person financial transactions globally. RPG hums along on IBM midrange systems. ABAP powers SAP implementations. These languages lack the public training data AI needs—and ironically, they run the systems with the largest budgets.&lt;/p&gt;

&lt;p&gt;Ask an AI to modify your COBOL billing module. It generates something plausible, but may be introducing subtle bugs you'll discover in production six months later.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Implication:&lt;/strong&gt; Legacy languages create unexpected insulation. Constrained talent pools cut both ways.&lt;/p&gt;

&lt;h2&gt;
  
  
  Domain Knowledge Doesn't Scale
&lt;/h2&gt;

&lt;p&gt;Software that works isn't software that sells. Software that works &lt;em&gt;correctly for a specific domain&lt;/em&gt; sells.&lt;/p&gt;

&lt;p&gt;Healthcare billing involves thousands of payer-specific rules changing quarterly. Energy trading handles physical delivery constraints across jurisdictions. Insurance policy administration encodes actuarial logic accumulated over decades. AI generates code that processes data—not code embodying expertise it never learned.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Implication:&lt;/strong&gt; Vertical software companies with genuine domain expertise have &lt;em&gt;stronger&lt;/em&gt; moats post-AI. The easy parts got easier. Encoding specialized knowledge stayed exactly as hard.&lt;/p&gt;

&lt;h2&gt;
  
  
  Battle Scars Have Value (and Institutional Knowledge Isn't In the Docs)
&lt;/h2&gt;

&lt;p&gt;Enterprise systems connect to ERPs, CRMs, data warehouses, payment processors, and countless internal tools. Each integration point represents hard-won understanding of how systems &lt;em&gt;actually&lt;/em&gt; behave—as opposed to how documentation claims they behave.&lt;/p&gt;

&lt;p&gt;AI writes integration code. It cannot anticipate that upstream systems send malformed JSON on Tuesdays, that authentication tokens expire differently in production, or that the "deprecated" field is actually required for three specific customer configurations. (Every developer reading this just nodded ruefully.)&lt;/p&gt;

&lt;p&gt;This knowledge lives in tribal memory and incident reports. Nobody writes it down because nobody realizes it's unusual until something breaks.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Implication:&lt;/strong&gt; Deep integrations create sticky moats. Competitors must relearn every lesson the hard way.&lt;/p&gt;

&lt;h2&gt;
  
  
  Relationships Beat Tech Chops
&lt;/h2&gt;

&lt;p&gt;CIOs choose vendors they trust to exist in five years who will navigate regulatory inquiries alongside them. A fintech startup with AI-generated code competes against thirty years of relationship and proven reliability. Good luck with that pitch.&lt;/p&gt;

&lt;p&gt;AI-generated code also creates compliance questions nobody has answered. Who's liable when it processes protected health information incorrectly? When AI-built credit decisioning introduces unintended discrimination? Until liability frameworks mature, AI faces adoption friction in precisely the industries with the largest software budgets.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Implication:&lt;/strong&gt; Relationship-intensive businesses in financial services, healthcare, and government hold moats invisible to technical assessments.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Maintenance Cliff
&lt;/h2&gt;

&lt;p&gt;Every line of code creates maintenance obligations. AI-generated code optimizes for working now, not being understood later—producing solutions humans wouldn't choose, using patterns inconsistently, generating technical debt at accelerated rates.&lt;/p&gt;

&lt;p&gt;Organizations rapidly building AI-assisted systems simultaneously build maintenance liabilities. Developers five years from now will struggle to understand logic no human designed. Black boxes that work but nobody understands may come to dominate codebases. How good must AI get before you trust it to fix its own mysteries? Sometimes it's already there. Sometimes we'll always need humans.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Implication:&lt;/strong&gt; Discount AI productivity gains by future maintenance costs. Ignoring accelerated technical debt means overestimating returns.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Pragmatic Bottom Line
&lt;/h2&gt;

&lt;p&gt;AI transforms software development. The competitive landscape for simple ground-up greenfield applications has shifted fundamentally. But enterprise software operates differently. Complexity, integration, domain expertise, relationships, and regulatory dynamics create moats AI hasn't breached.&lt;/p&gt;

&lt;p&gt;These moats aren't permanent. RLMs will erode context advantages. Training data will expand. Regulatory frameworks will mature and become more machine-compatible. But defensible positions today require understanding something deeply, integrating thoroughly, and maintaining relationships that transcend technical capability.&lt;/p&gt;

&lt;p&gt;Those advantages outlast the headlines—and organizations that distinguish temporary moats from durable ones will allocate capital more effectively than those chasing the narrative.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;The question isn't "can AI build this?" It's "can AI build it correctly, integrate it properly, and support it reliably?" The answer determines where you put your money.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>software</category>
      <category>ai</category>
      <category>technicaldebt</category>
      <category>convex</category>
    </item>
    <item>
      <title>The Irony of AI Development: How Context Engineering Is Taking Us Back to Waterfall</title>
      <dc:creator>Keith MacKay</dc:creator>
      <pubDate>Sun, 10 May 2026 22:26:50 +0000</pubDate>
      <link>https://forem.com/keithjmackay/the-irony-of-ai-development-how-context-engineering-is-taking-us-back-to-waterfall-31j3</link>
      <guid>https://forem.com/keithjmackay/the-irony-of-ai-development-how-context-engineering-is-taking-us-back-to-waterfall-31j3</guid>
      <description>&lt;h1&gt;
  
  
  The Irony of AI Development: How Context Engineering Is Taking Us Back to Waterfall
&lt;/h1&gt;

&lt;h2&gt;
  
  
  And Why That's Not Necessarily a Bad Thing
&lt;/h2&gt;

&lt;p&gt;For three decades, the software industry has been on a journey away from waterfall development toward agile methodologies. Now, in an unexpected twist, the rise of AI-powered development tools and "context engineering" is quietly pushing us back toward sequential, specification-heavy workflows.&lt;/p&gt;

&lt;p&gt;But this time, we're walking into a trap we've seen before—the waterbed problem. You must tackle this strategically and head-on in order to recognize AI efficiencies—otherwise AI acceleration will create more chaos than efficiency.&lt;/p&gt;




&lt;h2&gt;
  
  
  A Brief History: From Waterfall to Agile
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The Waterfall Era (1970s-1990s)&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Waterfall development emerged from manufacturing and engineering disciplines. The model was simple: define requirements completely, design the system, build it, test it, deploy it. Each phase flowed into the next like water over a cascade.&lt;/p&gt;

&lt;p&gt;The approach made sense for its time. Computing was expensive. Mistakes were costly. The assumption was that thorough upfront planning would prevent downstream problems.&lt;/p&gt;

&lt;p&gt;It didn't work out that way. Projects routinely ran over budget and behind schedule. By the time software shipped, requirements had changed. The market had moved. A running joke about large enterprise systems was that they were a perfect fit for the company...as of 18 months ago!&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Agile Revolution (2001-2020s)&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The Agile Manifesto was a direct response to waterfall's failures. Its core insight: in complex, uncertain environments, you can't plan your way to success. You must iterate, learn, and adapt.&lt;/p&gt;

&lt;p&gt;Agile shortened feedback loops. Instead of 18-month cycles, teams delivered working software in weeks. Requirements became conversations rather than contracts. Testing happened continuously, not just at the end.&lt;/p&gt;

&lt;p&gt;The results spoke for themselves. Agile teams shipped faster, responded to change better, and produced software that more closely matched what users actually needed.&lt;/p&gt;

&lt;p&gt;Note that there were exceptions where waterfall still made sense, like embedded software that needed to be tested against evolving hardware, or highly regulated industries.&lt;/p&gt;

&lt;p&gt;For the most part, however, for two decades the industry consensus has been clear: agile beats waterfall. Iterate fast. Embrace uncertainty. Deliver incrementally.&lt;/p&gt;




&lt;h2&gt;
  
  
  Enter Context Engineering: The Return of the Specification
&lt;/h2&gt;

&lt;p&gt;Now something interesting is happening.&lt;/p&gt;

&lt;p&gt;The most effective AI-assisted development doesn't look like agile at all. It looks remarkably like waterfall.&lt;/p&gt;

&lt;p&gt;When developers work with large language models like Claude or GPT-4, they quickly discover a pattern: the quality of the output is directly proportional to the quality of the input. Vague prompts produce vague code. Detailed specifications produce useful implementations.&lt;/p&gt;

&lt;p&gt;This has given rise to "context engineering"—the practice of carefully crafting the information, constraints, and examples you provide to AI systems. Context engineering is essentially specification writing for machines.&lt;/p&gt;

&lt;p&gt;The parallels to waterfall are striking:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Upfront investment in specification&lt;/strong&gt;: Before touching code, developers spend significant time writing detailed requirements, examples, and constraints&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Sequential phases&lt;/strong&gt;: Define the context, generate the code, review the output, refine the specification&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Heavy documentation&lt;/strong&gt;: The context window has become the new requirements document&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The irony is profound. After decades of moving away from heavy upfront specification, we're returning to it—not because humans need it, but because AI does.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Waterbed Problem Returns
&lt;/h2&gt;

&lt;p&gt;Here's where things get dangerous.&lt;/p&gt;

&lt;p&gt;In engineering, the "waterbed problem" describes a phenomenon where compressing one part of a system creates pressure elsewhere. Push down on a waterbed here, it bulges up over there. You can't eliminate the complexity; you can only move it around.&lt;/p&gt;

&lt;p&gt;AI development tools are creating exactly this dynamic.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Math Is Merciless&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Consider the numbers that are now being thrown around:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;AI can generate code 10x to 100x faster than manual development&lt;/li&gt;
&lt;li&gt;A single developer can now produce the output of a small team&lt;/li&gt;
&lt;li&gt;Features that took weeks now take hours&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This sounds like pure upside. It isn't.&lt;/p&gt;

&lt;p&gt;If development speed increases 100x, what happens to testing? Does your QA capacity magically scale by 100x? What about code review? Security audits? Documentation? Integration testing?&lt;/p&gt;

&lt;p&gt;The answer, of course, is that you've simply moved the bottleneck.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Where the Pressure Goes&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;When you compress development time through AI, the pressure shows up in predictable places:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Testing&lt;/strong&gt;: AI-generated code requires testing—often more testing than human-written code, because AI systems can produce subtle bugs that humans wouldn't make&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Review&lt;/strong&gt;: Someone still needs to verify that the code does what it should, follows security best practices, integrates properly with existing systems, and provides a clear, useful user experience for its users&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Architecture&lt;/strong&gt;: Faster code generation means architectural decisions come faster, with less time for deliberation&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Requirements&lt;/strong&gt;: If you can implement anything quickly, choosing what to implement becomes the constraint&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Operations&lt;/strong&gt;: More code shipping faster means more deployments, more incidents, more maintenance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;User Absorption&lt;/strong&gt;: Users need to be able to keep up with how to use their software, what features are available, and so forth&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Organizations that accelerate development without accelerating everything else are merely building technical debt at an unprecedented rate. They're pushing on the waterbed.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Whole-Lifecycle Imperative
&lt;/h2&gt;

&lt;p&gt;The lesson is clear: AI tools cannot be applied effectively in isolation. They must be applied across the entire development lifecycle.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;This Is Not Optional&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;If you're using AI to accelerate coding but relying on manual testing, you're setting yourself up for quality disasters. If you're generating code faster but reviewing it at the same pace, defects will slip through. If you're shipping features rapidly but operating infrastructure manually, you'll drown in incidents.&lt;/p&gt;

&lt;p&gt;The math doesn't work any other way.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What Whole-Lifecycle AI Looks Like&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Organizations that successfully navigate this transition are applying AI comprehensively:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;AI-assisted specification&lt;/strong&gt;: Using AI to help write, validate, and refine requirements&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI-accelerated development&lt;/strong&gt;: Code generation, completion, and transformation&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI-powered testing&lt;/strong&gt;: Automated test generation, coverage analysis, and regression detection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI-enhanced review&lt;/strong&gt;: Automated code review, security scanning, and compliance checking&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI-driven operations&lt;/strong&gt;: Incident detection, root cause analysis, and automated remediation&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI-supported architecture&lt;/strong&gt;: Design review, pattern matching, and technical debt detection&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The key insight: the acceleration ratio must be roughly consistent across all phases. If development gets 100x faster, testing needs to get close to 100x faster. Otherwise, testing becomes the bottleneck.&lt;/p&gt;

&lt;p&gt;Your overall throughput is gated by your slowest phase.&lt;/p&gt;




&lt;h2&gt;
  
  
  Strategic Implications for Leaders
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;1. Don't Chase Point Solutions&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The temptation is to start with the most visible opportunity—usually code generation—and optimize later. This is a mistake. Point solutions create imbalances. Imbalances create failures.&lt;/p&gt;

&lt;p&gt;We have seen organizations begin learning how to use AI by implementing it in specific parts of the organization:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;documentation (AI greatly reduces the key-person problem)&lt;/li&gt;
&lt;li&gt;test creation (going from 0 automated tests to comprehensive automated testing, including integration and end-to-end tests, is low-risk, fast, and hugely valuable)&lt;/li&gt;
&lt;li&gt;code review guidance (helping senior engineers more quickly zero in on the biggest challenges and learning opportunities from junior engineers to best use their valuable time)&lt;/li&gt;
&lt;li&gt;tech debt evaluation (reviewing the code base, looking for future challenges)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These strategies each increase quality and provide longer-term value, but they don't radically affect the speed of the software lifecycle, and, once optimized, they don't provide the same ongoing value. These are great mechanisms to leap to a higher level of maturity, but different solutions are required to maintain this new posture going forward.&lt;/p&gt;

&lt;p&gt;A different long-term approach is to start with a comprehensive view of your development lifecycle. Identify every phase where work happens. Map the current throughput of each. Then invest in AI capabilities for each phase.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Measure Throughput, Not Activity&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;It's easy to celebrate when developers report 10x productivity improvements. But developer productivity is not organizational throughput. If testing becomes the bottleneck, you haven't improved throughput—you've just moved work in progress from one queue to another.&lt;/p&gt;

&lt;p&gt;Measure end-to-end cycle time. Measure defect rates. Measure incidents. These metrics tell you whether you're actually moving faster or just generating more chaos.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Rethink Team Structure&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Traditional team structures assumed human-speed development. Ratios of developers to QA engineers, code reviewers to developers, ops engineers to services—all of these were calibrated to pre-AI velocities.&lt;/p&gt;

&lt;p&gt;Those ratios no longer hold. Organizations need to fundamentally reconsider how work is distributed across roles when development velocity changes by an order of magnitude.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Embrace the New Waterfall—Thoughtfully&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Context engineering and specification-heavy development aren't bad. They represent the right way to work with current AI capabilities. The key is to bring the benefits of agile thinking—fast feedback, iteration, continuous integration—to this new paradigm.&lt;/p&gt;

&lt;p&gt;Write specifications, but test them quickly. Generate code, but review it immediately. Ship features, but instrument them comprehensively. The phases may be more sequential than agile purists would like, but the cycles can still be fast.&lt;/p&gt;

&lt;p&gt;And one of the fundamental pillars of agile development -- frequent communication -- still adds tremendous value in context engineering. This communication is both agent-to-agent and human-agent communication via status and spec files and prompts. Frequent human-in-the-loop review is still required at every phase to make sure that systems are behaving as expected, but AI can be used to make sure these reviews are as streamlined and efficient as possible. "Trust but verify" is good policy.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Path Forward
&lt;/h2&gt;

&lt;p&gt;We're at an inflection point. AI tools offer genuine productivity improvements, but they also create genuine risks. The organizations that succeed will be those that:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Recognize that AI acceleration must be applied holistically&lt;/li&gt;
&lt;li&gt;Invest proportionally across the entire development lifecycle&lt;/li&gt;
&lt;li&gt;Measure system throughput rather than local optimization&lt;/li&gt;
&lt;li&gt;Adapt their organizational structures to new velocity assumptions&lt;/li&gt;
&lt;li&gt;Embrace specification-heavy approaches without abandoning fast feedback&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The waterbed problem isn't new. Neither is the tendency to optimize locally while ignoring systemic effects. But the stakes are higher now. AI acceleration is too powerful to apply carelessly.&lt;/p&gt;

&lt;p&gt;The choice isn't whether to adopt AI development tools. That's already inevitable. The choice is whether to adopt them strategically—across the whole lifecycle, in proper proportion, with clear-eyed understanding of the tradeoffs.&lt;/p&gt;

&lt;p&gt;Push on the waterbed intelligently, or watch it bulge in unexpected and costly places.&lt;/p&gt;

</description>
      <category>coding</category>
      <category>watercooler</category>
      <category>architecture</category>
      <category>leadership</category>
    </item>
    <item>
      <title>The AI Bullwhip: What The Beer Game Teaches Us About Uneven AI Adoption</title>
      <dc:creator>Keith MacKay</dc:creator>
      <pubDate>Sun, 10 May 2026 20:20:19 +0000</pubDate>
      <link>https://forem.com/keithjmackay/the-ai-bullwhip-what-the-beer-game-teaches-us-about-uneven-ai-adoption-2k9i</link>
      <guid>https://forem.com/keithjmackay/the-ai-bullwhip-what-the-beer-game-teaches-us-about-uneven-ai-adoption-2k9i</guid>
      <description>&lt;h1&gt;
  
  
  The AI Bullwhip: What The Beer Game Teaches Us About Uneven AI Adoption
&lt;/h1&gt;

&lt;p&gt;&lt;strong&gt;Why introducing AI to one team might break others—and how to avoid the chaos&lt;/strong&gt;&lt;/p&gt;




&lt;p&gt;Several decades ago, I was involved in building a digital version of The Beer Game for HBS, and from its first run the lessons became viscerally clear.&lt;/p&gt;

&lt;p&gt;What is the Beer Game? In 1960, MIT professor Jay Forrester created a deceptively simple simulation that would raise blood pressure for business school students for generations. Four players across a supply chain. Some poker chips representing beer. What could go wrong?&lt;/p&gt;

&lt;p&gt;Everything, it turns out. And sixty-five years later, organizations rushing to adopt AI are relearning the same painful lessons—with considerably higher stakes than simulated beer.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Beer Game: A Five-Minute Primer
&lt;/h2&gt;

&lt;p&gt;If you've never played The Beer Game, here's the setup: four players represent different stages of a beer supply chain—a retailer, a wholesaler, a distributor, and a factory. Each week, customers buy beer from the retailer. Each player can only see their own inventory and incoming orders, not what's happening elsewhere in the chain. There's a time delay between placing orders and receiving shipments.&lt;/p&gt;

&lt;p&gt;The goal seems simple: meet customer demand while minimizing costs from excess inventory or stockouts.&lt;/p&gt;

&lt;p&gt;The result is reliably catastrophic.&lt;/p&gt;

&lt;p&gt;Here's what happens: customer demand increases slightly—say, from four cases per week to eight. The retailer notices shelves emptying and orders more from the wholesaler. But shipments take time, so the shelves keep emptying. Panicking, the retailer orders even more. The wholesaler, now seeing a surge in orders, assumes demand is exploding and orders aggressively from the distributor. The distributor does the same to the factory. The factory ramps up production dramatically.&lt;/p&gt;

&lt;p&gt;Then the delayed shipments start arriving. Everywhere. All at once.&lt;/p&gt;

&lt;p&gt;Suddenly everyone is drowning in beer. The retailer stops ordering. The wholesaler, still receiving massive shipments, stops ordering. The distributor is buried. The factory has just finished a production run for demand that evaporated weeks ago. And beer begins to go stale in storage (which, to my collegiate colleagues, was a particularly egregious outcome).&lt;/p&gt;

&lt;p&gt;This is the &lt;strong&gt;bullwhip effect&lt;/strong&gt;: small fluctuations at the customer end create massive, destructive oscillations upstream. A 10% increase in consumer demand can translate to 40% swings at the factory. Careers are ruined. Simulated beer is wasted. Business school students stare at their inventory sheets in disbelief.&lt;/p&gt;

&lt;p&gt;The culprit isn't stupidity. Every player makes locally rational decisions. The problem is &lt;strong&gt;systemic&lt;/strong&gt;: limited visibility, time delays, and independent decision-making combine to amplify rather than dampen disruptions.&lt;/p&gt;

&lt;h2&gt;
  
  
  Now Replace "Beer" with "AI Productivity"
&lt;/h2&gt;

&lt;p&gt;Organizations introducing AI tools are playing their own version of The Beer Game—and most don't realize it.&lt;/p&gt;

&lt;p&gt;Consider a typical scenario: a development team adopts AI coding assistants. Productivity jumps. Code flows faster. Features that took weeks now take days. The team lead reports the wins. Leadership notices.&lt;/p&gt;

&lt;p&gt;But no one downstream adjusted.&lt;/p&gt;

&lt;p&gt;The QA team still has the same headcount. The same testing processes. The same throughput. Suddenly they're facing a tsunami of code. Defect backlogs balloon. Test coverage drops as testers scramble to keep pace. Quality issues slip into production.&lt;/p&gt;

&lt;p&gt;Meanwhile, upstream teams notice something strange: requirements that used to take the dev team three sprints now complete in one. Product managers haven't recalibrated how much work to queue up. The backlog empties unexpectedly. Roadmap meetings get chaotic. "We need more features defined!" becomes the cry—but the product team is still operating at their old cadence.&lt;/p&gt;

&lt;p&gt;The QA/Testing team has more tests to write, more features to evaluate. Often under-sized to begin with, they are swamped. With predictable quality results.&lt;/p&gt;

&lt;p&gt;The DevOps team, accustomed to a predictable deployment rhythm, now sees triple the deployment requests. CI/CD pipelines bottleneck. Infrastructure provisioning can't keep pace. Developers who were flying now sit waiting for environments.&lt;/p&gt;

&lt;p&gt;Each team is making locally rational decisions. Each team is overwhelmed or starved for reasons they can't quite see. The bullwhip cracks. (If this all feel familiar, in software circles this is sometimes also referred to as "the waterbed problem", and I wrote about it last week in those terms when talking about how AI is bringing us back to waterfall development)&lt;/p&gt;

&lt;h2&gt;
  
  
  How Organizations Are Approaching AI Adoption
&lt;/h2&gt;

&lt;p&gt;Most organizations fall into one of three patterns when introducing AI development tools:&lt;/p&gt;

&lt;h3&gt;
  
  
  The Piecemeal Pioneers
&lt;/h3&gt;

&lt;p&gt;The most common approach: individual teams or developers adopt AI tools organically. Someone tries GitHub Copilot. A team experiments with Claude Code. Results vary. Successes spread through word of mouth. There's no coordinated rollout, no systemic adjustment.&lt;/p&gt;

&lt;p&gt;This is The Beer Game with each player ordering independently, without coordination.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Mandate Push
&lt;/h3&gt;

&lt;p&gt;Leadership declares AI adoption a strategic priority. Tools are procured. Training is scheduled. Metrics are established. The development organization gets AI capabilities—often simultaneously.&lt;/p&gt;

&lt;p&gt;But adjacent functions don't. QA, product, DevOps, security review, documentation—they're still operating traditionally while development adopts new strategies. The mandate created a step function in (only) one part of the value stream.&lt;/p&gt;

&lt;p&gt;This is like one Beer Game player getting instant teleportation while everyone else still waits for truck deliveries.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Thoughtful Rollout
&lt;/h3&gt;

&lt;p&gt;Rare but effective: organizations that map their entire value stream before introducing acceleration. They ask: if development velocity triples, what breaks? Where do bottlenecks emerge? Which handoffs become flood points?&lt;/p&gt;

&lt;p&gt;Then they stage adoption to match capacity across the chain.&lt;/p&gt;

&lt;p&gt;This is the only approach that avoids the bullwhip—and almost nobody does it.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Bullwhip Effects of Uneven AI Adoption
&lt;/h2&gt;

&lt;p&gt;Let's map the specific oscillations that emerge when AI productivity hits an unprepared organization:&lt;/p&gt;

&lt;h3&gt;
  
  
  The Quality Whiplash
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Upstream acceleration:&lt;/strong&gt; Dev team ships code 3x faster with AI assistance.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Downstream bottleneck:&lt;/strong&gt; QA capacity unchanged.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Oscillation pattern:&lt;/strong&gt; Quality team rushes reviews → defects escape → production incidents spike → emergency slowdowns → dev team idles waiting for fixes → QA catches up → dev accelerates again → cycle repeats.&lt;/p&gt;

&lt;p&gt;Organizations stuck in this loop often conclude "AI is causing quality problems." The AI isn't causing anything—the uneven adoption is.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Requirements Vacuum
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Upstream bottleneck:&lt;/strong&gt; Product team defines work at traditional pace.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Downstream acceleration:&lt;/strong&gt; Dev team consumes requirements faster than they're created.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Oscillation pattern:&lt;/strong&gt; Backlog empties → devs pull partially-formed work → rework increases → devs slow down → backlog fills again → devs accelerate on clear requirements → backlog empties → cycle repeats.&lt;/p&gt;

&lt;p&gt;Teams trapped here often see erratic velocity charts and blame "unclear requirements." The requirements aren't less clear—they're just not flowing fast enough.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Deployment Gridlock
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Upstream acceleration:&lt;/strong&gt; More code, more features, more changes.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Downstream bottleneck:&lt;/strong&gt; Same CI/CD capacity, same deployment windows, same ops team.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Oscillation pattern:&lt;/strong&gt; Deployment queue grows → batching increases → batch sizes create risk → releases get delayed → pressure builds → risky big-bang release → incidents → release freezes → queue grows again.&lt;/p&gt;

&lt;p&gt;This pattern often ends with someone suggesting "maybe we should slow down development"—treating the symptom rather than the system.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Security Squeeze
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Upstream acceleration:&lt;/strong&gt; More code surface area, faster.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Downstream bottleneck:&lt;/strong&gt; Security review capacity fixed.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Oscillation pattern:&lt;/strong&gt; Security backlog grows → reviews become perfunctory → vulnerabilities ship → incident occurs → security becomes blocker → development halts for remediation → security catches up → development accelerates → security backlog grows.&lt;/p&gt;

&lt;p&gt;The security team isn't being obstructionist. They're being bullwhipped.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Compounding Problem
&lt;/h2&gt;

&lt;p&gt;What makes AI adoption particularly treacherous is that these oscillations compound.&lt;/p&gt;

&lt;p&gt;In The Beer Game, there's one supply chain with one bullwhip. In software development, there are multiple parallel flows—and they interact. A quality slowdown affects deployment timing. A deployment bottleneck affects security review scheduling. A security delay affects requirements prioritization.&lt;/p&gt;

&lt;p&gt;Introduce AI acceleration unevenly, and you don't get one bullwhip—you get several, out of phase, amplifying each other in unpredictable ways.&lt;/p&gt;

&lt;p&gt;The organization experiences this as chaos, politics, and blame. "The dev team is cowboying." "QA is a bottleneck." "Product can't get their act together." "DevOps is always blocking us."&lt;/p&gt;

&lt;p&gt;Nobody sees the system. Everyone sees their adjacent node failing them.&lt;/p&gt;

&lt;h2&gt;
  
  
  Planning to Avoid the Whip
&lt;/h2&gt;

&lt;p&gt;The good news: The Beer Game has a solution. It's called &lt;strong&gt;information sharing and coordinated decision-making&lt;/strong&gt;. When all players can see the entire supply chain and coordinate their orders, the bullwhip disappears.&lt;/p&gt;

&lt;p&gt;The same principle applies to AI adoption:&lt;/p&gt;

&lt;h3&gt;
  
  
  Map Before You Accelerate
&lt;/h3&gt;

&lt;p&gt;Before introducing AI to any team, map your value stream end-to-end. Identify every handoff. Measure current throughput at each stage. Find existing bottlenecks (you probably have some already).&lt;/p&gt;

&lt;p&gt;Then ask: if we 2x this stage, what happens to the stage immediately downstream? What about two stages down?&lt;/p&gt;

&lt;h3&gt;
  
  
  Accelerate Bottlenecks First
&lt;/h3&gt;

&lt;p&gt;Counterintuitively, the best place to introduce AI might not be where you'll see the biggest individual productivity gain—it's where you'll relieve the biggest systemic constraint.&lt;/p&gt;

&lt;p&gt;If QA is already struggling to keep pace, accelerating development is pouring water into a backed-up drain. Consider AI-assisted testing tools first. Or semi-automated code review (so senior engineers can focus on the right quality elements and teaching opportunities with less review time). Or AI-enhanced security scanning.&lt;/p&gt;

&lt;p&gt;Match AI adoption to system topology, not team enthusiasm.&lt;/p&gt;

&lt;h3&gt;
  
  
  Build Slack Intentionally
&lt;/h3&gt;

&lt;p&gt;The Beer Game punishes systems with no buffer capacity. When everyone operates at maximum efficiency, there's no room to absorb variation.&lt;/p&gt;

&lt;p&gt;As you introduce AI acceleration, deliberately create slack in adjacent functions. That might mean additional headcount. It might mean reduced WIP limits. It might mean explicit buffers between stages.&lt;/p&gt;

&lt;p&gt;Yes, slack feels inefficient. It's also what prevents oscillation from becoming catastrophe.&lt;/p&gt;

&lt;h3&gt;
  
  
  Make the System Visible
&lt;/h3&gt;

&lt;p&gt;The Beer Game's dysfunction persists because players can't see beyond their immediate neighbors. Create visibility across your development value stream:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;End-to-end cycle time dashboards&lt;/li&gt;
&lt;li&gt;WIP at each stage, visible to all&lt;/li&gt;
&lt;li&gt;Bottleneck indicators that surface automatically&lt;/li&gt;
&lt;li&gt;Regular cross-functional reviews of flow&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;When everyone can see the whole chain, locally rational decisions become globally rational decisions.&lt;/p&gt;

&lt;h3&gt;
  
  
  Stage Your Rollout
&lt;/h3&gt;

&lt;p&gt;If you must introduce AI capability unevenly (and you probably will—budgets and readiness vary), stage it deliberately:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Start with the current bottleneck&lt;/li&gt;
&lt;li&gt;Wait for throughput to stabilize&lt;/li&gt;
&lt;li&gt;Identify the new bottleneck&lt;/li&gt;
&lt;li&gt;Introduce AI there&lt;/li&gt;
&lt;li&gt;Repeat&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This is slower than a simultaneous rollout. It's also far less likely to create destructive oscillation.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Meta-Lesson
&lt;/h2&gt;

&lt;p&gt;The Beer Game has taught a consistent lesson for sixty-five years: &lt;strong&gt;optimizing parts degrades wholes&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;AI tools offer genuine, dramatic acceleration. They also offer the ability to create genuine, dramatic dysfunction if deployed without systemic thinking.&lt;/p&gt;

&lt;p&gt;The organizations that will succeed with AI aren't the ones that adopt fastest. They're the ones that adopt most coherently—matching capability to capacity across their entire value stream.&lt;/p&gt;

&lt;p&gt;Every team is connected to every other team. Accelerate one without adjusting the others, and you're not improving the system—you're just moving the bottleneck, amplifying the oscillation, and cracking the bullwhip.&lt;/p&gt;

&lt;p&gt;The beer, it turns out, was a metaphor all along.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>systemdesign</category>
      <category>management</category>
      <category>leadership</category>
    </item>
    <item>
      <title>We're Linear Thinkers in an Exponentially-Changing World</title>
      <dc:creator>Keith MacKay</dc:creator>
      <pubDate>Sun, 10 May 2026 18:00:54 +0000</pubDate>
      <link>https://forem.com/keithjmackay/were-linear-thinkers-in-an-exponentially-changing-world-53jc</link>
      <guid>https://forem.com/keithjmackay/were-linear-thinkers-in-an-exponentially-changing-world-53jc</guid>
      <description>&lt;h1&gt;
  
  
  We're Linear Thinkers in an Exponentially-Changing World
&lt;/h1&gt;

&lt;p&gt;&lt;strong&gt;The more time I spend in the AI ecosystem, the more convinced I become that the pace of change isn’t just fast—it's&lt;/strong&gt; &lt;strong&gt;&lt;em&gt;explosive&lt;/em&gt;&lt;/strong&gt;&lt;strong&gt;…and increasingly so. Most people still think in terms of linear change while the world is accelerating exponentially.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;That mismatch is where disruption happens.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;We’re linear thinkers living in an exponential world&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;We weren’t built to intuit compounding curves. It’s why exponential progress feels like it comes out of nowhere.&lt;/p&gt;

&lt;p&gt;Most AI charts are logarithmic for a reason: they’re trying to compress the 10X-after-10X reality into something our brains can process. Turn those accelerations we don't process into nice straight lines.&lt;/p&gt;

&lt;p&gt;And this is why the fast eat the slow. By the time a large company finishes planning, the curve has already bent again.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;I first learned this over 20 years ago—Kurzweil rewired my thinking&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;At a Ray Kurzweil talk at MIT in the mid-2000s, he described the “Law of Accelerating Returns.” It permanently rewired how I think about the pace of technology change.&lt;/p&gt;

&lt;p&gt;He wasn’t the first to notice this pattern:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Henry Adams&lt;/strong&gt; – law of acceleration&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Buckminster Fuller&lt;/strong&gt; – ephemeralization (doing more and more with less and less)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Moore’s Law&lt;/strong&gt; – exponential chip complexity&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Hans Moravec&lt;/strong&gt; – robotics advancing at Moore’s-law speed&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Kurzweil unified and expanded these ideas to encompass all technology and evolution—and even applied them to his business strategy. He claimed that he began designing products years ahead of time, and calculating &lt;em&gt;when&lt;/em&gt; the enabling technologies would be available. He developed a photocopier-sized text scanner for the blind in the 70s, and began designing his handheld version of the same soon afterwards. He was in production within months of when the enabling tech was finally small and performant enough to work for the product. Is it true? I don’t know – but the fact that it would be a striking example of true exponential thinking demonstrates how rarely it occurs.&lt;/p&gt;




&lt;p&gt;Critics point to Kurzweil’s dates being off in some cases by a decade. But I would argue that in most cases, the limiting factor wasn’t technological capability—it was political will, regulation, or distribution. Exponential growth of technical capability is the rule rather than the exception. Rapid adoption of what's possible is the exception rather than the rule.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;As William Gibson said: &lt;strong&gt;“The future is already here—it’s just not evenly distributed.”&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;It's never been more true. An example would be that Kurzweil's predictions for 2009 (made in 1999!) included “self-driving cars”. While not available to consumers in 2009, Google had, in fact, successfully logged over 200,000 miles with their self-driving technology by 2009, and Nevada was putting self-driving vehicle laws on the books. That future wasn't yet evenly distributed.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;Today’s AI acceleration makes those early curves look quaint&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Consider just a few signals as outlined in the Stanford AI Index Report 2025[1]:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Hardware costs have been dropping &lt;strong&gt;30% per year&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;Energy efficiency has been improving &lt;strong&gt;40% per year&lt;/strong&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And then on the software and training sides of the equation:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Google’s Gemini 3.1 showed that models can still gain intelligence through smarter training—not just more parameters as the recent trend had been (combine smarter training with more parameters, and there's no reason for the capability curves to flatten out yet)&lt;/li&gt;
&lt;li&gt;Breakthroughs (like Mixture-of-Experts, 1-bit quantization, the tabular foundation model, etc.) emerge constantly&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And finally, practitioners are learning new ways to get the most from the tools, and increasing efficiency while reducing costs:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;I code with AI &lt;em&gt;completely differently&lt;/em&gt; than I did a few months ago.&lt;/li&gt;
&lt;li&gt;And I expect the same a few months from now.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Every layer of the ecosystem is growing. And doing so faster.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;10X improvement per year changes everything&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Most of my clients invest with a 3–7 year horizon. If AI capability continues at the ~10X per year that we’ve seen since at least GPT2:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;3 years → &lt;strong&gt;1,000X more capable&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;7 years → &lt;strong&gt;10,000,000X more capable&lt;/strong&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Even “only” 5 more years of this curve gives you &lt;strong&gt;100,000X&lt;/strong&gt; improvement.&lt;/p&gt;

&lt;p&gt;Put simply: &lt;strong&gt;software moats have largely evaporated over the past year.&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;In a recent diligence project, I replicated ~80% of the target’s product in a single weekend (in spare time) for something like $60 of Claude Code time. When the target wrote the product two years ago, it was ground-breaking. Now it was a weekend’s part-time work.&lt;/li&gt;
&lt;li&gt;Another colleague did something similar on a recent project.&lt;/li&gt;
&lt;li&gt;We’re both experienced software developers, and deeply ensconced in the context engineering rabbit holes. But a motivated junior engineer could very likely do this in under a week.&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;&lt;strong&gt;It's never technology, always psychology&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;My colleagues and I have successfully used context engineering principles and the latest generation of AI coding tools and LLMs to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;rebuild significant legacy systems into modern stacks with full test infrastructure and mature coding practices&lt;/li&gt;
&lt;li&gt;build greenfield apps at unbelievable speed with legit UI frontend work&lt;/li&gt;
&lt;li&gt;create agents, skills, commands, and developer workflow tools to further accelerate our own work with these tools&lt;/li&gt;
&lt;li&gt;analyze legacy codebases and plan monolith decomposition and modularization&lt;/li&gt;
&lt;li&gt;read, analyze, and fix bugs in open source projects&lt;/li&gt;
&lt;li&gt;develop documentation and visualization for codebases, with no prior exposure to the code&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The hardest part?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Change management.&lt;/strong&gt; Humans can’t mentally accelerate at the same rate as the tools.&lt;/p&gt;

&lt;p&gt;Rote tasks? AI is already there. That future just isn’t yet evenly distributed. Creative tasks? They’re next.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;So what should you do?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Three things matter more than ever:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Master the tools&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Stay flexible and experiment constantly&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Build moats around relationships, distribution, and trust—not code&lt;/strong&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Because the curve is still bending upward. And it’s bending faster than most people realize.&lt;/p&gt;




&lt;p&gt;[1] Nestor Maslej, Loredana Fattorini, Raymond Perrault, Yolanda Gil, Vanessa Parli, Njenga Kariuki, Emily Capstick, Anka Reuel, Erik Brynjolfsson, John Etchemendy, Katrina Ligett, Terah Lyons, James Manyika, Juan Carlos Niebles, Yoav Shoham, Russell Wald, Toby Walsh, Armin Hamrah, Lapo Santarlasci, Julia Betts Lotufo, Alexandra Rome, Andrew Shi, Sukrut Oak. “The AI Index 2025 Annual Report,” AI Index Steering Committee, Institute for Human-Centered AI, Stanford University, Stanford, CA, April 2025. &lt;a href="https://doi.org/10.48550/arXiv.2504.07139" rel="noopener noreferrer"&gt;https://doi.org/10.48550/arXiv.2504.07139&lt;/a&gt;&lt;/p&gt;

</description>
      <category>aistrategy</category>
      <category>agents</category>
      <category>productivity</category>
      <category>cognitiveload</category>
    </item>
  </channel>
</rss>
