<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: John Lee</title>
    <description>The latest articles on Forem by John Lee (@johnonlee).</description>
    <link>https://forem.com/johnonlee</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3924610%2Fe3db9d16-677c-4971-9de6-071496991f48.jpeg</url>
      <title>Forem: John Lee</title>
      <link>https://forem.com/johnonlee</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/johnonlee"/>
    <language>en</language>
    <item>
      <title>How Are You Managing Your AI's Context Window?</title>
      <dc:creator>John Lee</dc:creator>
      <pubDate>Mon, 11 May 2026 12:34:28 +0000</pubDate>
      <link>https://forem.com/johnonlee/how-are-you-managing-your-ais-context-window-324g</link>
      <guid>https://forem.com/johnonlee/how-are-you-managing-your-ais-context-window-324g</guid>
      <description>&lt;p&gt;Your AI coding agent has a 200K token context window. Maybe 500K. Maybe a million.&lt;/p&gt;

&lt;p&gt;So... what actually changed?&lt;/p&gt;

&lt;p&gt;Honestly, I'm still figuring that out. I expected bigger windows to deliver better results. The reality has been more nuanced.&lt;/p&gt;




&lt;h2&gt;
  
  
  1. The Window Got Bigger. Did Anything Actually Change?
&lt;/h2&gt;

&lt;p&gt;The narrative is seductive: "200K tokens! I can dump my entire codebase in there." "1M tokens? Every issue, every doc, every chat log."&lt;/p&gt;

&lt;p&gt;This is like saying "my hard drive is 2TB, so I'll keep every file on my desktop." Technically possible. But do you actually do that?&lt;/p&gt;

&lt;p&gt;Research consistently shows that as context windows grow, retrieval accuracy degrades. The "lost in the middle" problem is real — AI pays most attention to the beginning and end, and everything in between fades. Bigger haystacks make needles harder to find.&lt;/p&gt;

&lt;p&gt;But here's what I find more interesting: &lt;strong&gt;how are we actually using these bigger windows?&lt;/strong&gt; Model spec comparisons are easy. "200K vs 1M" is a number you can compare. But "how well am I managing my context" has no number. It's invisible. So nobody looks at it.&lt;/p&gt;




&lt;h2&gt;
  
  
  2. What Actually Happens Inside a Claude Code Session
&lt;/h2&gt;

&lt;p&gt;Here's what I've observed over a few months of using Claude Code with my team. No quantified data — just experiential patterns. If you've done actual measurement on this, I'd honestly love to hear about it.&lt;/p&gt;

&lt;p&gt;A typical session has this rhythm:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Context gathering eats up a surprising amount of time.&lt;/strong&gt; Reading issues. Scanning docs. Exploring the codebase to figure out what's what. It repeats at the start of every session.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Re-verification is weirdly common.&lt;/strong&gt; My Claude discovers something. Tomorrow, my Claude (or my teammate's Claude) re-discovers the same thing. Not because the AI isn't capable. Because the AIs don't share memory.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Actual problem solving gets less time than you'd think.&lt;/strong&gt; After the first two phases, you finally get to the work you opened the session for.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Here's what matters: this isn't waste because the AI isn't smart enough. It's waste because &lt;strong&gt;the AIs don't share what they know&lt;/strong&gt;. We've built incredible systems for CI/CD, code review, documentation. But when it comes to how our AI agents share knowledge as a team? Almost nothing.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What about your team?&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  3. Three Patterns I Keep Seeing
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. The Dump Truck
&lt;/h3&gt;

&lt;p&gt;"I have 200K tokens. Here's every file in the repo, 47 issues, the company handbook. Go."&lt;/p&gt;

&lt;p&gt;I get it. You don't know what's relevant ahead of time. The temptation to "just put everything in" is real.&lt;/p&gt;

&lt;p&gt;But then your AI is reasoning against mostly irrelevant context. Finding patterns in noise. Confidently proposing solutions to problems you don't have. Unnecessary noise eventually eats away at reasoning quality.&lt;/p&gt;

&lt;p&gt;I did this early on. Still catch myself doing it. I haven't found a perfect solution — but just being aware of the pattern has helped.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Groundhog Day
&lt;/h3&gt;

&lt;p&gt;"Our project uses pnpm workspaces. Auth is in &lt;code&gt;packages/auth&lt;/code&gt;. Don't touch &lt;code&gt;legacy/&lt;/code&gt;. Alice owns deployments."&lt;/p&gt;

&lt;p&gt;Your human colleagues learned this on day one. Your AI has to re-learn it every single session.&lt;/p&gt;

&lt;p&gt;If a human teammate asked you to re-explain the project structure every morning before they could start working, you'd have a serious conversation. But we accept this from AI without question. Why haven't we automated this repetition away yet?&lt;/p&gt;

&lt;h3&gt;
  
  
  3. The Genius Silo
&lt;/h3&gt;

&lt;p&gt;This is the most fascinating one. And the most unsettling.&lt;/p&gt;

&lt;p&gt;Same Claude model. Wildly different outcomes. When a senior engineer who knows the product's bones by heart picks up Claude, the AI becomes a "genius." The codebase's history, known landmines, unwritten conventions — all this invisible context dissolves into the AI's reasoning. Sessions are fast, almost magical.&lt;/p&gt;

&lt;p&gt;When a junior engineer with less context picks up the exact same Claude, they come back empty-handed. Their Claude re-discovers, from scratch, what the senior's Claude figured out months ago. Burns tokens. Burns time. Builds frustration.&lt;/p&gt;

&lt;p&gt;Here's what this means: AI, as a tool, isn't lifting the team's collective productivity. It's &lt;strong&gt;trapped in individual silos of personal experience&lt;/strong&gt;. The senior gets faster and faster. The junior stays stuck. Claude has become a personal assistant, not a team tool.&lt;/p&gt;

&lt;p&gt;And the team lead sees none of this. Doesn't know what the senior's Claude knows. Doesn't know what the junior's Claude is painfully re-learning. &lt;strong&gt;These invisible walls are completely hidden.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Is this happening on your team too? Or have you found a different way?&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  4. What I've Been Trying (Hypothesis Stage)
&lt;/h2&gt;

&lt;p&gt;After months of experimenting, I've roughly settled on four principles. These are working hypotheses — if you've found better approaches, I genuinely want to hear them.&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Relevance Over Volume
&lt;/h3&gt;

&lt;p&gt;I stopped asking "how much can I fit?" and started asking "what actually matters right now?"&lt;/p&gt;

&lt;p&gt;A small, well-curated context beats a massive dump. I'm convinced of this through experience. What "well-curated" actually means in practice — still experimenting.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Persistence Over Repetition
&lt;/h3&gt;

&lt;p&gt;When my AI discovers something valuable — a pattern, a gotcha, an insight — I try not to let it die with the session.&lt;/p&gt;

&lt;p&gt;At the end of each Claude Code session, I ask myself: "What did my Claude learn today that my teammate's Claude should know tomorrow?" It's not perfect, but it has saved the opening minutes of my next session more times than I can count.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Domain Sync
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Transplanting the senior engineer's business context into the AI's baseline assumptions.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;When a senior tells their Claude "this component is perf-critical, O(n²) won't fly," that judgment has months of domain knowledge baked into it. Domain Sync is about making that knowledge accessible to every teammate's Claude.&lt;/p&gt;

&lt;p&gt;It's about converting individual expertise into the team's prompt assets. How far this can be automated — I don't know yet. But the direction feels right.&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Routinized Results Verification
&lt;/h3&gt;

&lt;p&gt;Not blindly trusting the AI's output. Systematically filtering it through past incidents and accumulated history.&lt;/p&gt;

&lt;p&gt;A senior developer, reviewing Claude's code, unconsciously checks: "We had a similar PR that broke tests last time..." "This pattern looks like the one that caused the outage last year..." This filtering instinct — knowing how to &lt;em&gt;reject&lt;/em&gt; well — is what truly separates seniors from juniors.&lt;/p&gt;

&lt;p&gt;The problem: this filtering instinct has remained private, tacit knowledge. How do we turn "knowing how to ask well" into "knowing how to filter well" — and make that filtering instinct a baseline routine for every team Claude? This is what I'm most preoccupied with lately.&lt;/p&gt;




&lt;h2&gt;
  
  
  5. What I Actually Want to Know — Let's Think Together
&lt;/h2&gt;

&lt;p&gt;With context windows exploding toward infinity, are we falling into the quantity trap while losing sight of quality?&lt;/p&gt;

&lt;p&gt;What actually determines real-world productivity isn't benchmark scores. It's &lt;strong&gt;the quality of context optimized for your specific product&lt;/strong&gt;. But that doesn't show up on any benchmark. So nobody looks at it. So I'm asking.&lt;/p&gt;

&lt;h3&gt;
  
  
  Four Questions
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;1. Experience Replication:&lt;/strong&gt;&lt;br&gt;
Is your senior engineer's AI know-how and business context being transferred to other team members — or is it trapped inside individual chat windows? How many Genius Silos exist on your team?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. The Noise Paradox:&lt;/strong&gt;&lt;br&gt;
As windows grow bigger, AI paradoxically loses the plot (Lost in the Middle). What filtering are you doing to counter this? Not just "use less context" — are there smarter ways to structure it?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Knowledge Expiration:&lt;/strong&gt;&lt;br&gt;
In the "store it and forget it" pile, is stale, contaminated context quietly poisoning your AI's judgment? Is last year's "never touch &lt;code&gt;legacy/&lt;/code&gt;" silently overriding this year's "migration complete, it's safe now"?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Building the Team Brain:&lt;/strong&gt;&lt;br&gt;
Is your team's AI getting smarter over time — or stuck in an endless loop of Groundhog Day explanations? Do you have any way to tell?&lt;/p&gt;




&lt;h2&gt;
  
  
  6. Closing
&lt;/h2&gt;

&lt;p&gt;I've been staring at this problem for months. Building tools. Running experiments with my team. But I don't have the answers. I'm still experimenting.&lt;/p&gt;

&lt;p&gt;So I'm asking: &lt;strong&gt;how are you managing your AI coding agent's context window?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Ideas? I want them. Disagreements? Even better. If your experience is "dumping everything into a big window works fine for us," I genuinely want to hear about that too. Let's figure this out together.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;All examples and scenarios in this post are based on real experiences, adapted for the blog format.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>productivity</category>
      <category>programming</category>
      <category>claude</category>
    </item>
  </channel>
</rss>
