<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Harry Floyd</title>
    <description>The latest articles on Forem by Harry Floyd (@harryfloyd).</description>
    <link>https://forem.com/harryfloyd</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3933548%2Fa644e757-fdc0-4213-a2d0-37774cbe6730.png</url>
      <title>Forem: Harry Floyd</title>
      <link>https://forem.com/harryfloyd</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/harryfloyd"/>
    <language>en</language>
    <item>
      <title>NVIDIA Q1 FY2027 Earnings Preview — 5 Signals the Market May Be Missing</title>
      <dc:creator>Harry Floyd</dc:creator>
      <pubDate>Sun, 17 May 2026 11:17:20 +0000</pubDate>
      <link>https://forem.com/harryfloyd/nvidia-q1-fy2027-earnings-preview-5-signals-the-market-may-be-missing-2j79</link>
      <guid>https://forem.com/harryfloyd/nvidia-q1-fy2027-earnings-preview-5-signals-the-market-may-be-missing-2j79</guid>
      <description>&lt;p&gt;NVIDIA reports Q1 FY2027 earnings on &lt;strong&gt;May 20, 2026&lt;/strong&gt;, after market close. The consensus expects approximately &lt;strong&gt;$78.1-78.8 billion in revenue and $1.74 EPS&lt;/strong&gt;, with Citi and Wells Fargo running slightly higher at ~$80B and $1.79 respectively.&lt;/p&gt;

&lt;p&gt;The stock closed at ~$225 on the most recent trading day, roughly a 27x forward P/E on FY2027 estimates. This is not a distressed entry point. It is a thesis-testing moment.&lt;/p&gt;

&lt;p&gt;Below are 5 signals to watch that go beyond the headline beat-or-miss narrative. Each maps to a specific structural claim about NVIDIA's position in the AI infrastructure stack, and each has a falsification trigger that would challenge that claim.&lt;/p&gt;




&lt;h2&gt;
  
  
  1. Purchase Commitments: Are They Still Rising?
&lt;/h2&gt;

&lt;p&gt;NVIDIA's supply-related purchase commitments rose from &lt;strong&gt;$50.3 billion to $95.2 billion&lt;/strong&gt; between Q3 and Q4 FY2026, nearly doubling in a single quarter. This is not optional inventory building. It is NVIDIA aggressively locking in component supply for constraints it believes are structural.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What to watch:&lt;/strong&gt; If commitments rise further in Q1, NVIDIA is deepening its supply chain lock-in through at least 2027. If they flatten, either supply is easing (bullish for margins) or suppliers have hit allocation limits (bearish for revenue growth).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Falsification trigger:&lt;/strong&gt; A flat or declining commitment trajectory would suggest NVIDIA sees peak demand behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  2. Optical Interconnect Mentions
&lt;/h2&gt;

&lt;p&gt;On May 6, NVIDIA announced a &lt;strong&gt;$500 million partnership with Corning (GLW)&lt;/strong&gt; — three new US optical factories, 10x capacity increase, and a warrant structure giving NVIDIA up to 15 million shares at $180. This follows &lt;strong&gt;$2B+ purchase commitments each to Lumentum (LITE)&lt;/strong&gt; (which reported $808M revenue, +90% YoY on May 5) and &lt;strong&gt;Coherent (COHR)&lt;/strong&gt; ($1.81B, +21% YoY on May 6).&lt;/p&gt;

&lt;p&gt;Combined, over &lt;strong&gt;$4.7B+ was committed to the optical supply chain in 10 weeks&lt;/strong&gt; across three independent layers: passive fiber (Corning), active components (LITE, COHR), and co-packaged optics (Ayar Labs, ~$155M NVIDIA portion).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What to watch:&lt;/strong&gt; If management references Corning, Lumentum, or Coherent by name on the call, it validates the thesis that optical interconnect is the next binding constraint beyond HBM memory. Silence on optical supply is a missed signal.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Falsification trigger:&lt;/strong&gt; If optical supply is described as "secured" or "no longer a concern," the bottleneck thesis for the photonics layer weakens meaningfully.&lt;/p&gt;




&lt;h2&gt;
  
  
  3. Blackwell Margin Trajectory
&lt;/h2&gt;

&lt;p&gt;The Street is fixated on gross margins during the Blackwell ramp. The concern is that the more complex B200/B300 packaging (CoWoS-L) compresses margins compared to the simpler H100/H200 designs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What to watch:&lt;/strong&gt; The directional trajectory matters more than the absolute number. Sequential margin expansion indicates the ramp is absorbing complexity costs. Compression suggests the packaging premium is permanent.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The deeper signal:&lt;/strong&gt; NVIDIA's gross margin has been the most-watched metric for 8 consecutive quarters. The market has already priced in margin compression, so an in-line or better margin print removes a major overhang. A miss amplifies the ASIC-competition narrative.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Context:&lt;/strong&gt; NVIDIA pre-booked approximately 60% of TSMC's total 2026 CoWoS output (per Morgan Stanley), with demand of ~700,000 wafers. CoWoS is the packaging constraint. If margins hold despite this capacity scramble, the GPU economics thesis is intact.&lt;/p&gt;




&lt;h2&gt;
  
  
  4. Inference Mix and Agent Workload Commentary
&lt;/h2&gt;

&lt;p&gt;The AI market narrative shifted in 2026 from "training is everything" to "inference is the growth vector." Jensen Huang's GTC keynote emphasized &lt;strong&gt;AI factories&lt;/strong&gt; as long-running inference infrastructure, not just training clusters. Agentic workloads -- autonomous systems that chain multiple model calls per task -- compound inference demand beyond what chat-era projections captured.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What to watch:&lt;/strong&gt; Any qualitative commentary about inference workload growth, token demand trajectories, or agentic infrastructure spend. If management quantifies inference as a growing share of data center revenue, it supports the thesis that AI compute demand has structural legs beyond model training.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Falsification trigger:&lt;/strong&gt; If inference growth is described as "migrating to edge devices" or "handled by CPU-based systems," the GPU-inference thesis weakens. If there is silence on this topic entirely, the market may be overestimating inference demand relative to what the company sees.&lt;/p&gt;




&lt;h2&gt;
  
  
  5. ASIC Competition Framing
&lt;/h2&gt;

&lt;p&gt;The most credible competitive threat to NVIDIA is custom hyperscaler ASICs: Google's &lt;strong&gt;TPU 8t/8i&lt;/strong&gt; (April 22 launch, Anthropic committed to 1M TPU v7 chips), Amazon's &lt;strong&gt;Trainium 2&lt;/strong&gt;, and Meta's &lt;strong&gt;MTIA&lt;/strong&gt; (accelerating with Broadcom through 2029). Industry analysts project custom ASIC shipments growing significantly faster than GPU shipments in 2026 as hyperscalers vertically integrate.&lt;/p&gt;

&lt;p&gt;However, every ASIC still needs &lt;strong&gt;HBM memory, optical interconnect, and CoWoS packaging&lt;/strong&gt; -- all of which NVIDIA has pre-booked at scale. ASIC growth at the margin does not necessarily mean NVIDIA loses revenue. It means the total AI compute pie is growing, and NVIDIA captures the GPU slice while participating in the broader ecosystem via supply chain positioning.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What to watch:&lt;/strong&gt; How management frames ASIC competition. If they dismiss it as irrelevant, that would suggest they are not tracking the custom silicon trend. If they acknowledge it and frame NVIDIA's counter-position (CUDA moat, NVLink, ecosystem), that signals clear-eyed strategy.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Falsification trigger:&lt;/strong&gt; A hyperscaler announcing that their internal chip has reached &amp;gt;50% utilization across their own AI workload would signal structural erosion. Not expected this quarter.&lt;/p&gt;




&lt;h2&gt;
  
  
  Putting It Together
&lt;/h2&gt;

&lt;p&gt;The standard earnings framework (beat, miss, guide, P/E) tells you how the market feels about NVIDIA today. The 5 signals above test whether NVIDIA's structural position is &lt;strong&gt;improving or eroding.&lt;/strong&gt;&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Signal&lt;/th&gt;
&lt;th&gt;What It Tests&lt;/th&gt;
&lt;th&gt;Bullish&lt;/th&gt;
&lt;th&gt;Bearish&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Purchase commitments&lt;/td&gt;
&lt;td&gt;Supply chain conviction&lt;/td&gt;
&lt;td&gt;Rising&lt;/td&gt;
&lt;td&gt;Flat/falling&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Optical mentions&lt;/td&gt;
&lt;td&gt;Bottleneck migration thesis&lt;/td&gt;
&lt;td&gt;Named by management&lt;/td&gt;
&lt;td&gt;Not discussed&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Blackwell margins&lt;/td&gt;
&lt;td&gt;Ramp economics&lt;/td&gt;
&lt;td&gt;Expanding&lt;/td&gt;
&lt;td&gt;Compressing&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Inference mix&lt;/td&gt;
&lt;td&gt;Demand durability&lt;/td&gt;
&lt;td&gt;Quantified growth&lt;/td&gt;
&lt;td&gt;Silent/edge-focused&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASIC framing&lt;/td&gt;
&lt;td&gt;Competitive awareness&lt;/td&gt;
&lt;td&gt;Acknowledged + countered&lt;/td&gt;
&lt;td&gt;Dismissed&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;The Durability Curve framework&lt;/strong&gt; rates NVIDIA as a Law I (Bottleneck Migration) and Law II (Difficulty Is Load-Bearing) play. The falsification triggers above test both laws. Through this lens, May 20 is not about whether NVIDIA beats by $1B. It is about whether the structural evidence supports or challenges the durability thesis.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This analysis is derived from the **Durability Curve&lt;/em&gt;* research framework, a systematic approach to identifying AI infrastructure bottlenecks before they are priced. The full 36-page NVIDIA Q1 FY2027 earnings research report with detailed falsification triggers, supply chain signal verification across all 5 layers, and options positioning framework is available at:*&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://harryfloyd.gumroad.com/l/nvda-q1-fy2027-earnings-research-report?utm_source=devto" rel="noopener noreferrer"&gt;📄 NVIDIA Q1 FY2027 Earnings Research Report (36 pages, £9)&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Follow &lt;a href="https://techhub.social/@durabilitycurve" rel="noopener noreferrer"&gt;@durabilitycurve&lt;/a&gt; on Mastodon for real-time signal monitoring during the earnings call. Free weekly analysis at &lt;a href="https://harryfloyd.substack.com?utm_source=devto" rel="noopener noreferrer"&gt;harryfloyd.substack.com&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Not financial advice. All data points verified against public sources as of May 17, 2026. Verify independently before making investment decisions.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>infrastructure</category>
      <category>investing</category>
      <category>analysis</category>
    </item>
    <item>
      <title>A Developer's Guide to AI Inference Costs in 2026</title>
      <dc:creator>Harry Floyd</dc:creator>
      <pubDate>Sat, 16 May 2026 21:45:26 +0000</pubDate>
      <link>https://forem.com/harryfloyd/a-developers-guide-to-ai-inference-costs-in-2026-1h6g</link>
      <guid>https://forem.com/harryfloyd/a-developers-guide-to-ai-inference-costs-in-2026-1h6g</guid>
      <description>&lt;p&gt;If you're building AI features in 2026, your gross margin depends on a question most developers don't have a good answer to: &lt;strong&gt;what does one inference actually cost?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The answer isn't in the model card. It's in the physical infrastructure chain that runs from a fab in Taiwan to a data centre in Virginia. Here's how to estimate it.&lt;/p&gt;

&lt;h2&gt;
  
  
  The easy part: API pricing
&lt;/h2&gt;

&lt;p&gt;If you're using an API (OpenAI, Anthropic, Together, Groq), your per-token cost is known. The hidden variable is &lt;strong&gt;cache-hit rate&lt;/strong&gt;. Prompt caching drops cost by 2-10x depending on how much of your system prompt is shared across requests. If you haven't measured your cache-hit ratio, you don't know your true cost.&lt;/p&gt;

&lt;p&gt;Most teams I've seen get 30-50% cache hits on well-structured prompts and close to 0% on dynamic ones. That's a 2x difference in effective cost hiding in plain sight.&lt;/p&gt;

&lt;h2&gt;
  
  
  The harder part: self-hosted inference
&lt;/h2&gt;

&lt;p&gt;Running your own models means paying for GPU time whether you're using it or not. The number that matters: &lt;strong&gt;utilization rate&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;A single H100 at $2-3/hour needs to be generating tokens &amp;gt;60% of the time to beat API pricing at scale. Below 30% utilization, you'd have been better off on an API. Most self-hosted deployments I see run at 15-25% because of traffic spikes and idle standby capacity.&lt;/p&gt;

&lt;p&gt;The break-even math:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;API: pay per token, zero fixed cost&lt;/li&gt;
&lt;li&gt;Self-hosted: ~$2k/month per GPU all-in, first ~2M tokens are effectively "paying off the fixed cost"&lt;/li&gt;
&lt;li&gt;Breakeven: ~4-5M tokens/month per GPU, assuming 60% utilization&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The hidden constraint: hardware availability
&lt;/h2&gt;

&lt;p&gt;This is the part most infrastructure analyses miss. In 2026, GPU lead times are still 12-18 months for new deployments. H200s are shipping but allocated. The secondary market for A100s is active but prices haven't dropped as much as expected — because demand from inference workloads has replaced training demand.&lt;/p&gt;

&lt;p&gt;What this means for your deployment plan: &lt;strong&gt;if you need GPUs in a quarter, you're renting&lt;/strong&gt;. If you're renting, you can't amortize hardware cost. If you can't amortize, you're at the mercy of spot pricing — which has swung 40% in a single month twice this year already.&lt;/p&gt;

&lt;h2&gt;
  
  
  The one number to track
&lt;/h2&gt;

&lt;p&gt;For any AI feature, track &lt;strong&gt;cost per completed interaction&lt;/strong&gt; — not cost per token. Token counts hide the real metric: how many tokens does your average user interaction consume?&lt;/p&gt;

&lt;p&gt;A chatbot using Claude Sonnet 4.6 ($3/M input, $15/M output) averaging 2,000 tokens per conversation with a typical 70/30 input/output split costs roughly &lt;strong&gt;$0.013 per conversation&lt;/strong&gt;. At 100k conversations/month, that's $1,300 — significant enough that a 10% improvement in token efficiency pays for an engineer's time.&lt;/p&gt;

&lt;p&gt;Most teams don't know their average cost per interaction. That's the first number worth instrumenting — without it, you can't tell whether optimisation matters or not.&lt;/p&gt;

&lt;h2&gt;
  
  
  Summary
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Measure your cache-hit ratio (30-50% is typical; anything below 20% means expensive redundant computation)&lt;/li&gt;
&lt;li&gt;Track cost per completed interaction, not per token&lt;/li&gt;
&lt;li&gt;Know your self-host breakeven point (~4M tokens/month per GPU)&lt;/li&gt;
&lt;li&gt;Assume 12-month lead times for hardware — plan accordingly&lt;/li&gt;
&lt;li&gt;Spot GPU pricing can swing 40% in a month; don't build on spot for production&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The infrastructure layer is the part of AI most developers treat as someone else's problem. It isn't. The teams that understand their cost-per-interaction will build features that survive the margin compression that's coming.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>infrastructure</category>
      <category>cloud</category>
      <category>architecture</category>
    </item>
    <item>
      <title>The Model Is Not the Moat</title>
      <dc:creator>Harry Floyd</dc:creator>
      <pubDate>Fri, 15 May 2026 17:00:32 +0000</pubDate>
      <link>https://forem.com/harryfloyd/the-model-is-not-the-moat-3b4d</link>
      <guid>https://forem.com/harryfloyd/the-model-is-not-the-moat-3b4d</guid>
      <description>&lt;p&gt;I keep hearing the same assumption underneath AI strategy talk: the winner will be whoever has the strongest model. Smarter model wins. Everything else is secondary.&lt;/p&gt;

&lt;p&gt;That sounds plausible right up until you look at how people actually choose tools in real life.&lt;/p&gt;

&lt;p&gt;If scale laws keep holding, local models probably won't beat the frontier on raw intelligence. But users don't adopt "raw intelligence." They adopt something that fits into their day: something fast enough, private enough, legible enough, reliable enough, cheap enough, and integrated enough that using it becomes natural.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The benchmark measures the visible product. The moat forms one layer out, in the surrounding package.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The Product and the Seat
&lt;/h2&gt;

&lt;p&gt;People compare models as if users experience them as isolated intelligence engines. Most of the time they don't. They experience a bundle: interface, defaults, permissions, speed, memory, privacy, cost, and how much the tool asks them to rearrange their behaviour.&lt;/p&gt;

&lt;p&gt;That bundle is where trust accumulates. It's also where switching costs quietly form.&lt;/p&gt;

&lt;p&gt;A frontier model can win the benchmark and still lose the seat.&lt;/p&gt;

&lt;p&gt;By "seat" I mean the position a product earns inside a user's actual workflow. The place where their context lives. The place they trust not to embarrass them, leak data, slow them down, or force them to relearn everything.&lt;/p&gt;

&lt;p&gt;The product is the thing they evaluate. The seat is the thing they get used to living in. These are not the same asset.&lt;/p&gt;

&lt;h2&gt;
  
  
  You Can Already See It in Coding Tools
&lt;/h2&gt;

&lt;p&gt;A model that is slightly worse on a public benchmark can still be the one people prefer if it lives inside the editor, sees the repo, responds instantly, keeps sensitive code local, and fits the way they already work.&lt;/p&gt;

&lt;p&gt;This is also why "just use the best model" is often bad product advice. The best model according to a leaderboard may carry the wrong latency, the wrong privacy posture, the wrong integration burden, or the wrong failure mode for the actual job.&lt;/p&gt;

&lt;h2&gt;
  
  
  How Moats Actually Form
&lt;/h2&gt;

&lt;p&gt;Apple is the familiar version of this dynamic outside AI. The moat was never just one visible feature. It was the package: ecosystem fit, convenience, defaults, identity, and the low-grade friction of leaving. The product got attention. The package became hard to leave.&lt;/p&gt;

&lt;p&gt;I think a lot of AI products will work the same way. Moats usually form in the residue, not in the headline claim. In habit. In muscle memory. In stored context. In predictable behaviour. In the feeling that this tool understands how you work and doesn't make you pay a tax every time you use it.&lt;/p&gt;

&lt;h2&gt;
  
  
  What This Means for Builders
&lt;/h2&gt;

&lt;p&gt;Local models don't need to win the intelligence race to matter. They can win a different race entirely: trust, control, governance, latency, privacy, and workflow fit.&lt;/p&gt;

&lt;p&gt;Once a tool becomes the place where your context lives, your defaults settle, and your work starts to flow, a better benchmark somewhere else is not enough to dislodge it.&lt;/p&gt;

&lt;p&gt;If I were building in this market, I'd treat that as a design rule. Don't ask only, "How do we make the model look stronger?" Ask:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Where does the user feel risk right now?&lt;/li&gt;
&lt;li&gt;What part of the workflow still feels awkward or fragile?&lt;/li&gt;
&lt;li&gt;What context would make the tool more useful after 30 days than on day 1?&lt;/li&gt;
&lt;li&gt;What would make leaving this product feel expensive in a good way?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That is how you build the seat. Not by winning one benchmark snapshot, but by creating a package that compounds through use.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>infrastructure</category>
      <category>analysis</category>
      <category>architecture</category>
    </item>
  </channel>
</rss>
