<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Jose Crespo, PhD</title>
    <description>The latest articles on Forem by Jose Crespo, PhD (@jose_crespo_phd).</description>
    <link>https://forem.com/jose_crespo_phd</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3592257%2Fc78c73bf-6bd3-499e-9590-1325927e1223.png</url>
      <title>Forem: Jose Crespo, PhD</title>
      <link>https://forem.com/jose_crespo_phd</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/jose_crespo_phd"/>
    <language>en</language>
    <item>
      <title>The Two Programming Styles of AI — and Why Everyone Uses the Wrong One</title>
      <dc:creator>Jose Crespo, PhD</dc:creator>
      <pubDate>Wed, 26 Nov 2025 11:04:37 +0000</pubDate>
      <link>https://forem.com/jose_crespo_phd/the-two-programming-styles-of-ai-and-why-everyone-uses-the-wrong-one-c0n</link>
      <guid>https://forem.com/jose_crespo_phd/the-two-programming-styles-of-ai-and-why-everyone-uses-the-wrong-one-c0n</guid>
      <description>&lt;h2&gt;
  
  
  AI keeps crashing against the same walls
&lt;/h2&gt;

&lt;p&gt;Yep, everybody, even Tesla, is using the old damn math from 2 centuries ago, and hence it’s not surprising to watch many of scenes like this all over YouTube when you rely on AI driving your Tesla:&lt;/p&gt;

&lt;p&gt;Apparently, not even Tesla - with its 1.4 Trillion valuation and army of PhDs - knows about this math. Or maybe they do, and just enjoy watching their cars perform interpretive dance routines at 60 mph.&lt;/p&gt;

&lt;p&gt;Either way, here’s the greatest hits compilation you’ve seen all over YouTube:&lt;/p&gt;

&lt;p&gt;The Tesla Self-Driving Blooper Reel:&lt;/p&gt;

&lt;p&gt;🎬 &lt;a href="https://www.reddit.com/r/TeslaUK/comments/1lzojc5/phantom_braking_in_2025/" rel="noopener noreferrer"&gt;Phantom Braking&lt;/a&gt; - The car slams the brakes for a shadow. Because apparently, shadows are the #1 threat to highway safety in the 21st century.&lt;/p&gt;

&lt;p&gt;🎬 &lt;a href="https://static.nhtsa.gov/odi/rcl/2023/RCLRPT-23V838-8276.PDF" rel="noopener noreferrer"&gt;The Surprise Party Turn&lt;/a&gt; - Takes curves at full speed, then goes “OH SHIT A CURVE!” and throws a mini-chicane out of nowhere, comedy for everyone, except your neck..&lt;/p&gt;

&lt;p&gt;🎬 &lt;a href="https://static.nhtsa.gov/odi/rcl/2025/RCLRPT-25V092-6812.PDF" rel="noopener noreferrer"&gt;The Seizure Shuffle &lt;/a&gt;- Steering adjustments so jerky you’d think the car is having an existential crisis. Left, right, left, right.. it’s not driving, it’s vibrating down the highway.&lt;/p&gt;

&lt;p&gt;🎬 &lt;a href="https://www.tesla.com/fsd/safety" rel="noopener noreferrer"&gt;The “Why Did It Do That?” &lt;/a&gt;- Does something so inexplicable that even the AI researchers watching the logs just shrug and mutter “gradient descent, probably.”&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;If this post is sparking your curiosity, you may enjoy exploring deeper analyses, geometric AI concepts, and ongoing research on my personal site:&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
👉 &lt;a href="https://josecrespo.substack.com" rel="noopener noreferrer"&gt;https://josecrespo.substack.com&lt;/a&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  The Fix That Nobody’s Using
&lt;/h2&gt;

&lt;p&gt;Tesla could solve this - easily - by using second derivatives (Hessian-vector products, or HVP for the cool kids).&lt;/p&gt;

&lt;p&gt;So could Google, Meta, OpenAI, and pretty much every company with an “AI Strategy” PowerPoint deck.&lt;/p&gt;

&lt;p&gt;But they’re not. See the table below - notice a pattern?&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqx37hrvq7hzvd3d9y1vx.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqx37hrvq7hzvd3d9y1vx.png" alt=" " width="556" height="366"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Wait — These Are Different Problems, Right?
&lt;/h2&gt;

&lt;p&gt;Not exactly. They are different symptoms, but the same disease.&lt;/p&gt;

&lt;p&gt;They’re all using math that can answer “Which way should I go?”&lt;/p&gt;

&lt;p&gt;but not “How sharply is this about to change?”&lt;/p&gt;

&lt;p&gt;It’s like asking a GPS for directions but never checking if there’s a cliff ahead.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Root Cause: Your Great-great-grandfather’s Calculus
&lt;/h2&gt;

&lt;p&gt;As said, in the case of Tesla what is happening is that their cars are reacting to what’s happening right now, not anticipating what’s about to happen.&lt;/p&gt;

&lt;p&gt;It’s like playing chess by only looking at the current board position - no planning, no strategy, just “I see a piece, I move a piece.”&lt;/p&gt;

&lt;p&gt;Chess players call this “beginner level.” Tesla calls it “Full Self-Driving.”&lt;/p&gt;

&lt;p&gt;Ready for the diagnosis? Tesla engineers, like everyone else in Silicon Valley, are still using 19th-century limit-based calculus — the math equivalent of trying to stream Netflix on a telegraph machine.&lt;/p&gt;

&lt;p&gt;Meanwhile, the solution has been sitting on the shelf for 60 years: dual/jet numbers.&lt;/p&gt;

&lt;p&gt;Nobody thought to check the manual. Seriously, who bothers with that “wacko, exotic math” they don’t teach in university CS programs?s?&lt;/p&gt;

&lt;p&gt;And yet, these hyperreal-related algebras (duals and jets) make second derivatives (HVP) a computationally trivial operation through the elegant composition of two first-order operators (JVP ∘ VJP).&lt;/p&gt;

&lt;h2&gt;
  
  
  Hold Up — Are You Telling Me…
&lt;/h2&gt;

&lt;p&gt;that the “gold-standard” h-limit calculus makes it a slog, while duals/jets make it trivial.…that what’s computationally intractable with the traditional h-limit calculus so many Ivy-League courses treat as the gold standard is trivial with dual/jet numbers, that can fix most of those damn curve-related problems in our current AI?&lt;/p&gt;

&lt;p&gt;Yes. Exactly that.&lt;/p&gt;

&lt;p&gt;And it gets worse.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Hyperreal Revolution: Your Calculus Professor Never Told You This
&lt;/h2&gt;

&lt;p&gt;The calculus you learned in college — the one that got you through differential equations, optimization theory, and machine learning courses — isn’t wrong. It’s just incomplete.&lt;/p&gt;

&lt;p&gt;It’s like learning arithmetic but never being taught that multiplication is just repeated addition. You can still do math, but you’re doing it the hard way.&lt;/p&gt;

&lt;p&gt;Here’s the specific problem:&lt;/p&gt;

&lt;p&gt;Traditional calculus (the h-limit approach):&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;f'(x) = lim[h→0] (f(x+h) - f(x)) / h
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This defines derivatives as limits — which means:&lt;/p&gt;

&lt;p&gt;✅ Mathematically rigorous&lt;br&gt;
✅ Great for proving theorems&lt;br&gt;
❌ Computationally nightmarish for anything beyond first derivatives&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;f'(x+h) = lim[h'→0] (f(x+h+h') - f(x+h)) / h'
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;em&gt;But f'(x+h) itself requires computing&lt;/em&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;f'(x+h) = lim[h'→ 0] (f(x+h+h') - f(x+h)) / h'
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;So, summing up: either you end up with nested limits and two step sizes (h,h′) that interact unstably, or you resort to higher-order stencils that are exquisitely sensitive to step size and noise. In both cases you lose derivative structure, so two first-derivative passes (JVP → VJP) don’t compose into a true second derivative - you’re rebuilding guesses instead of carrying derivatives.&lt;/p&gt;

&lt;p&gt;For a third derivative? Three nested limits or or use even higher-order stencils.&lt;/p&gt;

&lt;p&gt;For the k-th derivative: either nest k layers or use wider stencils - noise blows up as O(h^-k), truncation depends on stencil order, and you still lose derivative structure, so JVP→VJP won’t compose into HVP in an FD pipeline.&lt;/p&gt;

&lt;p&gt;So your self-driving car keeps crashing against sun-set lit walls.&lt;/p&gt;

&lt;p&gt;And for GPT-5’s approximately 1.8 trillion parameters? Computational impossibility.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sharp Readers Will Notice:
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;“Hold on, if we know the function f, can’t we just compute f’ and f’’ analytically? Why do we need any of this limit or dual number stuff?”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Great question!&lt;/p&gt;

&lt;p&gt;Here’s why that doesn’t work for neural networks:&lt;/p&gt;

&lt;p&gt;The Problem: Neural Networks Are Black Boxes&lt;br&gt;
When you write a simple function, you can compute derivatives analytically:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;# Simple case - analytic derivatives work fine
f(x) = x² + 3x + 5
f'(x) = 2x + 3      # Easy to derive by hand
f''(x) = 2          # Even easier
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;But a neural network with 1.8 trillion parameters looks like this:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;f(x) = σ(W₁₇₅·σ(W₁₇₄·σ(...σ(W₂·σ(W₁·x))...)))


Where:
- Each `W` is a matrix with billions of parameters
- Each `σ` is a nonlinear activation function
- There are hundreds of layers (GPT-style)
- The composition is **dynamically computed** during runtime

You literally cannot write down the analytic form of f'(x) because:
1. The function changes every time you update parameters (every training step)
2. It's too large to express symbolically
3. It contains billions of nested compositions


### Why Traditional Calculus Fails Here

**The h-limit formula:**

f''(x) = lim[h→0] (f'(x+h) - f'(x)) / h


**Requires you to evaluate 

f'(x+h)`**, which means:
f'(x+h) = lim[h'→0] (f(x+h+h') - f(x+h)) / h'
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;And here’s the trap:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;You can’t compute f' analytically (the function is too complex)&lt;/li&gt;
&lt;li&gt;So you approximate it using finite differences (the h-limit)&lt;/li&gt;
&lt;li&gt;Now you need f'(x+h) for the second derivative&lt;/li&gt;
&lt;li&gt;So you approximate that using another finite difference (with step size h’)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Result: You’re approximating an approximation — errors compound catastrophically.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The skeptical reader might continue objecting: &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;”But can’t we use something like SymPy or Mathematica to compute derivatives symbolically?”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;In theory, yes. In practice, we face a similar problem.&lt;/p&gt;

&lt;p&gt;For a 1.8 trillion parameter model!:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The symbolic expression for f' would be larger than the model itself 👀&lt;/li&gt;
&lt;li&gt;Computing it would take years&lt;/li&gt;
&lt;li&gt;Storing it would require more memory than exists&lt;/li&gt;
&lt;li&gt;Simplifying it would be computationally intractable&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Example: Even for a tiny 3-layer network with 1000 neurons per layer:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Symbolic f' lands in the millions of terms.&lt;/li&gt;
&lt;li&gt;Symbolic f'' jumps to the billions of terms.&lt;/li&gt;
&lt;li&gt;Growth is combinatorial with depth/width; common-subexpression tricks don’t save you enough.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For hundred of layers? Forget it.&lt;/p&gt;

&lt;p&gt;clear now?&lt;/p&gt;

&lt;p&gt;Let’s Bring Back Our Hyperreals Flavor for AI Computing and let’s see what happens when hyperreals face similar scenarios:&lt;/p&gt;

&lt;p&gt;What Dual/Jet Numbers Do Differently: Automatic Differentiation&lt;br&gt;
Dual numbers don’t use limits at all. Instead, they:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;- Encode the differentiation rules in the arithmetic&lt;/li&gt;
&lt;li&gt;- Evaluate f with special numbers that carry derivative info&lt;/li&gt;
&lt;li&gt;- Derivatives emerge through rule-following arithmetic&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Jets generalize this. k-jets carry truncated Taylor lanes up to order k (nilpotent ε^k+1=0), so higher-order derivatives fall out in one pass.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Here’s the key: The calculus rules (power rule, chain rule, etc.) are built into the jet arithmetic operations, not applied symbolically! So you get all the advantages of analytical solution without using them!&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The Three Fundamental Differences&lt;br&gt;
Calculus with Symbolic Rule Application ( impractical at modern AI scale)&lt;br&gt;
Process:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Write down the function: f(x) = x³&lt;/li&gt;
&lt;li&gt;Recall the power rule: d/dx[xⁿ] = n·xⁿ⁻¹&lt;/li&gt;
&lt;li&gt;Apply it symbolically: f’(x) = 3x²&lt;/li&gt;
&lt;li&gt;Store both formulas separately&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;For neural networks: Must build the entire derivative expression — exponential memory explosion.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Traditional h-Limit Calculus: Numerical Approximation&lt;/strong&gt;:&lt;/p&gt;

&lt;p&gt;Process:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Choose a step size h (guesswork)&lt;/li&gt;
&lt;li&gt;Evaluate: (f(x+h) — f(x))/h&lt;/li&gt;
&lt;li&gt;Get an approximation with error&lt;/li&gt;
&lt;li&gt;Problems:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Not exact (always has truncation or roundoff error)&lt;br&gt;
Can’t compose cleanly&lt;br&gt;
Breaks down at higher orders&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Dual/Jet Numbers Algebra: Evaluation with Augmented Arithmetic(practical at modern AI scale)&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Process:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Extend the number system with ε where ε² = 0&lt;/li&gt;
&lt;li&gt;Evaluate f at (x + ε) using this arithmetic&lt;/li&gt;
&lt;li&gt;Derivatives appear as ε-coefficients automatically&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;For neural networks&lt;/strong&gt;: No expression built — just evaluate once with special numbers. Linear memory scaling.&lt;/p&gt;
&lt;h2&gt;
  
  
  How It Actually Works: The Binomial Magic with dual numbers
&lt;/h2&gt;

&lt;p&gt;Let’s see as a toy example how the power rule emerge without applying any calculus:&lt;/p&gt;

&lt;p&gt;Example: compute derivative of f(x) = x³&lt;/p&gt;

&lt;p&gt;Step 1: Evaluate at augmented input&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example: compute derivative of f(x) = x³&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Step 1: Evaluate at augmented input&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;f(x + ε) = (x + ε)³
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Step 2: Expand using binomial theorem (combinatorics, not calculus)&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;(x + ε)³ = x³ + 3x²ε + 3xε² + ε³
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Step 3: Apply nilpotent algebra (ε² = 0)&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;= x³ + 3x²ε + 0 + 0
= x³ + 3x²ε
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Step 4: Read the dual number&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;x³ + 3x²ε = (x³) + ε·(3x²)
            ↑         ↑
         value   derivative
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The derivative f’(x) = 3x² emerged through:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Binomial expansion (algebra)&lt;/li&gt;
&lt;li&gt;Nilpotent simplification (ε² = 0)&lt;/li&gt;
&lt;li&gt;Coefficient reading&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;NOT through:&lt;/p&gt;

&lt;p&gt;❌ Power rule application&lt;br&gt;
❌ h-limit formula&lt;br&gt;
❌ Symbolic differentiation&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;gt; You don’t apply the power rule — you let binomial expansion reveal it.&lt;/code&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why This Scales When Symbolic Differentiation Doesn’t
&lt;/h2&gt;

&lt;p&gt;Symbolic Differentiation (Analytical):&lt;br&gt;
With AI working with neural networkd you must build expressions:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Layer 1 derivative: thousands terms&lt;/li&gt;
&lt;li&gt;Layer 2 derivative: millions terms (combinatorial explosion)&lt;/li&gt;
&lt;li&gt;Hundreds of layers: expression size grows exponentially in depth/width; even with common-subexpression elimination it becomes intractable to construct, store, or simplify.
&lt;strong&gt;Memory required: More than all atoms in the universe 👀&lt;/strong&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Dual Number Evaluation:&lt;/strong&gt;&lt;br&gt;
Never builds expressions:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Each instrumented tensor stored value + ε·derivative&lt;/li&gt;
&lt;li&gt;Memory: 2× base model (for k=1)&lt;/li&gt;
&lt;li&gt;Or 3× base model with Jets (for k=2 with second derivative)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For GPT-5 (1.8T parameters):&lt;br&gt;
&lt;a href="https://www.nvidia.com/en-us/data-center/h100/" rel="noopener noreferrer"&gt;k=1: ~14.4 TB → 18.0 TB (totally practical)&lt;/a&gt;&lt;br&gt;
&lt;a href="https://arxiv.org/abs/2203.15556" rel="noopener noreferrer"&gt;k=2: ~14.4 TB → 21.6 TB (fits on ~34 H100 nodes)&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  BUT WAIT — YOU’RE FLYING FIRST CLASS IN AI MATH
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;And there’s still more.&lt;/strong&gt;&lt;br&gt;
The algebra of dual/jet numbers lets you use composition of functions (yup, if you want to do yourself a favor and write real AI that works, learn category theory now!).&lt;/p&gt;

&lt;p&gt;Here’s your genius move:&lt;/p&gt;

&lt;p&gt;With composition of functions, we can get second derivatives for the price of a first derivative!!&lt;/p&gt;

&lt;p&gt;Woah. 🤯&lt;/p&gt;

&lt;p&gt;How? Just by using composition of functions — otherwise structurally impossible with limit-based calculus.&lt;/p&gt;

&lt;h2&gt;
  
  
  In Plain English: Why Composition Fails With h-Limits
&lt;/h2&gt;

&lt;p&gt;Traditional calculus can’t do JVP∘VJP = HVP because:&lt;br&gt;
JVP via finite differences gives you a number (an approximation of f’(x)·v)&lt;br&gt;
That number has no derivative structure for VJP to differentiate&lt;br&gt;
You must start over with a new finite-difference approximation&lt;br&gt;
The operations don’t chain — each one discards the structure the next one needs&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Dual numbers CAN do JVP∘VJP = HVP because:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;JVP with duals gives you a dual number (f(x), f'(x)·v)&lt;/li&gt;
&lt;li&gt;That dual number carries derivative structure in its ε-coefficient&lt;/li&gt;
&lt;li&gt;VJP can differentiate it directly by treating it as input&lt;/li&gt;
&lt;li&gt;The operations chain naturally — each preserves the structure the next needs&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Dual numbers are algebraically closed under composition.&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
The Practical Consequence&lt;br&gt;
what the new paradigm can compute that the old one can’t:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7cs4zhk6prctiqcemfuf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7cs4zhk6prctiqcemfuf.png" alt=" " width="566" height="307"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Why This Is The Key To Fixing AI&lt;br&gt;
&lt;strong&gt;Current AI (k=1 only):&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Can answer: “Which direction should I go?”&lt;/li&gt;
&lt;li&gt;Cannot answer: “How sharply is this direction changing?”&lt;/li&gt;
&lt;li&gt;Result: Reactive, not anticipatory&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;With composition (JVP∘VJP):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Get second derivatives for 2× the cost of first derivatives&lt;/li&gt;
&lt;li&gt;Can anticipate curves, detect trajectory changes&lt;/li&gt;
&lt;li&gt;Result: one of many examples, &lt;a href="https://www.reuters.com/business/autos-transportation/us-probes-416000-tesla-vehicles-over-unexpected-braking-reports-2022-02-17/" rel="noopener noreferrer"&gt;Tesla stops phantom braking&lt;/a&gt;; &lt;a href="https://www.reuters.com/business/autos-transportation/us-probes-416000-tesla-vehicles-over-unexpected-braking-reports-2022-02-17/" rel="noopener noreferrer"&gt;AI stops hallucinating&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;With explicit k=3 jets:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Get third derivatives for 3× the cost&lt;/li&gt;
&lt;li&gt;Can verify topological consistency (winding numbers)&lt;/li&gt;
&lt;li&gt;Result: Mathematically certified AI outputs&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The Functors + Composition Advantage
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;And why Hyperreal Algebra Matters:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Without it (finite differences):&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Each derivative order requires starting from scratch&lt;/li&gt;
&lt;li&gt;Errors accumulate with each nesting&lt;/li&gt;
&lt;li&gt;No compositional structure to exploit&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;With it (dual/jet numbers):&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Higher-order derivatives = compose lower-order operations&lt;br&gt;
Exact (within floating-point)&lt;br&gt;
Automatic (chain rule built into ε-arithmetic)&lt;br&gt;
This is why:&lt;/p&gt;

&lt;p&gt;✅ Dual/Jet numbers scale to hundred of layers (linear memory)&lt;/p&gt;

&lt;p&gt;✅ Composition works (JVP∘VJP = HVP automatically)&lt;/p&gt;

&lt;p&gt;✅ Higher orders accessible with Jet numbers ( k=3, k=4 feasible)&lt;/p&gt;

&lt;p&gt;And why:&lt;/p&gt;

&lt;p&gt;❌ Symbolic differentiation explodes (exponential expressions)&lt;br&gt;
❌ Finite differences can’t compose (no functoriality)&lt;br&gt;
❌ h-limit methods break at higher orders (error compounds)&lt;/p&gt;




&lt;h2&gt;
  
  
  SUMMING UP
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The entire AI industry is stuck at first-order optimization because:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;They learned calculus as h-limits (doesn’t scale)&lt;/li&gt;
&lt;li&gt;They implement derivatives as finite differences (doesn’t compose)&lt;/li&gt;
&lt;li&gt;They never learned about Group Theory and Hyperreal Numbers (not in CS curricula)&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Meanwhile:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Dual numbers make derivatives algebraic objects (not approximations)&lt;/li&gt;
&lt;li&gt;Jets make higher orders linear in cost (not exponential)&lt;/li&gt;
&lt;li&gt;Functorial composition makes second derivatives cheap (JVP∘VJP)&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The math to fix Tesla’s phantom braking, OpenAI’s hallucinations, and Meta’s moderation chaos has been sitting in textbooks since 1960s.&lt;/p&gt;

&lt;p&gt;Waiting for someone to connect the dots among: the binomial theorem (~400 years old), nilpotent algebra (~150 years old), and functorial composition + hyperreals (~60 years old).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;To the biggest unsolved problems in AI.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Now you know what Silicon Valley doesn’t and see what they cannot.&lt;/strong&gt;&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;NOTE&lt;/strong&gt;: &lt;em&gt;In this article, “traditional calculus” means the finite-difference (h-limit) implementation used in practice — pick an h, approximate, repeat — not analytic/symbolic derivatives.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;If this post has sparked your curiosity, you may enjoy exploring deeper analyses, geometric AI concepts, and ongoing research on my personal site:&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
👉 &lt;a href="https://josecrespo.substack.com" rel="noopener noreferrer"&gt;https://josecrespo.substack.com&lt;/a&gt;&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>algorithms</category>
    </item>
    <item>
      <title>LLMs Are Dying - The New AI Is Killing Them</title>
      <dc:creator>Jose Crespo, PhD</dc:creator>
      <pubDate>Sat, 15 Nov 2025 09:07:15 +0000</pubDate>
      <link>https://forem.com/jose_crespo_phd/llms-are-dying-the-new-ai-is-killing-them-1aa4</link>
      <guid>https://forem.com/jose_crespo_phd/llms-are-dying-the-new-ai-is-killing-them-1aa4</guid>
      <description>&lt;h2&gt;
  
  
  LMs are already museum pieces
&lt;/h2&gt;

&lt;p&gt;Yes, ChatGPT, Claude, Gemini, all of them. Brilliant fossils of a linguistic age that’s already ending. They’re decomposing in public, billions are still being spent to polish their coffins: bigger models, longer contexts, more hallucinations per watt.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The “predict-the-next-word” LLM era is over.&lt;/strong&gt;&lt;br&gt;
The new killer on the stage isn’t language, it’s world modeling.&lt;br&gt;
An AI that understands reality like a conceptual puzzle, where you don’t need every piece to see the whole picture.&lt;/p&gt;

&lt;p&gt;We just haven’t dragged the old LLM stars off the stage yet.&lt;br&gt;
If you’re still building your career around prompting chatbots, you’re preparing for a world that’s already gone.&lt;/p&gt;

&lt;p&gt;We don’t need machines that shovel tokens like brainless parrots. We need machines that carry objects through time, that bind cause to effect, that keep the scene consistent. “Stochastically parroted” was cute for 2023. It’s not intelligence. It’s probability cosplay.&lt;/p&gt;

&lt;p&gt;And here’s what many of us have seen already before in the tech industry: if AI keeps stitching likelihood instead of reality, the industry that burned 1 trillion dollars to make chatty cockatoos will burst. Soon.&lt;/p&gt;

&lt;h2&gt;
  
  
  Is JEPA the New AI Hope?
&lt;/h2&gt;

&lt;p&gt;In a galaxy not so far from Menlo Park, a new acronym rises from the ashes of the chatbot empire: JEPA!&lt;/p&gt;

&lt;p&gt;The promised one. The architecture said to restore balance to the Force after the LLM wars.&lt;/p&gt;

&lt;p&gt;Before we let Meta’s marketing department baptize it as the next Messiah, let’s decode the acronym.&lt;br&gt;
JEPA stands for Joint Embedding Predictive Architecture - four words that sound profound until you realize they describe a fairly old idea wearing a new coat of GPU varnish:&lt;/p&gt;

&lt;p&gt;Joint, because it trains two halves of a scene — one visible, one masked — and forces them to agree in a shared latent space.&lt;br&gt;
Embedding, Because instead of dealing with raw pixels or words like LLMs do, it operates in dense vector representations: the hidden space where modern AI stores meaning.&lt;br&gt;
Predictive, because its only trick is to predict the missing chunk from the one it still sees.&lt;br&gt;
Architecture, because every new AI concept needs an impressive noun at the end to look academic.&lt;br&gt;
As Yann LeCun - the new high priest of the post-LLM cult - likes to put it, in plain English:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Intelligence isn’t language, nor is it predicting the next word. Intelligence is predicting what will happen in the world.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;And that’s exactly what JEPA tries to do.&lt;br&gt;
It learns to guess what’s missing in a world it only half perceives — not by generating text or pixels, but by aligning internal representations so that context can explain absence.&lt;/p&gt;

&lt;p&gt;It doesn’t write; it completes.&lt;br&gt;
It doesn’t imagine; it infers.&lt;/p&gt;

&lt;h2&gt;
  
  
  Too Good to Be True ?
&lt;/h2&gt;

&lt;p&gt;But you’re probably wondering whether this is the real deal - or just another half-baked tech product with brighter LEDs and duller intelligence&lt;/p&gt;

&lt;p&gt;Let’s cut through the grand talk for a second.&lt;br&gt;
Meta presents JEPA like a revelation carved in silicon: the dawn of “true intelligence,” the end of chatbots, the start of world-model gods.&lt;/p&gt;

&lt;p&gt;And at first glance, it really does look that way — until you&lt;br&gt;
strip off the marketing halo, and you’ll see what’s actually there.&lt;br&gt;
Not quite the miracle LeCun sells, but still something valuable: a step in the right direction, but still long way to go&lt;/p&gt;

&lt;p&gt;For now, JEPA lives mostly in videos and physical-world scenarios, while the wording layer still leans heavily on the same stochastic parrot LLMs they claim to have surpassed (see chart 1 below).&lt;/p&gt;

&lt;p&gt;Of course, Meta just prefers not to mention that part in the brochures. No surprise there.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F71b5lelm423yljqem1ag.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F71b5lelm423yljqem1ag.png" alt=" " width="717" height="762"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;So yes, In the blueprint JEPA looks clever - new modules, shiny arrows, and a fresh sense of purpose, but underneath, we’re still stirring the same pot.&lt;br&gt;
The word-soup problem just got an upgrade to concept-soup.&lt;br&gt;
Different flavor, same indigestion.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;To be fair, JEPA isn’t pure marketing incense.&lt;br&gt;
As the chart below shows — humorously stripped of Meta’s triumphal glow its context encoder, target encoder, and predictor form a neat little triad that genuinely breaks from the token tyranny of LLMs.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdbob66ps3jnq9n9mjraj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdbob66ps3jnq9n9mjraj.png" alt=" " width="800" height="529"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;So what can you actually do with JEPA, you’re asking?&lt;br&gt;
Well - you can show it half an image, and it predicts the missing part for you, not by painting pixels, but by reasoning in latent space.&lt;br&gt;
That’s progress: perception over parroting.&lt;/p&gt;

&lt;p&gt;And here’s the lineup:&lt;/p&gt;

&lt;p&gt;I-JEPA recognized ImageNet objects with only a few labeled samples, impressive. But it still fails under fixed distractor noise and never touches language.&lt;/p&gt;

&lt;p&gt;Then comes V-JEPA, trained on videos to learn motion intuition, what moves, when, and why. But not real physics: collisions, forces, contact still out of reach.&lt;/p&gt;

&lt;p&gt;More into robotics? V-JEPA 2 guides robotic arms, predicting how objects behave before acting. But it still trails humans on physical benchmarks and, irony of ironies, needs an LLM to talk about what it sees.&lt;/p&gt;

&lt;p&gt;So yes, progress, enough to declare LLM technology as a fossil in artificial life maintenance - but still flatland thinking dressed up as revelation.&lt;/p&gt;

&lt;p&gt;And that’s the part Meta doesn’t want you to see.&lt;/p&gt;

&lt;p&gt;Here’s the fatal secret hiding behind JEPA’s slick diagrams and LeCun’s confident declarations: they solved the linguistic problem by creating an architectural one that’s mathematically far worse.&lt;/p&gt;

&lt;p&gt;They escaped the word-prison only to build a concept-prison with thicker walls.&lt;/p&gt;

&lt;p&gt;Think about it: LLMs were bad because they treated reality as a linear sequence of tokens: a one-dimensional march through probability space. JEPA says “we’re smarter than that” and jumps into high-dimensional representation space, where concepts float as vectors in a 768-dimensional cloud.&lt;/p&gt;

&lt;p&gt;Sounds impressive, right?&lt;/p&gt;

&lt;p&gt;Wrong.&lt;/p&gt;

&lt;p&gt;They just traded a bad neighborhood for a worse one: the kind where the laws of mathematics themselves guarantee failure.&lt;/p&gt;

&lt;p&gt;The Fatal Mathematical Doom of the New AI Architectures: They Still See a Flat World&lt;br&gt;
And now, get ready, dear reader, to learn what you won’t read anywhere else. How the mathematical poison is silently killing every AI architecture — from the LLM dinosaurs to the new kids who swear they’ll fix everything.&lt;/p&gt;

&lt;p&gt;Stop. Breathe 😄 This is where the story gets dangerous.&lt;/p&gt;

&lt;p&gt;But don’t worry, we have the antidote. And you can use it to your own advantage.&lt;/p&gt;

&lt;p&gt;Before you go on, a small recommendation.&lt;br&gt;
If you want to dive deeper - into the gritty, nerdy, beautiful mathematics that prove beyond doubt how the dual-number toroidal model cures hallucinations and the myopic doom haunting today’s AI architectures - I’ve left a few doors open for you:&lt;/p&gt;

&lt;p&gt;→ &lt;a href="https://josecrespo.substack.com/p/mathematical-core-equations-ai-jepas" rel="noopener noreferrer"&gt;[JEPA-AI: Core Technologies &amp;amp; Programming Stack]&lt;/a&gt;&lt;br&gt;
→ &lt;a href="https://josecrespo.substack.com/p/the-mathematical-myopia-of-new-ai" rel="noopener noreferrer"&gt;The Mathematical Myopia of New AI Architectures&lt;/a&gt;&lt;br&gt;
→ &lt;a href="https://josecrespo.substack.com/p/mathematical-core-equations-ai-jepas" rel="noopener noreferrer"&gt;Mathematical Core Equations: AI JEPA’s Failures vs. AI-Toroidal Truth&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Take your time there if you wish.&lt;br&gt;
If not, stay with us.&lt;br&gt;
We’re about to cross the event horizon.&lt;/p&gt;

&lt;p&gt;Now, let’s go on.&lt;/p&gt;

&lt;p&gt;Imagine you’re trying to organize your massive music library. You could throw all your songs into one giant folder and hope your computer can tell the difference between “Stairway to Heaven” and “Highway to Hell” based on… vibes? That’s basically what JEPA and the new AI architectures are doing.&lt;/p&gt;

&lt;p&gt;Here’s the problem: these systems live in what mathematicians call “Euclidean space”, basically a flat, infinite spreadsheet where everything is a bunch of numbers floating around. Sounds reasonable, right? Wrong.&lt;/p&gt;

&lt;p&gt;This is why you’ll find the same mathematical doom baked right into the so-called “next generation” AI systems: the ones sold as the antidote to LLM poison.&lt;br&gt;
They promise salvation but inherit the same broken math.&lt;/p&gt;

&lt;h2&gt;
  
  
  — Welcome again to the Hall of AI Shame. Here they are. —
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The Birthday Party Disaster&lt;/strong&gt;&lt;br&gt;
You know the birthday paradox? Put 23 random people in a room, and there’s a 50% chance two share a birthday. Now imagine your AI has to store a million concepts in this flat number-space. The math says: collision guaranteed. “White truck” ends up looking almost identical to “bright sky” because they landed in nearly the same spot in this giant number soup, and there’s how you got those self-driving Tesla killer cars.&lt;/p&gt;

&lt;p&gt;It’s like trying to organize a library by randomly throwing books into a warehouse and hoping you can find them later by remembering their approximate coordinates… you can’t.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Gradient Descent Hamster Wheel&lt;/strong&gt;&lt;br&gt;
Current AI architectures use something called “gradient descent” to find the minimum error given an error function, which is a fancy way of saying they stumble around in the dark, poking things with a stick, hoping to eventually find the exit.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fluyxfmd8wximf43a20v7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fluyxfmd8wximf43a20v7.png" alt=" " width="716" height="637"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The problem? They use fake infinitesimals with point-wise myopic vision. They can’t see the shape of the hill they’re trying to climb down, just one pebble at a time. It’s like trying to navigate San Francisco with a blindfold and a magnifying glass that only shows you one square inch of sidewalk.&lt;/p&gt;

&lt;p&gt;But wait, it gets dumber: you have to pick your epsilon (the step size) before you start stumbling. Pick it too big? You’re that drunk guy at a party taking huge steps and crashing into walls. Too small? You’re inching forward like a paranoid snail, and you’ll die of old age before you get anywhere. Yup, this whole mess comes from 19th-century calculus and its epsilon–delta limit formalism.&lt;/p&gt;

&lt;p&gt;But the craziest thing of all happens during training: AI runs billions of these tiptoeing optimization steps trying to minimize its loss function. Billions! Each one either jumps off the cliff like a buggy robot or advances like Windows loading at 1%. The computational waste is absolutely bonkers - all because this outdated framework, born in the 19th century, forces you to pick an epsilon value beforehand.&lt;/p&gt;

&lt;p&gt;The second error caused by this outdated way of doing infinitesimal calculus is the compounding effect of tiny approximation errors. You start with something like 10^-8 and think, “eh, close enough to zero, right?” Wrong. Square it and you get 10^-16. Still. Not. Zero. After billions of iterations, these pretend infinitesimals pile up like compound interest from hell, spawning numerical explosions, instabilities, and rounding errors that eventually turn into full-blown AI hallucinations.&lt;/p&gt;

&lt;p&gt;Yup, there is an easy solution: switch to a dual-number system that laughs at this entire clown show. No limits. No epsilon guessing games. No billion-step hamster wheel. When ε² = 0 by definition, not approximation, actual mathematical law: derivatives are exact, and topology just tells you where everything belongs. No stumbling required.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Attention Apocalypse&lt;/strong&gt;&lt;br&gt;
Transformers (the tech behind ChatGPT, not the robots) use something called “attention” where every word looks at every other word. That’s N-squared complexity: which means if you double your text length, the computation goes up 4x.&lt;/p&gt;

&lt;p&gt;With 1000 words? That’s a million comparisons. With 10000 words? 100 million comparisons. Your AI is basically reading a book by comparing every single word to every other word simultaneously. Exhausting and expensive.&lt;/p&gt;

&lt;p&gt;How Our Toroidal Model Fixes the AI Flatland Doom&lt;br&gt;
Instead of a flat spreadsheet, we use a donut (mathematically, a torus). Stay with me here.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft57qx68qs535wqxfs4yg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft57qx68qs535wqxfs4yg.png" alt=" " width="721" height="601"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;On a donut, you can wrap string around it in different ways: around the hole, through the hole, or both. These “winding patterns” give every concept a unique address that cannot collide. It’s not probability, it’s topology. Different winding patterns are as different as a circle and a figure-8. They literally cannot become each other.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Infinitesimals
&lt;/h2&gt;

&lt;p&gt;We use dual numbers where ε² = 0 isn’t an approximation - it’s the definition. This means our layers are separated by actual infinitesimals, not fake ones. No numerical explosions. No gradient descent needed. The topology just… works.&lt;/p&gt;

&lt;p&gt;Sparse by Design&lt;br&gt;
Our attention mechanism only connects compatible winding patterns. Most connections are exactly zero - not “close to zero” but structurally impossible. This drops complexity from N-squared to linear. That 100 million comparisons? Now it’s 10,000.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Bottom Line&lt;/strong&gt;&lt;br&gt;
JEPA and these new AI architectures are the right idea to replace our very primitive state towards truly AI. But still, like LLMs, are repeating the same error by trying to navigate a flat world with a bad compass and approximations.&lt;/p&gt;

&lt;p&gt;The real leap won’t come from another tweak in parameters.&lt;br&gt;
It will come from changing the space itself.&lt;/p&gt;

&lt;p&gt;We need to abandon the Euclidean coffin that traps intelligence in two dimensions, and build in a topology where meaning can breathe.&lt;br&gt;
In our toroidal model, concepts don’t collide or blur.They live in separate, protected neighborhoods: each one infinitesimally distinct, each one safe from the chaos of false merges.&lt;/p&gt;

&lt;p&gt;Why Toroidal AI Is Not Being Built — Yet&lt;br&gt;
Now, the sharp reader may ask:&lt;br&gt;
“If this is so obvious, why hasn’t anyone built it yet?”&lt;/p&gt;

&lt;p&gt;Fair question - and the answer itself proves the point.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Institutional Inertia&lt;br&gt;
The entire AI establishment has sunk trillions into Euclidean architectures. Every framework, every GPU kernel, every optimization routine assumes a flat world. Replacing that geometry would mean rebuilding the cathedral from its foundations - and few engineers dare shake the pillars of their own temple.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The Workforce Barrier&lt;br&gt;
A full generation of machine-learning engineers has been trained to think in gradients, not geometry. Retraining them to reason with curvature, continuity, and dual numbers is not a weekend tutorial - it’s a civilizational shift in mathematical literacy.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Patents and IP Locks&lt;br&gt;
Big Tech doesn’t innovate; it defends its moat. Any true geometric paradigm shift would invalidate entire layers of existing intellectual property and licensing chains. The system isn’t optimized for truth - it’s optimized for control.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The Sunk-Cost Fallacy&lt;br&gt;
From cloud infrastructure to AI chips, everything has been built for the flatland paradigm. Even when engineers know it’s broken, the machinery keeps running - because admitting it would collapse too many balance sheets and too many egos.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;So yes - it hasn’t been built yet.&lt;br&gt;
Not because it’s wrong.&lt;br&gt;
But because it’s too right, too disruptive, and too costly for a world addicted to its own mistakes.&lt;/p&gt;

&lt;p&gt;And that’s precisely why it will happen.&lt;br&gt;
Because math doesn’t care who resists it.&lt;br&gt;
It simply wins - always.&lt;/p&gt;

&lt;p&gt;And soon, the new startups will notice the gap. You’ll watch Toroidal AI evolve exactly as every disruptive technology has before it:&lt;br&gt;
First ignored - too threatening to a trillion dollars in sunk investments.&lt;br&gt;
Then ridiculed “crackpot” accusations from people who don’t understand topology.&lt;br&gt;
And finally, triumphantly accepted “Of course! We always knew Euclidean space was wrong.”&lt;/p&gt;

&lt;p&gt;History doesn’t repeat itself.&lt;br&gt;
It curves. 😁&lt;/p&gt;

&lt;h2&gt;
  
  
  Top 10 Essential References
&lt;/h2&gt;

&lt;p&gt;On JEPA Architectures:&lt;br&gt;
1.- LLM-JEPA: Large Language Models Meet Joint Embedding Predictive Architectures (September 2025)&lt;/p&gt;

&lt;p&gt;2.- ACT-JEPA: Novel Joint-Embedding Predictive Architecture for Efficient Policy Representation Learning (April 2025)&lt;/p&gt;

&lt;p&gt;3.- Point-JEPA: A Joint Embedding Predictive Architecture for Self-Supervised Learning on Point Cloud (February 2025)&lt;/p&gt;

&lt;p&gt;On Transformer Attention Complexity:&lt;br&gt;
4.- The End of Transformers? On Challenging Attention and the Rise of Sub-Quadratic Architectures (October 2024)&lt;/p&gt;

&lt;p&gt;5.- FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning (October 2023, still widely cited in 2024–2025)&lt;/p&gt;

&lt;p&gt;On Toroidal/Topological Neural Networks:&lt;br&gt;
6.- Toroidal Topology of Population Activity in Grid Cells (January 2022, Nature — still foundational for 2024–25 work)&lt;/p&gt;

&lt;p&gt;7.- Deep Networks on Toroids: Removing Symmetries Reveals the Structure of Flat Regions in the Landscape Geometry (June 2022)&lt;/p&gt;

&lt;p&gt;On Dual Numbers &amp;amp; Automatic Differentiation:&lt;br&gt;
8-. Dual Numbers for Arbitrary Order Automatic Differentiation (January 2025)&lt;/p&gt;

&lt;p&gt;On LLM Hallucinations:&lt;br&gt;
9.- Why Language Models Hallucinate (September 2025)&lt;/p&gt;

&lt;p&gt;Bonus - Yann LeCun’s Vision:&lt;br&gt;
10.- Navigation World Models (April 2025)&lt;/p&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>mathematic</category>
      <category>llm</category>
    </item>
    <item>
      <title>Why the Oldest, Simplest Algorithms Are Beating AI</title>
      <dc:creator>Jose Crespo, PhD</dc:creator>
      <pubDate>Sat, 01 Nov 2025 11:12:45 +0000</pubDate>
      <link>https://forem.com/jose_crespo_phd/why-the-oldest-simplest-algorithms-are-beating-ai-1n13</link>
      <guid>https://forem.com/jose_crespo_phd/why-the-oldest-simplest-algorithms-are-beating-ai-1n13</guid>
      <description>&lt;p&gt;Let’s put numbers to the AI overhype. Companies are burning more than $200 billion every year by choosing AI over simple, proven algorithms. That’s not an opinion. That’s math. See the chart below if you’re still rubbing your eyes in disbelief.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1huqxxzsdk4qv5ft3wym.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1huqxxzsdk4qv5ft3wym.png" alt="Chart 1 – The AI machine burning more than $200B yearly" width="800" height="469"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Chart 1: The numbers behind the $200B conservative burn.&lt;/strong&gt; Independent studies point the same way: MIT reports &lt;strong&gt;95% of GenAI implementations show no measurable P&amp;amp;L impact&lt;/strong&gt; (&lt;a href="https://www.tomshardware.com/tech-industry/artificial-intelligence/95-percent-of-generative-ai-implementations-in-enterprise-have-no-measurable-impact-on-p-and-l-says-mit-flawed-integration-key-reason-why-ai-projects-underperform" rel="noopener noreferrer"&gt;Tom’s Hardware coverage&lt;/a&gt;), BCG finds &lt;strong&gt;74% of companies struggle to achieve and scale value&lt;/strong&gt; (&lt;a href="https://www.bcg.com/press/24october2024-ai-adoption-in-2024-74-of-companies-struggle-to-achieve-and-scale-value" rel="noopener noreferrer"&gt;BCG press release&lt;/a&gt;), HBR documents &lt;strong&gt;2.5–14% profit drag&lt;/strong&gt; from AI misuse (&lt;a href="https://hbr.org" rel="noopener noreferrer"&gt;Harvard Business Review&lt;/a&gt;), and CFO surveys show &lt;strong&gt;6–16% gross-margin hits&lt;/strong&gt; (&lt;a href="https://www.cfodive.com/news/one-in-four-firms-miss-ai-cost-projections-50percent-or-more-survey/760197/" rel="noopener noreferrer"&gt;CFO Dive&lt;/a&gt;). Brookings adds the macro verdict: &lt;strong&gt;no measurable productivity gains to date&lt;/strong&gt; (&lt;a href="https://www.brookings.edu" rel="noopener noreferrer"&gt;Brookings&lt;/a&gt;). Against an &lt;a href="https://www.idc.com" rel="noopener noreferrer"&gt;IDC&lt;/a&gt; baseline of &lt;strong&gt;~$307B in 2025 AI spend&lt;/strong&gt;, even conservative failure/no-value multipliers imply &lt;strong&gt;≈$200B/year&lt;/strong&gt; in burned spend that simpler algorithms could have avoided. &lt;em&gt;Plot created by the author using Matplotlib library.&lt;/em&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  But here’s the question: Is AI the villain — or are CEOs burning money on systems they barely understand?
&lt;/h2&gt;

&lt;p&gt;The truth is more brutal than either narrative. &lt;a href="https://medium.com/ai-advances/ai-is-producing-more-garbage-code-than-ever-97a37104e22c?sk=80d1da92b6a3a5e0de1f004b40613d9e" rel="noopener noreferrer"&gt;AI isn’t the disaster — it’s the excuse for the disaster&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Look at the chart below. This is your American Horror Story in three acts:&lt;/p&gt;

&lt;p&gt;&lt;u&gt;&lt;strong&gt;Act I&lt;/strong&gt;&lt;/u&gt;: The Promise Solo AI vendors promise revolutionary decision-making. Fortune 500 CEOs, hypnotized by NVIDIA’s stock price and OpenAI’s demos, write checks their companies can’t cash.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;u&gt;Act II&lt;/u&gt;&lt;/strong&gt;: The Bloodbath That red line? That’s $500 billion annually achieving -45% ROI (Return on Investment — how much you get back compared to what you spend). Not a typo. Negative forty-five percent. Companies are paying premium prices to make worse decisions slower. It’s like hiring McKinsey to tell you what you already know, except McKinsey occasionally delivers positive value.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;u&gt;Act III&lt;/u&gt;&lt;/strong&gt;: The Coverup Nobody admits failure. The AI system becomes “strategic infrastructure.” The losses get buried in “digital transformation.” The vendors — Anthropic, OpenAI, the whole choir — keep singing about the future while cashing checks in the present.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fad2aifdnkk4wivfbry9b.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fad2aifdnkk4wivfbry9b.png" alt="Chart 2 – ROI of simple algorithms vs solo AI vs hybrid AI" width="800" height="506"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Chart 2: The $2.2 Trillion Reality Check.&lt;/strong&gt; This is what burning money looks like in three colors. The red line isn’t a projection — it’s &lt;a href="https://spectrum.ieee.org/how-ibm-watson-overpromised-and-underdelivered-on-ai-health-care" rel="noopener noreferrer"&gt;IBM Watson’s $4 billion loss&lt;/a&gt;, &lt;a href="https://www.cnbc.com" rel="noopener noreferrer"&gt;Zillow’s $500M quarterly disaster&lt;/a&gt;, and hundreds of other AI failures achieving a median &lt;strong&gt;-45% ROI&lt;/strong&gt; (&lt;a href="https://www.gartner.com/en/documents/4598199" rel="noopener noreferrer"&gt;Gartner 2023&lt;/a&gt;). Meanwhile, that green line soaring at 1200%? That’s &lt;a href="https://investor.costco.com/financials/annual-reports-and-proxy-statements/default.aspx" rel="noopener noreferrer"&gt;Costco’s one-line markup rule&lt;/a&gt;, Southwest’s 47-year profit streak, and algorithms older than your grandparents still crushing it (&lt;a href="https://info.mheducation.com/rs/128-SJW-347/images/Preface_Hillier_Intro_Operations_Research_11e.pdf" rel="noopener noreferrer"&gt;Operations Research, Hillier &amp;amp; Lieberman&lt;/a&gt;). The blue hybrid line shows the only scenario where AI doesn’t destroy value — when it’s kept on a tight leash by old-school algorithms. Even then, &lt;strong&gt;450%&lt;/strong&gt; is less than half what simple algorithms achieve alone (&lt;a href="https://www.bcg.com/publications/2023/how-people-create-and-destroy-value-with-gen-ai" rel="noopener noreferrer"&gt;MIT–BCG Study 2023&lt;/a&gt;). &lt;em&gt;Plot created by the author using Matplotlib library.&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;Remember the California Gold Rush from the history books? You know who got rich in 1849. Not the miners. The pickaxe sellers.&lt;/p&gt;

&lt;p&gt;Today’s pickaxe merchants sell GPU clusters and API tokens to CEOs panning for digital gold. They whisper sweet promises: “AI will revolutionize your business.” What they don’t mention is that graph showing AI never achieves positive ROI for operational decisions.&lt;/p&gt;

&lt;p&gt;Instead they boast about having more billion parameters than the competition’s AI monster. Of course theirs is bigger, more insanely convoluted — an expanding minefield of compounding errors, &lt;a href="https://medium.com/ai-advances/the-math-that-cures-ai-blindness-3aa99b05a3f2?sk=dfb2db83bc16ec154460e12388673fe9" rel="noopener noreferrer"&gt;Types I through IV, each cascade multiplying the disaster&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Meanwhile, see that green line in the chart above — simple algorithms from 1913 — keeps printing money at 1200% ROI. But nobody’s selling conferences about the Economic Order Quantity formula. There’s no TED talk about Little’s Law and the other older algorithms.&lt;/p&gt;

&lt;p&gt;Yep, unfortunately you can’t raise billions in funding rounds by selling an algorithm that fits on a napkin from that expensive restaurant where you’re shaking hands with investors interested in a bigger AI beast.🤔&lt;/p&gt;

&lt;p&gt;The tragedy is more nuanced than AI sellers admit — and more complex than CEOs with their broken quarterly-profit-maximization algorithms want to hear.&lt;/p&gt;

&lt;p&gt;Yes, there’s a hybrid approach where AI becomes just one component alongside those ancient algorithms when complexity grows. But that requires two things the AI revolution explicitly avoids:&lt;/p&gt;

&lt;p&gt;First, you need to understand your business from the molecular level to the stratosphere — every intricacy of your model, what you’re actually selling versus what you think you’re selling, where costs hide, where value emerges. These insights come from human minds. AI won’t solve what you can’t articulate, despite the sales pitch.&lt;/p&gt;

&lt;p&gt;Second, you need to hire and respect professionals who combine programming excellence with mathematical rigor and the rare ability to translate both into business value. But your AI recruiting system will never surface these people. It’s optimized for commodities, not talent — screening for keywords, not capability.&lt;/p&gt;

&lt;p&gt;Here’s the brutal reality with a massively failed AI that most companies are facing right now:&lt;br&gt;
Look at the carnage in the table below (Chart 3) . This isn’t speculation — it’s documented history:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Costco&lt;/strong&gt;: Running on a decision that fits on an index card. Investment: approximately zero. Return: Industry dominance for four decades.&lt;br&gt;
IBM Watson Health: The crown jewel of AI healthcare. Investment: $4 billion. Return: Sold for parts at a 95% loss.&lt;br&gt;
&lt;strong&gt;Southwest Airlines&lt;/strong&gt;: One simple rule about load factors. Still the only major US airline to avoid bankruptcy.&lt;br&gt;
&lt;strong&gt;Zillow iBuying&lt;/strong&gt;: Cutting-edge AI pricing models. Burned $500 million in one quarter before shutting down entirely.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhlg7bw98pty2vzbimw2z.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhlg7bw98pty2vzbimw2z.png" alt="Chart 3 – ROI table showing algorithmic vs AI vs hybrid approaches" width="800" height="480"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Chart 3: The companies destroying their competition aren’t using AI&lt;/strong&gt; — they’re using formulas older than your grandparents. That $4 billion IBM lost? Costco made 100× that with a rule that fits on a Post-it. Look at the ROI column. Now look again. That’s not a typo. For references see Chart 2 above. &lt;em&gt;Table created by the author using Plotly.&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;The only success story with AI? Amazon’s hybrid approach — yup, not pure AI, it’s old-school EOQ with some ML sprinkled on top when absolutely necessary. Even then, the ROI is a fraction of what simple algorithms achieve alone.&lt;/p&gt;

&lt;p&gt;But let’s be honest here, let’s not throw the baby out with the bathwater: when your business complexity genuinely grows — meaning the actual number of variables you must account for — you need more flexibility. That might be AI/ML, or more likely, it’s just good programmers who understand your business and can architect the math you actually need.&lt;/p&gt;

&lt;p&gt;Here’s what 20 years in the trenches taught most of us: around 90% of “complex” business problems can be solved by a competent programmer with decent math skills and simple algorithms, properly wired together.&lt;/p&gt;

&lt;p&gt;Occam’s Razor still cuts: The simplest solution that works is usually right.&lt;/p&gt;

&lt;p&gt;But simplicity doesn’t sell conference tickets. Simplicity doesn’t raise Series B funding. Simplicity doesn’t get you on the cover of Wired. 😂&lt;/p&gt;

&lt;p&gt;So instead, companies take the lazy, prestigious route: throw their problems at whatever AI the vendors are pushing this quarter. It’s like hiring a top surgeon to apply a Band-Aid — expensive, unnecessary, and probably going to make things worse.&lt;/p&gt;

&lt;p&gt;The real tragedy? That competent programmer with the simple solution costs $150K/year. The AI system that fails costs $15M. But the programmer doesn’t come with a sales team, a PowerPoint deck, or a promise to “transform your digital future.”&lt;/p&gt;

&lt;p&gt;Guess which one the board approves.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Seven Heroes of Business History
&lt;/h2&gt;

&lt;p&gt;Our position shouldn’t surprise you by now, dear reader. We’re betting on what’s worked brilliantly for decades through every crisis, disaster, and market shift while delivering massive profits: the algorithm way.&lt;/p&gt;

&lt;p&gt;The formula is simple: Hire, value, and respect your best asset — &lt;a href="https://medium.com/ai-advances/the-maths-you-need-to-survive-ai-and-nobodys-teaching-you-9167a2ccd885?sk=a3989896ff7822a627b3022f6d1570f8" rel="noopener noreferrer"&gt;skilled programmers with solid math foundations&lt;/a&gt; who convert business problems into algorithmic solutions. Add modern cloud infrastructure and, when genuinely appropriate (maybe 5–10% of cases), deploy the hybrid Algorithm + AI/ML approach.&lt;/p&gt;

&lt;p&gt;But let’s be clear: forget the magic they’re selling — the fantasy of throwing data at AI and getting perfect solutions every time. That’s not happening.&lt;/p&gt;

&lt;p&gt;Yes, LLMs are useful as search tools, data navigators, even coding assistants. But they’re far from autonomous. In non-trivial cases, you need spend lot of time cross-referencing to catch their false positives and, worse, the errors nobody notices — Type III and IV errors that spaghettify and collide concepts into dangerous nonsense.&lt;/p&gt;

&lt;p&gt;Meet the Magnificent Seven&lt;br&gt;
More than 500 years of combined service. Trillions in value created. Under 1K lines of code total.&lt;/p&gt;

&lt;p&gt;Look at this table (Chart 4). These seven algorithms have never failed. For decades — in some cases over a century — they’ve consistently delivered business value to most of mankind. The epitome of simplicity.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5fvu6kcfphp8wwe4r4sf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5fvu6kcfphp8wwe4r4sf.png" alt="Chart 4 – Timeline of the most profitable algorithms ever invented" width="800" height="608"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Chart 4: The Magnificent 7.&lt;/strong&gt; Still blooming today, hidden inside most business software and powering modern networks. They quietly run the world’s critical domains — inventory, finance, perishables, risk, projects, operations, information — and they keep compounding value. Think about it: 468 years of proven success, trillions of dollars created, and all it takes is less than 1K lines of code. A handful of math functions generating returns that outshine most of today’s AI hype. This isn’t the full list — just a curated glimpse at the most profitable algorithms ever invented. Reference: &lt;a href="https://arxiv.org/pdf/2407.13839" rel="noopener noreferrer"&gt;Algorithms and ROI&lt;/a&gt;. &lt;em&gt;Plot created by the author using Matplotlib library. Table created by the author using Plotly.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Each one is a specialist with embarrassingly simple code:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;EOQ (1913): The inventory gunslinger&lt;/li&gt;
&lt;li&gt;Square root of (2 × demand × order cost / holding cost). One line. Tells you exactly how much to order.&lt;/li&gt;
&lt;li&gt;DuPont (1920): The financial sharpshooter&lt;/li&gt;
&lt;li&gt;ROE = Profit Margin × Asset Turnover × Leverage. Three numbers multiplied. Instant diagnosis of what’s broken.&lt;/li&gt;
&lt;li&gt;Newsvendor (1950s): The perishables ranger&lt;/li&gt;
&lt;li&gt;Order up to the point where the chance of selling out matches the cost of running short vs. overstocking. A single threshold for “how much to make.”&lt;/li&gt;
&lt;li&gt;Kelly (1956): The risk-sizing marshal&lt;/li&gt;
&lt;li&gt;Bet a fraction of your bankroll based on edge and odds — when your advantage is bigger, size up; when it’s small, size down. Never overbet.&lt;/li&gt;
&lt;li&gt;CPM (1957): The project management tracker&lt;/li&gt;
&lt;li&gt;Find the longest path through your network. That’s your deadline. Everything else can slip; this can’t.&lt;/li&gt;
&lt;li&gt;Little’s Law (1961): The operations enforcer&lt;/li&gt;
&lt;li&gt;Items in system = arrival rate × time in system. It’s physics, not statistics. Works for everything.&lt;/li&gt;
&lt;li&gt;PageRank (1998): The young gun who built an empire&lt;/li&gt;
&lt;li&gt;A page’s importance is the sum of votes from important pages, each vote split by their outlinks, with a small random-jump factor to keep it stable. Keep iterating. Built Google.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;Total code for all seven: comfortably &amp;lt;1K lines. The math that runs the world fits in a single GitHub gist.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;u&gt;&lt;strong&gt;Is a programming career still viable? Looking at this table — absolutely. Learning algorithms and programming built the past and will build the future, despite this transient AI-hypnotized present.&lt;/strong&gt;&lt;/u&gt;&lt;/p&gt;

&lt;p&gt;The AI hype has its place — pattern finding, searching, identifying. But it needs years of improvement to correct its statistical errors from Type I through IV. Even then, algorithms in the hands of competent developers remain irreplaceable and far from commodities.&lt;/p&gt;

&lt;p&gt;The bottom line: While everyone chases AI complexity, these seven simple formulas keep generating thousands of percent returns. They don’t need updates. They don’t hallucinate. They just work.&lt;/p&gt;

&lt;p&gt;That’s not outdated. That’s immortal. Let’s keep programming with algorithms while the true researchers work on fixing AI errors with better math.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>algorithms</category>
      <category>machinelearning</category>
    </item>
  </channel>
</rss>
