<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: H33.ai</title>
    <description>The latest articles on Forem by H33.ai (@h33ai).</description>
    <link>https://forem.com/h33ai</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1023869%2Fb3b3e26e-8be6-45c9-8641-2a9634940365.png</url>
      <title>Forem: H33.ai</title>
      <link>https://forem.com/h33ai</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/h33ai"/>
    <language>en</language>
    <item>
      <title>Building a Sub-Microsecond Cache for a Billion-User Mining Platform</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:57:22 +0000</pubDate>
      <link>https://forem.com/h33ai/building-a-sub-microsecond-cache-for-a-billion-user-mining-platform-4j67</link>
      <guid>https://forem.com/h33ai/building-a-sub-microsecond-cache-for-a-billion-user-mining-platform-4j67</guid>
      <description>&lt;h2&gt;
  
  
  The Problem: 100-250ms Middleware Tax
&lt;/h2&gt;

&lt;p&gt;Our Node.js Express backend had a dirty secret: before any route handler ran, the middleware stack consumed 100-250ms. Four separate Redis round-trips for rate limiting, session load, XSS sanitization — all serial.&lt;/p&gt;

&lt;p&gt;For a platform targeting billions of users, this was unacceptable.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Solution: CacheeEngine in Rust
&lt;/h2&gt;

&lt;p&gt;We replaced the entire JS middleware stack with an in-process Rust cache engine called CacheeEngine:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;What It Does&lt;/th&gt;
&lt;th&gt;Memory&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;CacheeLFU eviction&lt;/td&gt;
&lt;td&gt;Frequency counters with periodic halving decay&lt;/td&gt;
&lt;td&gt;Inline&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Count-Min Sketch&lt;/td&gt;
&lt;td&gt;Admission doorkeeper — rejects one-hit-wonders&lt;/td&gt;
&lt;td&gt;512 KiB constant&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;DashMap storage&lt;/td&gt;
&lt;td&gt;Lock-free concurrent reads&lt;/td&gt;
&lt;td&gt;O(entries)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;SWR support&lt;/td&gt;
&lt;td&gt;Stale-while-revalidate built in&lt;/td&gt;
&lt;td&gt;Built-in&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h3&gt;
  
  
  CacheeLFU vs W-TinyLFU
&lt;/h3&gt;

&lt;p&gt;We chose CacheeLFU over W-TinyLFU (Caffeine/moka):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No window cache&lt;/li&gt;
&lt;li&gt;Direct frequency comparison with CMS admission&lt;/li&gt;
&lt;li&gt;Periodic halving decay instead of reset-on-epoch&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Count-Min Sketch: 512 KiB That Saves Everything
&lt;/h3&gt;

&lt;p&gt;The CMS is the doorkeeper. Layout: 4 rows x 131,072 counters x 1 byte = 512 KiB. Constant regardless of entry count.&lt;/p&gt;

&lt;h2&gt;
  
  
  Results
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Path&lt;/th&gt;
&lt;th&gt;JS (Express)&lt;/th&gt;
&lt;th&gt;Rust (Axum + Cachee)&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Middleware&lt;/td&gt;
&lt;td&gt;100-250ms&lt;/td&gt;
&lt;td&gt;&amp;lt;5ms&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Trust score hit&lt;/td&gt;
&lt;td&gt;5-10ms&lt;/td&gt;
&lt;td&gt;&amp;lt;1us&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Mining sync&lt;/td&gt;
&lt;td&gt;50ms-2s&lt;/td&gt;
&lt;td&gt;&amp;lt;10ms&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Swap quote stale&lt;/td&gt;
&lt;td&gt;40-150ms&lt;/td&gt;
&lt;td&gt;&amp;lt;1us&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Rate limit&lt;/td&gt;
&lt;td&gt;30-50ms&lt;/td&gt;
&lt;td&gt;&amp;lt;100ns&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The binary is 5.2MB stripped. 15 tests pass.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://revmine.ai" rel="noopener noreferrer"&gt;RevMine&lt;/a&gt; is live. Check &lt;a href="https://github.com/H33ai-postquantum" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt; for open-source components.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Built on Cachee — H33 post-quantum cache engine with CacheeLFU eviction.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>rust</category>
      <category>performance</category>
      <category>caching</category>
      <category>webdev</category>
    </item>
    <item>
      <title>Why Loyalty Points Are Broken (And What Revenue-Backed Tokens Fix)</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:57:16 +0000</pubDate>
      <link>https://forem.com/h33ai/why-loyalty-points-are-broken-and-what-revenue-backed-tokens-fix-4mmm</link>
      <guid>https://forem.com/h33ai/why-loyalty-points-are-broken-and-what-revenue-backed-tokens-fix-4mmm</guid>
      <description>&lt;h2&gt;
  
  
  The $300B Loyalty Problem
&lt;/h2&gt;

&lt;p&gt;Traditional loyalty points have a &lt;strong&gt;54% abandonment rate&lt;/strong&gt;. Customers earn points, forget about them, and churn.&lt;/p&gt;

&lt;p&gt;What if loyalty points were actually &lt;em&gt;worth something&lt;/em&gt;?&lt;/p&gt;

&lt;h2&gt;
  
  
  Revenue-Backed Tokens
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://revmine.ai" rel="noopener noreferrer"&gt;RevMine&lt;/a&gt; replaces static points with revenue-backed Solana tokens. When a business grows, the tokens grow. This creates real switching costs — 25-40% churn reduction measured across early adopters.&lt;/p&gt;

&lt;h3&gt;
  
  
  How It Works
&lt;/h3&gt;

&lt;ol&gt;
&lt;li&gt;Business embeds a widget (2 lines of code)&lt;/li&gt;
&lt;li&gt;Users mine tokens by engaging&lt;/li&gt;
&lt;li&gt;Mining rates are AI-optimized&lt;/li&gt;
&lt;li&gt;Tokens backed by revenue via Stripe verification&lt;/li&gt;
&lt;li&gt;On-chain settlement on Solana&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  The Tech
&lt;/h3&gt;

&lt;p&gt;The backend is 100% Rust:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Sub-15ms API latency (Axum + tokio)&lt;/li&gt;
&lt;li&gt;CacheeLFU in-process cache with Count-Min Sketch (512 KiB constant)&lt;/li&gt;
&lt;li&gt;Post-quantum encryption via H33/NIST FIPS 204&lt;/li&gt;
&lt;li&gt;Offline-first WASM mining engine&lt;/li&gt;
&lt;li&gt;Single SQL transaction for mining validation&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  For SaaS Founders
&lt;/h3&gt;

&lt;p&gt;If you are fighting churn:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Free tier, no risk&lt;/li&gt;
&lt;li&gt;5-minute setup, widget embed&lt;/li&gt;
&lt;li&gt;No crypto knowledge required for users&lt;/li&gt;
&lt;li&gt;White-label, your brand&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://revmine.ai/token-wizard/" rel="noopener noreferrer"&gt;Try the token wizard&lt;/a&gt; to see token creation.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Built with Rust, Solana, and Cachee post-quantum caching.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>saas</category>
      <category>blockchain</category>
      <category>rust</category>
      <category>webdev</category>
    </item>
    <item>
      <title>20 Rust Microservices, Zero Node.js: How We Rebuilt Our Entire Video Platform</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:22 +0000</pubDate>
      <link>https://forem.com/h33ai/20-rust-microservices-zero-nodejs-how-we-rebuilt-our-entire-video-platform-4okg</link>
      <guid>https://forem.com/h33ai/20-rust-microservices-zero-nodejs-how-we-rebuilt-our-entire-video-platform-4okg</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/20-rust-microservices-zero-nodejs.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/20-rust-microservices-zero-nodejs.html" rel="noopener noreferrer"&gt;v100.ai/blog/20-rust-microservices-zero-nodejs&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>rust</category>
      <category>microservices</category>
      <category>architecture</category>
      <category>video</category>
    </item>
    <item>
      <title>10 Microseconds: How We Built the Fastest API Gateway in Rust</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:20 +0000</pubDate>
      <link>https://forem.com/h33ai/10-microseconds-how-we-built-the-fastest-api-gateway-in-rust-4o60</link>
      <guid>https://forem.com/h33ai/10-microseconds-how-we-built-the-fastest-api-gateway-in-rust-4o60</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/10-microsecond-api-gateway.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/10-microsecond-api-gateway.html" rel="noopener noreferrer"&gt;v100.ai/blog/10-microsecond-api-gateway&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>rust</category>
      <category>performance</category>
      <category>api</category>
      <category>webdev</category>
    </item>
    <item>
      <title>Rust vs C++ for Video Server Performance: Why We Chose Rust</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:19 +0000</pubDate>
      <link>https://forem.com/h33ai/rust-vs-c-for-video-server-performance-why-we-chose-rust-351f</link>
      <guid>https://forem.com/h33ai/rust-vs-c-for-video-server-performance-why-we-chose-rust-351f</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/rust-vs-cpp-video-server-performance.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/rust-vs-cpp-video-server-performance.html" rel="noopener noreferrer"&gt;v100.ai/blog/rust-vs-cpp-video-server-performance&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>rust</category>
      <category>cpp</category>
      <category>performance</category>
      <category>video</category>
    </item>
    <item>
      <title>Fastest WebRTC Server in 2026: Benchmarks Against coturn and LiveKit</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:17 +0000</pubDate>
      <link>https://forem.com/h33ai/fastest-webrtc-server-in-2026-benchmarks-against-coturn-and-livekit-1mi0</link>
      <guid>https://forem.com/h33ai/fastest-webrtc-server-in-2026-benchmarks-against-coturn-and-livekit-1mi0</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/fastest-webrtc-server-2026.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/fastest-webrtc-server-2026.html" rel="noopener noreferrer"&gt;v100.ai/blog/fastest-webrtc-server-2026&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>webrtc</category>
      <category>rust</category>
      <category>performance</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Sub-Microsecond Video Processing: Inside Our 263ns Pipeline</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:16 +0000</pubDate>
      <link>https://forem.com/h33ai/sub-microsecond-video-processing-inside-our-263ns-pipeline-h92</link>
      <guid>https://forem.com/h33ai/sub-microsecond-video-processing-inside-our-263ns-pipeline-h92</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/sub-microsecond-video-processing.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/sub-microsecond-video-processing.html" rel="noopener noreferrer"&gt;v100.ai/blog/sub-microsecond-video-processing&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>rust</category>
      <category>performance</category>
      <category>video</category>
      <category>api</category>
    </item>
    <item>
      <title>31 Nanoseconds: How Cachee Powers the Fastest Video API Cache</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:14 +0000</pubDate>
      <link>https://forem.com/h33ai/31-nanoseconds-how-cachee-powers-the-fastest-video-api-cache-381m</link>
      <guid>https://forem.com/h33ai/31-nanoseconds-how-cachee-powers-the-fastest-video-api-cache-381m</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/cachee-31-nanosecond-cache.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/cachee-31-nanosecond-cache.html" rel="noopener noreferrer"&gt;v100.ai/blog/cachee-31-nanosecond-cache&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>rust</category>
      <category>caching</category>
      <category>performance</category>
      <category>database</category>
    </item>
    <item>
      <title>How We Built Post-Quantum Encrypted Video Conferencing</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:12 +0000</pubDate>
      <link>https://forem.com/h33ai/how-we-built-post-quantum-encrypted-video-conferencing-23ep</link>
      <guid>https://forem.com/h33ai/how-we-built-post-quantum-encrypted-video-conferencing-23ep</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/post-quantum-encrypted-meetings.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/post-quantum-encrypted-meetings.html" rel="noopener noreferrer"&gt;v100.ai/blog/post-quantum-encrypted-meetings&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>security</category>
      <category>encryption</category>
      <category>rust</category>
      <category>quantum</category>
    </item>
    <item>
      <title>Open-Sourcing RustTURN: Post-Quantum Video Infrastructure</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:11 +0000</pubDate>
      <link>https://forem.com/h33ai/open-sourcing-rustturn-post-quantum-video-infrastructure-15h9</link>
      <guid>https://forem.com/h33ai/open-sourcing-rustturn-post-quantum-video-infrastructure-15h9</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/open-sourcing-rustturn.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/open-sourcing-rustturn.html" rel="noopener noreferrer"&gt;v100.ai/blog/open-sourcing-rustturn&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>rust</category>
      <category>opensource</category>
      <category>webrtc</category>
      <category>video</category>
    </item>
    <item>
      <title>938 Tests, Zero Failures: Why We're the Most Tested Video API</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:09 +0000</pubDate>
      <link>https://forem.com/h33ai/938-tests-zero-failures-why-were-the-most-tested-video-api-i3h</link>
      <guid>https://forem.com/h33ai/938-tests-zero-failures-why-were-the-most-tested-video-api-i3h</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/938-tests-zero-failures.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/938-tests-zero-failures.html" rel="noopener noreferrer"&gt;v100.ai/blog/938-tests-zero-failures&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>rust</category>
      <category>testing</category>
      <category>devops</category>
      <category>codequality</category>
    </item>
    <item>
      <title>Real-Time Video Intelligence: Our AI Pipeline Running at 220K RPS</title>
      <dc:creator>H33.ai</dc:creator>
      <pubDate>Thu, 16 Apr 2026 01:31:07 +0000</pubDate>
      <link>https://forem.com/h33ai/real-time-video-intelligence-our-ai-pipeline-running-at-220k-rps-1jii</link>
      <guid>https://forem.com/h33ai/real-time-video-intelligence-our-ai-pipeline-running-at-220k-rps-1jii</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Originally published at &lt;a href="https://v100.ai/blog/real-time-video-intelligence-220k-rps.html" rel="noopener noreferrer"&gt;v100.ai&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;At &lt;a href="https://v100.ai" rel="noopener noreferrer"&gt;V100&lt;/a&gt;, we build AI video infrastructure entirely in Rust. 20 microservices. 0.01ms server processing. 220,000+ requests per second. Post-quantum encryption on every call.&lt;/p&gt;

&lt;p&gt;This post is a deep dive into our architecture and the engineering decisions behind it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Rust for Video Infrastructure?
&lt;/h2&gt;

&lt;p&gt;Video infrastructure has unique constraints that make Rust the ideal choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Zero GC pauses&lt;/strong&gt; — Real-time video processing can't tolerate garbage collection stops&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory safety&lt;/strong&gt; — Buffer overflows in media pipelines are a security nightmare&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Performance&lt;/strong&gt; — Our gateway processes requests in 10 microseconds, not 10 milliseconds&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Small binaries&lt;/strong&gt; — Our meeting signaling server is a 2MB binary serving WebRTC at scale&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Our Stack
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Component&lt;/th&gt;
&lt;th&gt;Technology&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Web Framework&lt;/td&gt;
&lt;td&gt;Axum + Tokio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;PostgreSQL + TimescaleDB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cache&lt;/td&gt;
&lt;td&gt;Redis + Cachee (in-process)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Media&lt;/td&gt;
&lt;td&gt;FFmpeg (sidecar)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Crypto&lt;/td&gt;
&lt;td&gt;ML-KEM-768 + ML-DSA-65 (post-quantum)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Infra&lt;/td&gt;
&lt;td&gt;Docker + AWS ECS Fargate&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;20 Rust microservices&lt;/strong&gt; — gateway, AI, transcription, video, conferencing, billing, broadcasting&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;0.01ms&lt;/strong&gt; server processing latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;220K+ RPS&lt;/strong&gt; sustained throughput&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;263ns&lt;/strong&gt; pipeline latency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;938 tests&lt;/strong&gt; with zero failures&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;40+ languages&lt;/strong&gt; for real-time transcription&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;7 platforms&lt;/strong&gt; for social publishing (YouTube, TikTok, Instagram, LinkedIn, X, Facebook, Vimeo)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Services
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Gateway&lt;/strong&gt; (Axum) — JWT auth, rate limiting, CSRF/SSRF protection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI Orchestration&lt;/strong&gt; — Claude/Gemini proxy, streaming, compliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transcription&lt;/strong&gt; — Deepgram + Whisper, word-level timestamps, 40+ languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Meeting Signaling&lt;/strong&gt; — WebRTC SDP/ICE, DashMap concurrency, 2MB binary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;v100-turn&lt;/strong&gt; — Full broadcast platform: ABR, DRM, DVR, spatial audio, deepfake detection&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;V100 has a free tier — &lt;a href="https://v100.ai/pricing" rel="noopener noreferrer"&gt;v100.ai/pricing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;API docs at &lt;a href="https://docs.v100.ai" rel="noopener noreferrer"&gt;docs.v100.ai&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Read the full technical deep-dive at &lt;a href="https://v100.ai/blog/real-time-video-intelligence-220k-rps.html" rel="noopener noreferrer"&gt;v100.ai/blog/real-time-video-intelligence-220k-rps&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;V100 is built by &lt;a href="https://h33.ai" rel="noopener noreferrer"&gt;H33.ai&lt;/a&gt; — post-quantum security for production systems.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>rust</category>
      <category>ai</category>
      <category>video</category>
      <category>performance</category>
    </item>
  </channel>
</rss>
