<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Pacy</title>
    <description>The latest articles on Forem by Pacy (@kryptopacy).</description>
    <link>https://forem.com/kryptopacy</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3285413%2F0d1f519e-789a-4d81-95e9-b7c3e98a58d0.png</url>
      <title>Forem: Pacy</title>
      <link>https://forem.com/kryptopacy</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/kryptopacy"/>
    <language>en</language>
    <item>
      <title>Introducing TCG, The Cruise god - the world's first multiplayer voice AI concierge, game master, and local guide.</title>
      <dc:creator>Pacy</dc:creator>
      <pubDate>Fri, 27 Mar 2026 13:10:33 +0000</pubDate>
      <link>https://forem.com/kryptopacy/introducing-tcg-the-cruise-god-the-worlds-first-multiplayer-voice-ai-concierge-game-master-35fm</link>
      <guid>https://forem.com/kryptopacy/introducing-tcg-the-cruise-god-the-worlds-first-multiplayer-voice-ai-concierge-game-master-35fm</guid>
      <description>&lt;p&gt;&lt;strong&gt;We Built a Multiplayer Voice AI That Actually Runs the Party&lt;/strong&gt;&lt;br&gt;
Live demo: &lt;a href="https://dev.tourl"&gt;thecruisegod.vercel.app&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fadjscyjswt2byyf31obt.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fadjscyjswt2byyf31obt.png" alt=" " width="800" height="451"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Six friends. One city. Zero agreement on what to do next.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Here's the thing nobody talks about: every AI voice assistant ever built assumes you're alone. You whisper into it. It whispers back. Clean, personal, useless for a group.&lt;/p&gt;

&lt;p&gt;Real life is messier. Three problems hit at once:&lt;/p&gt;

&lt;p&gt;Where are we going? Someone's Googling. Someone's scrolling TikTok for vibes. A third texted "idk, you pick." Nobody picks.&lt;br&gt;
Who do we call? The group needs a DJ for Saturday or a barber open right now. Someone texts their cousin. The cousin doesn't respond. The moment dies.&lt;br&gt;
What are we doing right now? You're already together but the energy's flat. The night either catches fire or it doesn't — and it usually doesn't.&lt;br&gt;
No AI handles any of this. And none of them can do it for a whole room at once.&lt;/p&gt;

&lt;p&gt;So we built &lt;strong&gt;TCG&lt;/strong&gt; — &lt;em&gt;The Cruise God&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What Is TCG?&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;TCG&lt;/strong&gt; is the world's first multiplayer voice AI concierge, game master, and local guide. It's a live, conversational AI that physically drives its own React UI while it talks — serving an entire group simultaneously.&lt;/p&gt;

&lt;p&gt;You tap the character. TCG wakes up. You tell it where you are and what the vibe is. It runs the night — out loud, with personality, without anyone touching a single button.&lt;/p&gt;

&lt;p&gt;Three modes. Each purpose-built for a real group scenario.&lt;/p&gt;

&lt;p&gt;🟢 &lt;strong&gt;Locator Mode&lt;/strong&gt;&lt;br&gt;
"Find us somewhere to go."&lt;/p&gt;

&lt;p&gt;Tell TCG the vibe — rooftop bar, late-night spot, live music — and it:&lt;/p&gt;

&lt;p&gt;Dynamically builds a contextual search query from your location and energy&lt;br&gt;
Live-scrapes the web via Firecrawl&lt;br&gt;
Reads back real venue recommendations while the UI snaps into a card layout&lt;br&gt;
It's not a Google search. It knows your city, your crew size, and what kind of night you're trying to have.&lt;/p&gt;

&lt;p&gt;🔌 &lt;strong&gt;Plug Mode&lt;/strong&gt;&lt;br&gt;
"Find me someone who can handle this."&lt;/p&gt;

&lt;p&gt;Need a DJ for next weekend? A same-day mechanic? A barber open right now?&lt;/p&gt;

&lt;p&gt;Plug Mode rewrites queries with urgency context — "available now open today same-day" — runs a live search, and voices the results. It's the "who do you know?" answer, surfaced by AI.&lt;/p&gt;

&lt;p&gt;🎮 &lt;strong&gt;Game Master Mode&lt;/strong&gt;&lt;br&gt;
"Let's play something."&lt;/p&gt;

&lt;p&gt;This is where TCG earns its name.&lt;/p&gt;

&lt;p&gt;Say "teach us a drinking game for 6 people, chaotic energy" and TCG doesn't just name a game. It scrapes the full rule set in Markdown via Firecrawl and reads the rules aloud, step by step, to the room.&lt;/p&gt;

&lt;p&gt;The UI opens a live Game Session dashboard — scoreboard, current turn, player list, rules summary — that TCG updates as the game progresses.&lt;/p&gt;

&lt;p&gt;Plus 9 built-in party tools available on command:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Tool&lt;/strong&gt;    &lt;strong&gt;Details&lt;/strong&gt;&lt;br&gt;
Spin the Bottle Synced with the room's live guest list&lt;br&gt;
Truth or Dare   3 intensity levels&lt;br&gt;
Charades    4 categories&lt;br&gt;
Coin Flip / Dice Roll   Voice-triggered&lt;br&gt;
Randomizer  Splits guests into named groups&lt;br&gt;
Timer   Countdown, voice-controlled&lt;br&gt;
Scoreboard  Live tracked&lt;br&gt;
Bill Splitter   Camera → Gemini Vision → split result read aloud&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Tech Stack&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;ElevenLabs&lt;/strong&gt; — &lt;em&gt;The Heartbeat&lt;/em&gt;&lt;br&gt;
We didn't use ElevenLabs just for TTS. We architected 10 custom Client Tools that let the agent autonomously control the React frontend in real-time:&lt;/p&gt;

&lt;p&gt;switchMode       → Snaps the UI between Locator / Plug / Game Master / Tools&lt;br&gt;
openTool         → Opens any party tool on command&lt;br&gt;
showQR           → Displays the CruiseHQ join QR code&lt;br&gt;
randomizeGroups  → Splits guests into named groups&lt;br&gt;
setGroupLeader   → Elects and announces a group captain&lt;br&gt;
updateGameState  → Syncs the live scoreboard&lt;br&gt;
displayResults   → Populates venue/service result cards&lt;br&gt;
analyzeImage     → Opens camera → Gemini Vision&lt;br&gt;
createMemory     → Screenshot → Supabase Storage → Trophy + viral caption&lt;br&gt;
stopListening    → Cuts the mic at hardware level&lt;br&gt;
TCG says "let me pull that up" — and the UI snaps. The conversation never breaks.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Firecrawl&lt;/strong&gt; — &lt;em&gt;The Search Pipeline&lt;/em&gt;&lt;br&gt;
A production-grade 3-tier pipeline:&lt;/p&gt;

&lt;p&gt;Supabase (7-day cache) → Upstash Redis (15-min hot cache) → Firecrawl live scrape&lt;br&gt;
Queries are dynamically rewritten for context. "Chaotic game for 6" becomes:&lt;/p&gt;

&lt;p&gt;"for 6 players wild hilarious high-energy party game rules how to play"&lt;/p&gt;

&lt;p&gt;Game searches extract full Markdown rule sets. TCG teaches the room, rule by rule, out loud.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Supabase Realtime&lt;/strong&gt; — &lt;em&gt;The Multiplayer Backbone&lt;/em&gt;&lt;br&gt;
Friends scan a QR code from their phones and land in a full CruiseHQ interface. They're not just watching — they're in the room.&lt;/p&gt;

&lt;p&gt;Supabase Realtime handles:&lt;/p&gt;

&lt;p&gt;Presence tracking — live roster of who's in the room&lt;br&gt;
Broadcast channels — dares, group chats, poll votes, co-host voice transcripts&lt;br&gt;
Sub-second latency across every connected device&lt;br&gt;
When a guest sends TCG a dare from their phone, it's injected as a live user message into the active ElevenLabs voice session. TCG hears it. TCG acts on it. In front of the whole room.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Gemini 2.5 Flash Vision&lt;/strong&gt; — &lt;em&gt;Eyes for the Room&lt;/em&gt;&lt;br&gt;
Vision isn't just about reading text—it's TCG's way of physically "seeing" the party. It integrates directly with the live ElevenLabs voice session as natural context, so TCG actually sees the room and talks about it.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3qht2ng9i5q4v0yyawpp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3qht2ng9i5q4v0yyawpp.png" alt=" " width="800" height="451"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Act as a Referee&lt;/em&gt;: Show TCG a chaotic game board, a Charades drawing, or a physical challenge, and its Game Vision acts as an impartial, live judge.&lt;br&gt;
&lt;em&gt;Split the Check&lt;/em&gt;: Point the camera at a receipt → TCG reads the total, splits it per person, auto-detects dynamically the currency from your GPS location, and reads the result aloud to the group.&lt;br&gt;
&lt;em&gt;Verify Items&lt;/em&gt;: Scan a barcode → instantly validated against the Open Food Facts API to tell you exactly what you're holding.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What Makes It Actually Different&lt;/strong&gt;&lt;br&gt;
We didn't add "multiplayer" as a feature. We rebuilt the entire architecture around the assumption that the AI is serving a room, not a person.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;CruiseHQ submissions&lt;/em&gt;: A word submitted from a guest's phone can be silently inserted into the agent's context — invisible to the host's transcript, visible only to TCG. The AI knows the word. The host doesn't. The game works.IT can also be public, leaving the power in your hands. This lets you post dares, suggestions- anything that requires a submission, CruiseHQ is equipped to process.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Groups trick&lt;/strong&gt;: When the host says "randomize the groups," the randomizeGroups Client Tool runs, CruiseHQ auto-creates group tabs, every guest's interface updates, and TCG announces the result out loud with personality — all in one voice command.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Wingman Protocols&lt;/strong&gt;: Users can set secret instructions ("always suggest dive bars", "roast me constantly") that are injected into the ElevenLabs system prompt on every session start. TCG already knows you before you say a word.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Trophy Room&lt;/strong&gt;: When a moment lands, TCG proactively captures it — screenshot via html2canvas, uploaded to Supabase Storage, saved as a Trophy with a generated viral caption, shared from /trophy-room.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Try It&lt;/strong&gt;&lt;br&gt;
👉 &lt;a href="https://dev.tourl"&gt;thecruisegod.vercel.app&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Open it on one device. Share the QR code with your friends. Tell TCG where you are.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Let it run the night.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmtpx8ivmio5np6wb6l8x.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmtpx8ivmio5np6wb6l8x.png" alt=" " width="800" height="451"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>elevenagents</category>
      <category>firecrawl</category>
      <category>elevenhacks</category>
      <category>agents</category>
    </item>
    <item>
      <title>From Clinic to Code: Engineering GozAI, a Clinically- grounded AI Copilot for Low Vision &amp; Visually impaired</title>
      <dc:creator>Pacy</dc:creator>
      <pubDate>Mon, 16 Mar 2026 14:33:49 +0000</pubDate>
      <link>https://forem.com/kryptopacy/from-clinic-to-code-engineering-a-clinically-grounded-ai-copilot-for-low-vision-4l7a</link>
      <guid>https://forem.com/kryptopacy/from-clinic-to-code-engineering-a-clinically-grounded-ai-copilot-for-low-vision-4l7a</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fceskx1idnsac35kqctoh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fceskx1idnsac35kqctoh.png" alt=" " width="640" height="640"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;For the 2.2 billion people experiencing vision impairment globally, standard assistive technology often feels like a burden rather than a bridge. &lt;br&gt;
The seed for my latest project, &lt;strong&gt;GozAI&lt;/strong&gt;, was planted back in 2022 during my 500-level Optometry externship at the University of Ilorin Teaching Hospital (UITH) in Nigeria. Working under the guidance of a low-vision and pediatric specialist, I spent my days with countless patients navigating life with severe visual impairments. &lt;br&gt;
Day after day, I witnessed the same frustrating reality: traditional low-vision aids are incredibly fragmented and prohibitively expensive. Patients were being asked to carry and pay a premium for &lt;em&gt;clunky, single-purpose digital magnifiers&lt;/em&gt;. &lt;br&gt;
Even then, it raised a glaring question: &lt;em&gt;We are in the digital age. Why hasn't this been solved using the devices already in our pockets?&lt;/em&gt; At the time, I had the clinical domain expertise to know exactly what the solution should look like, but I lacked the software engineering skills to actually build it. The idea stayed on the shelf.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Catalyst: Agentic Coding and Clinical Reality
&lt;/h2&gt;

&lt;p&gt;Fast forward to 2026. Recent clinical research has completely validated what I saw in the clinic years ago, showing a staggering abandonment rate of up to 50% for specialized assistive gadgets. The primary culprits? Cost, severe battery drain, and the social stigma of carrying them. People don’t want another heavy device to manage. They just want the smartphone they already own to act as their eyes.&lt;br&gt;
At the same time, we entered the era of agentic coding. With recent leaps in AI development tools, the technical barriers to entry that once held me back vanished. By building entirely with &lt;strong&gt;Google's Antigravity&lt;/strong&gt;, I was finally able to take the clinical blueprints from my externship experience and code them into reality.&lt;/p&gt;

&lt;h2&gt;
  
  
  Building GozAI: A Multimodal Accessibility Copilot
&lt;/h2&gt;

&lt;p&gt;When I entered the &lt;strong&gt;Gemini Live Agent Challenge&lt;/strong&gt;, I set out with an uncompromising goal: build an accessibility copilot that is battery-optimized, emotionally reassuring, and rooted entirely in clinical reality. &lt;br&gt;
GozAI is a real-time voice and vision assistant powered by the &lt;strong&gt;Gemini 2.0 Flash Multimodal Live API&lt;/strong&gt; and hosted on &lt;strong&gt;Google Cloud Run&lt;/strong&gt;. Instead of just building a "talking camera" that overwhelms the user, I engineered it to act as an intelligent, empathetic filter. &lt;/p&gt;

&lt;h3&gt;
  
  
  Bridging the Gap: Clinical Need &amp;amp; Technical Execution
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Verified Clinical Grounding:&lt;/strong&gt; To ensure GozAI wasn't just another tech novelty, I didn't want it just guessing. I engineered the backend RAG architecture (built using the Google ADK) to strictly rely on real, verified optometry data and textbooks. It operates under strict clinical guardrails, ensuring the advice and spatial orientation it provides are safe and accurate.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Semantic Vibro-Acoustics:&lt;/strong&gt; Continuous audio feedback causes massive cognitive fatigue for low-vision users. GozAI captures frames at a battery-saving 1 FPS and uses synchronized haptics for navigation. It only delivers audio warnings for immediate hazards, keeping the user's mental bandwidth clear.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The Glaucoma Protocol:&lt;/strong&gt; Progressive vision loss carries a high risk of depression and anxiety. GozAI’s persona is strictly constrained to be an anchor of psychological safety—calm, steady, and reassuring. It doesn't just see for the user; it acts as a supportive companion.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The UI Navigator:&lt;/strong&gt; For digital environments, GozAI acts as a bridge where standard screen readers fail. Using Gemini’s multimodal capabilities, it can interpret unlabelled buttons on a screen, read medical labels offline (via Google ML Kit), and even inject synthetic screen taps for the user.
## The Technical Backbone
To make this vision a reality, we leveraged a robust Google-native stack:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Flutter PWA / Mobile&lt;/strong&gt;&lt;br&gt;
  └── Gemini Multimodal Live API (Bidirectional WebSocket)&lt;br&gt;
        └── Google ADK Agent (Cloud Run)&lt;br&gt;
              ├── Tool: Optometry Guidelines (Verified RAG)&lt;br&gt;
              ├── Tool: Medication Lookup (OpenFDA)&lt;br&gt;
              ├── Tool: SOS → Firestore&lt;br&gt;
              └── Tool: Clinical Stats&lt;br&gt;
                    └── Firebase / Firestore (Companion Memory)&lt;br&gt;
The mobile client connects directly to the Gemini Multimodal Live API via a bidirectional WebSocket, streaming real-time audio and camera frames simultaneously — ensuring near-zero latency for those critical safety moments.&lt;/p&gt;

&lt;p&gt;Closing the Gap&lt;br&gt;
GozAI proves that when you combine deep clinical empathy with bleeding-edge tools like the Gemini Live API and Google's Antigravity, you can build solutions that actually restore independence, rather than just adding another gadget to a patient's bag.&lt;/p&gt;

&lt;p&gt;Everyone deserves to navigate the world with confidence.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpe5a26k56jn87zpt2cqj.png" alt=" " width="640" height="640"&gt;
&lt;/h2&gt;

&lt;p&gt;This piece was created for the purposes of entering the Gemini Live Agent Challenge. 🚀&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Check out the code&lt;/strong&gt;: &lt;a href="https://github.com/Kryptopacy/GozAI" rel="noopener noreferrer"&gt;https://github.com/Kryptopacy/GozAI&lt;/a&gt;&lt;br&gt;
🌐 &lt;strong&gt;Try the Demo&lt;/strong&gt;: &lt;a href="https://dev.tourl"&gt;gozai-app.web.app&lt;/a&gt;&lt;br&gt;
🚀 Built for: &lt;em&gt;Gemini Live Agent Challenge&lt;/em&gt; | Tracks: Live Agents + UI Navigator&lt;/p&gt;

</description>
      <category>geminiliveagentchallenge</category>
      <category>flutter</category>
      <category>googlecloud</category>
      <category>ai</category>
    </item>
  </channel>
</rss>
