<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Swagatika Beura</title>
    <description>The latest articles on Forem by Swagatika Beura (@swagatika_beura2).</description>
    <link>https://forem.com/swagatika_beura2</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3568353%2F0199401b-7105-4031-a024-8a97b43a58f9.jpg</url>
      <title>Forem: Swagatika Beura</title>
      <link>https://forem.com/swagatika_beura2</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/swagatika_beura2"/>
    <language>en</language>
    <item>
      <title>🚀Mission Accomplished: Building a Real-Time AI Spatial Agent | Vision Possible Hackathon &amp; The Power of Learning by Building</title>
      <dc:creator>Swagatika Beura</dc:creator>
      <pubDate>Sun, 01 Mar 2026 04:09:34 +0000</pubDate>
      <link>https://forem.com/swagatika_beura2/mission-accomplished-building-a-real-time-ai-spatial-agent-vision-possible-hackathon-the-40m5</link>
      <guid>https://forem.com/swagatika_beura2/mission-accomplished-building-a-real-time-ai-spatial-agent-vision-possible-hackathon-the-40m5</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;When &lt;strong&gt;curiosity&lt;/strong&gt; meets &lt;strong&gt;constraints&lt;/strong&gt;, &lt;strong&gt;innovation&lt;/strong&gt; happens.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;And when you choose to just &lt;strong&gt;learn by building and enjoy the journey — growth becomes unstoppable&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Recently, I built &lt;strong&gt;VisionMate AI&lt;/strong&gt; as my submission for the &lt;strong&gt;Vision Possible: Agent Protocol Hackathon&lt;/strong&gt; — an event that challenged developers worldwide to build cutting-edge &lt;strong&gt;generative AI solutions using multimodal agents.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Hosted by &lt;a href="https://www.wemakedevs.org/hackathons/vision" rel="noopener noreferrer"&gt;WeMakeDevs&lt;/a&gt; , the hackathon challenged participants to push the limits of what’s possible with &lt;strong&gt;real-time AI&lt;/strong&gt; — from blazing-fast inference to true multimodal interactions. &lt;/p&gt;

&lt;p&gt;And that’s exactly where &lt;strong&gt;VisionMateAI&lt;/strong&gt;  was born.&lt;/p&gt;




&lt;h3&gt;
  
  
  🚀 What is VisionMate AI?
&lt;/h3&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/Y04gNrPA3kk"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

&lt;p&gt;Standard computer vision is entirely passive. If you are a cyclist, an e-scooter rider, or someone using a smart dashcam, knowing **"there is a car" **in a video feed is useless. You already know there are cars on the road! &lt;/p&gt;

&lt;p&gt;The real problem is knowing if a specific car or bus is &lt;strong&gt;approaching too fast or getting dangerously close to you&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;VisionMate AI&lt;/strong&gt; is a comprehensive mobility platform that allows users to &lt;strong&gt;turn any standard camera into an active spatial hazard agent&lt;/strong&gt; — transforming the way we handle personal safety on the road.&lt;/p&gt;

&lt;p&gt;It’s built with &lt;strong&gt;Streamlit&lt;/strong&gt;, &lt;strong&gt;OpenCV&lt;/strong&gt;, and &lt;strong&gt;Python&lt;/strong&gt;, powered by &lt;strong&gt;Stream's Vision Agents SDK&lt;/strong&gt; and &lt;strong&gt;Ultralytics YOLOv8&lt;/strong&gt;, running on the &lt;strong&gt;Gemini Realtime LLM&lt;/strong&gt; for unparalleled inference speed and voice interaction.&lt;/p&gt;




&lt;h3&gt;
  
  
  🧠 How I Used Stream Vision Agents &amp;amp; Gemini Together :
&lt;/h3&gt;

&lt;p&gt;At the core of ** VisionMate AI **lies a unique combination:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;🧩 &lt;strong&gt;Ultralytics YOLOv8&lt;/strong&gt; for highly accurate entity detection and bounding box generation.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;⚡ &lt;strong&gt;Stream Vision Agents SDK&lt;/strong&gt; for lightning-fast edge network routing — ensuring real-time, snappy voice-to-vision pipelines.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;🗣️ &lt;strong&gt;Gemini Realtime LLM&lt;/strong&gt; for instant, conversational audio warnings.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  ⚙️ Architecture Highlights:
&lt;/h3&gt;

&lt;p&gt;Instead of just identifying objects passively, &lt;strong&gt;VisionMate AI&lt;/strong&gt; actively measures spatial threats across three core modes:&lt;/p&gt;

&lt;h4&gt;
  
  
  📷 1. Image Processing (High-Fidelity Frame Inspection)
&lt;/h4&gt;

&lt;p&gt;Contextual safety thrives on precision. You simply &lt;strong&gt;upload an image&lt;/strong&gt;, and the model instantly processes it to accurately detect, label, and draw bounding boxes around the objects present in the frame. It maps out the immediate environment, providing structured, &lt;strong&gt;high-quality safety&lt;/strong&gt; responses in &lt;strong&gt;milliseconds&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffcwz3kmjcpa9giv0wwly.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffcwz3kmjcpa9giv0wwly.jpeg" alt=" "&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h4&gt;
  
  
  🎬 2. Video Pipeline (Time-Series Hazard Tracking)
&lt;/h4&gt;

&lt;p&gt;This feature is where the custom spatial math truly shines. You upload a pre-recorded dashcam video, and the system processes it &lt;strong&gt;frame-by-frame&lt;/strong&gt;. It features a synchronous dashboard with a &lt;strong&gt;live telemetry&lt;/strong&gt; HUD that successfully detects and tracks objects throughout the footage, exporting a final security audit in &lt;strong&gt;JSON format&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb82pyn7by1b3jwcn0ifl.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb82pyn7by1b3jwcn0ifl.jpeg" alt=" "&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h4&gt;
  
  
  🎥 3. Live Protocol (Real-Time Agent Interface)
&lt;/h4&gt;

&lt;p&gt;This is where the custom hackathon logic comes in. It features a Native Hardware Test for local proximity calculations and a WebRTC Edge Network. While the AI seamlessly detects everything in your &lt;strong&gt;live camera feed&lt;/strong&gt;, it isolates true spatial threats in real-time. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F565x2q6hq9vfb4kjpzge.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F565x2q6hq9vfb4kjpzge.png" alt=" "&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I wrote the &lt;strong&gt;"40% Rule"&lt;/strong&gt;—a custom proximity math algorithm inside the bounding box processor:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;
&lt;span class="c1"&gt;# The Core Spatial Hazard Math
&lt;/span&gt;
&lt;span class="k"&gt;def&lt;/span&gt; &lt;span class="nf"&gt;process_bounding_boxes&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;image&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;results&lt;/span&gt;&lt;span class="p"&gt;):&lt;/span&gt;

    &lt;span class="c1"&gt;# ... setup code ...
&lt;/span&gt;
    &lt;span class="k"&gt;for&lt;/span&gt; &lt;span class="n"&gt;box&lt;/span&gt; &lt;span class="ow"&gt;in&lt;/span&gt; &lt;span class="n"&gt;results&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;boxes&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;

        &lt;span class="n"&gt;x1&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;y1&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;x2&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;y2&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nf"&gt;map&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="nb"&gt;int&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;box&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;xyxy&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="mi"&gt;0&lt;/span&gt;&lt;span class="p"&gt;])&lt;/span&gt;

        &lt;span class="n"&gt;vertical_span&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;y2&lt;/span&gt; &lt;span class="o"&gt;-&lt;/span&gt; &lt;span class="n"&gt;y1&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="o"&gt;/&lt;/span&gt; &lt;span class="n"&gt;img_h&lt;/span&gt; &lt;span class="c1"&gt;# Object height relative to frame
&lt;/span&gt;


        &lt;span class="k"&gt;if&lt;/span&gt; &lt;span class="n"&gt;vertical_span&lt;/span&gt; &lt;span class="o"&gt;&amp;gt;&lt;/span&gt; &lt;span class="mf"&gt;0.4&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;

            &lt;span class="n"&gt;status&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;color&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;thickness&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;HAZARD&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;0&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;0&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;255&lt;/span&gt;&lt;span class="p"&gt;),&lt;/span&gt; &lt;span class="mi"&gt;4&lt;/span&gt;

        &lt;span class="k"&gt;elif&lt;/span&gt; &lt;span class="n"&gt;vertical_span&lt;/span&gt; &lt;span class="o"&gt;&amp;gt;&lt;/span&gt; &lt;span class="mf"&gt;0.2&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;

            &lt;span class="n"&gt;status&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;color&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;thickness&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;MID&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;0&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;255&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;255&lt;/span&gt;&lt;span class="p"&gt;),&lt;/span&gt; &lt;span class="mi"&gt;2&lt;/span&gt;

        &lt;span class="k"&gt;else&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;

            &lt;span class="n"&gt;status&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;color&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;thickness&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;FAR&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;0&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;255&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;0&lt;/span&gt;&lt;span class="p"&gt;),&lt;/span&gt; &lt;span class="mi"&gt;2&lt;/span&gt;

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;If an object (like a car or bus) suddenly takes up more than 40% of the camera frame, my system mathematically determines it is an imminent collision risk and triggers the Gemini agent to issue an instant &lt;strong&gt;gTTS&lt;/strong&gt; voice alert!&lt;/p&gt;




&lt;h3&gt;
  
  
  🧩 Tech Stack Overview :
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Frontend:&lt;/strong&gt; Streamlit, Custom CSS UI Components&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Core Logic:&lt;/strong&gt; Python, OpenCV, NumPy&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Authentication:&lt;/strong&gt; Bcrypt, JSON-based Local DB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;AI &amp;amp; Vision:&lt;/strong&gt; Stream Vision Agents SDK, Ultralytics YOLOv8, Gemini Realtime LLM&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Audio:&lt;/strong&gt; Google Text-to-Speech (gTTS)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Deployment/Environment:&lt;/strong&gt; Dotenv, Nest_Asyncio&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  ⚡ The Experience — When Curiosity Met Constraints :
&lt;/h3&gt;

&lt;p&gt;The &lt;strong&gt;Vision Possible: Agent Protocol Hackathon&lt;/strong&gt; was centered around one powerful idea: building &lt;strong&gt;real-time multimodal AI agents&lt;/strong&gt; that go beyond traditional cloud pipelines.&lt;/p&gt;

&lt;p&gt;For me, this was an opportunity to test a belief I’ve always lived by:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"You don’t need to be an expert to &lt;strong&gt;start&lt;/strong&gt;. You just need to start  &lt;strong&gt;doing&lt;/strong&gt;."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;I joined the hackathon organized by &lt;strong&gt;WeMakeDevs&lt;/strong&gt; almost spontaneously. There was no massive roadmap—just curiosity and a desire to build something meaningful.&lt;/p&gt;

&lt;h4&gt;
  
  
  🛠️ Engineering Around the "OOM" Wall
&lt;/h4&gt;

&lt;p&gt;To be honest, my development machine is not a high-end workstation. Running &lt;strong&gt;Ultralytics YOLOv8&lt;/strong&gt; locally was a struggle, and under strict &lt;strong&gt;1GB cloud RAM limits&lt;/strong&gt;, I repeatedly faced &lt;strong&gt;Out of Memory (OOM)&lt;/strong&gt; crashes.&lt;/p&gt;

&lt;p&gt;Instead of scaling down, I treated these constraints as engineering challenges:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Custom OpenCV Scaler:&lt;/strong&gt; I engineered an aspect-ratio scaler to intelligently reduce the memory footprint without compromising spatial accuracy.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Memory Management:&lt;/strong&gt; I minimized redundant tensor allocations and tuned inference batching to keep the system stable on low-tier instances.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;WebRTC Integration:&lt;/strong&gt; The real breakthrough came with &lt;strong&gt;Stream’s Vision Agents SDK&lt;/strong&gt; and &lt;strong&gt;Google’s Gemini Realtime LLM&lt;/strong&gt;.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  🚀 The Result: Sub-30ms Latency :
&lt;/h4&gt;

&lt;p&gt;Rather than relying on a laggy HTTP pipeline, the &lt;strong&gt;Vision Agents SDK&lt;/strong&gt; enabled a &lt;strong&gt;WebRTC-based streaming architecture&lt;/strong&gt; over Stream’s Edge network. This allowed YOLO detections to bind directly with Gemini’s reasoning engine.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Stream&lt;/strong&gt; handled efficient edge routing of heavy video data, while &lt;strong&gt;Gemini&lt;/strong&gt; provided refined multimodal understanding—allowing the AI to &lt;strong&gt;“see”&lt;/strong&gt; through the camera and &lt;strong&gt;“speak”&lt;/strong&gt; contextual warnings in real-time.&lt;/p&gt;




&lt;h3&gt;
  
  
  🏁 Looking back, this was about proving something to myself.
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;✅ &lt;strong&gt;You don’t need a perfect plan.&lt;/strong&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;✅ &lt;strong&gt;You don’t need elite hardware.&lt;/strong&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;✅ &lt;strong&gt;You don’t need to feel fully ready.&lt;/strong&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;You need &lt;strong&gt;curiosity&lt;/strong&gt;. You need &lt;strong&gt;resilience&lt;/strong&gt;. And you need to &lt;strong&gt;enjoy the journey&lt;/strong&gt; while building.&lt;/p&gt;

&lt;p&gt;Because when curiosity meets constraints — and you choose to build anyway — &lt;strong&gt;innovation becomes inevitable.&lt;/strong&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  🌍 What’s Next for VisionMate AI :
&lt;/h3&gt;

&lt;p&gt;This is only the beginning. I’m working on turning &lt;strong&gt;VisionMate AI&lt;/strong&gt; into a production-grade safety ecosystem where:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Edge Deployment:&lt;/strong&gt; Porting the weights to run natively on NPU-enabled hardware like Raspberry Pi for dedicated smart dashcams.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Predictive Analytics:&lt;/strong&gt; Moving beyond the "40% Rule" to include velocity-based trajectory prediction for higher accuracy.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Unified Safety Dashboard:&lt;/strong&gt; Allowing users to review past hazard logs, security audits, and spatial data in one centralized interface.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  🙌 A Huge Thanks to &lt;strong&gt;WeMakeDevs&lt;/strong&gt; :
&lt;/h3&gt;

&lt;p&gt;A massive shoutout to &lt;strong&gt;WeMakeDevs&lt;/strong&gt; and &lt;strong&gt;Kunal Kushwaha&lt;/strong&gt; for organizing such a well-structured and inspiring hackathon experience. &lt;/p&gt;

&lt;p&gt;From the very beginning, they ensured that participants had everything they needed—from technical resources and community support to sessions with the &lt;strong&gt;Stream&lt;/strong&gt; team explaining their technology in depth.&lt;/p&gt;

&lt;p&gt;They hosted interactive workshops on how to effectively use the &lt;strong&gt;Vision Agents SDK&lt;/strong&gt;, which really helped me understand how to build a low-latency, impactful project. The &lt;strong&gt;WeMakeDevs community&lt;/strong&gt; was always active and approachable, answering doubts instantly and fostering an environment where learning and innovation could thrive.&lt;/p&gt;

&lt;p&gt;I’m truly grateful for the opportunity to be a part of such an event — it wasn’t just a competition, it was a &lt;strong&gt;collaborative learning experience&lt;/strong&gt; that empowered me to build, explore, and create with real-time AI.&lt;/p&gt;




&lt;h3&gt;
  
  
  &amp;lt;/&amp;gt; Github Repo
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://github.com/swagatika60/Vision-Agent" rel="noopener noreferrer"&gt;Github Repo&lt;/a&gt;&lt;/strong&gt; &lt;br&gt;
⚠️ &lt;strong&gt;Disclaimer:&lt;/strong&gt; This project was developed during the &lt;em&gt;Vision Possible Hackathon&lt;/em&gt;. The current architecture is optimized for low-memory environments (1GB RAM) to ensure accessibility for users without high-end hardware.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;VisionMate AI&lt;/strong&gt; is an active spatial hazard agent that empowers users to turn any camera into a real-time safety tool. With advanced CV capabilities, sub-30ms WebRTC latency, and conversational AI alerts, VisionMate transforms passive video feeds into life-saving interactions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://youtube.com/watch?v=Y04gNrPA3kk" rel="noopener noreferrer"&gt;Demo Video&lt;/a&gt;&lt;/strong&gt; &lt;br&gt;
&lt;strong&gt;&lt;a href="https://vision-agent.streamlit.app/" rel="noopener noreferrer"&gt;Live Application&lt;/a&gt;&lt;/strong&gt; &lt;br&gt;
&lt;strong&gt;Built with ❤️ by Swagatika Beura for the Vision Possible Hackathon.&lt;/strong&gt;&lt;/p&gt;

</description>
      <category>visionagents</category>
      <category>wemakedevs</category>
      <category>learnbydoing</category>
      <category>ai</category>
    </item>
    <item>
      <title>wemakedevs</title>
      <dc:creator>Swagatika Beura</dc:creator>
      <pubDate>Sun, 01 Mar 2026 01:11:30 +0000</pubDate>
      <link>https://forem.com/swagatika_beura2/wemakedevs-mdo</link>
      <guid>https://forem.com/swagatika_beura2/wemakedevs-mdo</guid>
      <description></description>
    </item>
    <item>
      <title>2025 Wrap Up</title>
      <dc:creator>Swagatika Beura</dc:creator>
      <pubDate>Wed, 31 Dec 2025 17:23:21 +0000</pubDate>
      <link>https://forem.com/swagatika_beura2/2025-wrap-up--5kc</link>
      <guid>https://forem.com/swagatika_beura2/2025-wrap-up--5kc</guid>
      <description>&lt;p&gt;1.Explored many fields and worked on some amazing 2.stuff — tried, experimented, and learned by doing.&lt;br&gt;
Learned a lot this year, not just technical skills but also how things work in the real world.&lt;br&gt;
3.Became part of some incredible communities from around the world 🌍 — open source, tech, and beyond.&lt;br&gt;
4.Midway through the year, I finally found clarity about my path and what I truly want to focus on.&lt;br&gt;
5.Failed many times. Restarted again and again. Each failure taught me something new and helped me discover some pretty cool things.&lt;br&gt;
6.This year taught me how to handle myself — how to manage pressure, stay disciplined, and keep going even in difficult situations.&lt;br&gt;
7.Faced some really bad situations, but those moments showed me how growth actually happens. One lesson stayed with me: keep your head down, stay focused, and let your work speak.&lt;br&gt;
Grateful for every experience — the good, the bad, and everything in between.&lt;br&gt;
Still learning. Still building. 🚀&lt;/p&gt;

&lt;h1&gt;
  
  
  2025WrapUp #Growth #DevLife
&lt;/h1&gt;

</description>
    </item>
    <item>
      <title>✨ A few days ago, my first PR on Kestra got merged! ✨</title>
      <dc:creator>Swagatika Beura</dc:creator>
      <pubDate>Mon, 29 Dec 2025 12:29:31 +0000</pubDate>
      <link>https://forem.com/swagatika_beura2/a-few-days-ago-my-first-pr-on-kestra-got-merged-482g</link>
      <guid>https://forem.com/swagatika_beura2/a-few-days-ago-my-first-pr-on-kestra-got-merged-482g</guid>
      <description>&lt;p&gt;I was completely new to this project, and honestly, it wasn’t easy. But this journey taught me something important: you don’t need to be an expert to start—you just need to start.&lt;br&gt;
I failed many times, tried repeatedly, and kept pushing with the hope that one day it would be solved. And finally, that day came—my PR got merged!&lt;br&gt;
Huge thanks to the #Kestra team for their support.&lt;br&gt;
Lesson learned: don’t give up, always try, and learn from failure. One day, your time will come—you don’t need to chase it; it will reach you.&lt;/p&gt;

</description>
      <category>motivation</category>
      <category>beginners</category>
      <category>showdev</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Start First, Learn by Doing: My Thoughts on Growth 🚀</title>
      <dc:creator>Swagatika Beura</dc:creator>
      <pubDate>Fri, 17 Oct 2025 14:01:28 +0000</pubDate>
      <link>https://forem.com/swagatika_beura2/post-titlestart-first-learn-by-doing-my-thoughts-on-growth-1cef</link>
      <guid>https://forem.com/swagatika_beura2/post-titlestart-first-learn-by-doing-my-thoughts-on-growth-1cef</guid>
      <description>&lt;p&gt;Hello Dev Community! 👋&lt;/p&gt;

&lt;p&gt;I want to share a simple but important thought I’ve realized on my journey:&lt;/p&gt;

&lt;p&gt;You don’t need to be perfect to start.&lt;/p&gt;

&lt;p&gt;I’m not perfect, and I don’t know everything yet — but I’ve decided to start doing things instead of overthinking. I’m exploring Python, building beginner-friendly projects, contributing to open source, participating in communities, and practicing DSA.&lt;/p&gt;

&lt;p&gt;The key is learning by doing practical things. Every project, every contribution, and every small step teaches me more than endless planning ever could.&lt;/p&gt;

&lt;p&gt;If you’re also learning something new or thinking about starting, just start. Don’t wait for perfection — growth comes from action and experience 🌱💻&lt;/p&gt;

</description>
      <category>networking</category>
      <category>webdev</category>
    </item>
  </channel>
</rss>
