<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: ARmedia</title>
    <description>The latest articles on Forem by ARmedia (@sns_ar).</description>
    <link>https://forem.com/sns_ar</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3668324%2F5384dfe2-8fc8-4869-96c1-cc4d869da042.png</url>
      <title>Forem: ARmedia</title>
      <link>https://forem.com/sns_ar</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/sns_ar"/>
    <language>en</language>
    <item>
      <title>Building a Smart Refrigerator with a $15 IoT Camera and SAM 3: Solving the "Warm Water" Problem</title>
      <dc:creator>ARmedia</dc:creator>
      <pubDate>Thu, 18 Dec 2025 07:43:46 +0000</pubDate>
      <link>https://forem.com/sns_ar/building-a-smart-refrigerator-with-a-15-iot-camera-and-sam-3-solving-the-warm-water-problem-1pa3</link>
      <guid>https://forem.com/sns_ar/building-a-smart-refrigerator-with-a-15-iot-camera-and-sam-3-solving-the-warm-water-problem-1pa3</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;I've been battling with AI (Claude) for 14 hours a day. Couldn't be happier.&lt;/p&gt;

&lt;p&gt;— Akio Shiki (@ar_akio) &lt;a href="https://twitter.com/ar_akio/status/1980185472078070040?ref_src=twsrc%5Etfw" rel="noopener noreferrer"&gt;October 20, 2025&lt;/a&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  Black Friday Haul and the Warm Water Problem
&lt;/h2&gt;

&lt;p&gt;Hi, I'm Akio, an engineer at an AI startup.&lt;/p&gt;

&lt;p&gt;So, how did everyone spend &lt;strong&gt;Amazon Black Friday&lt;/strong&gt; this year? As usual, I bulk-ordered sparkling water and bottled drinks purely because they were "on sale," and now my entryway is buried under a tower of cardboard boxes.&lt;/p&gt;

&lt;p&gt;And then comes the inevitable: &lt;strong&gt;"I forgot to move them to the fridge, and now I'm stuck drinking warm water."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;"Can technology solve this?" ...Half joking, half serious.&lt;/p&gt;

&lt;p&gt;This time, as a PoC (Proof of Concept) with an eye toward future business applications, I built an &lt;strong&gt;object recognition system combining the ultra-cheap IoT microcontroller "ESP32" with Meta's latest model "SAM 3 (Segment Anything Model 3)."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvrfgvrebjv3tr9d66cvz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvrfgvrebjv3tr9d66cvz.png" width="800" height="446"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Mountains of bottled water still sitting in cardboard boxes at the office&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Why "water bottles in a refrigerator"? Because hidden within this seemingly trivial problem are &lt;strong&gt;technical challenges that any AI engineer can appreciate&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why "Refrigerator × Water Bottles" Is the Perfect PoC
&lt;/h2&gt;

&lt;p&gt;This might seem like a joke topic, but from a computer vision (CV) perspective, the inside of a refrigerator is an &lt;strong&gt;S-rank difficulty dungeon&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Here are the three reasons I chose this as my PoC target. These same challenges apply directly to industrial robotics and autonomous vehicles:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;The Difficulty of Transparent/Translucent Objects&lt;/strong&gt;: Water bottles are transparent. Not only does the background show through, but they create complex light reflections from the interior lighting. Traditional CNN-based object detection often fails to capture the contours and processes them as part of the background.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;The "Transparent Shelf" Trap&lt;/strong&gt;: Modern refrigerators have transparent plastic or glass shelves. Even powerful segmentation models like SAM can misidentify shelf edges as "object boundaries," or detect objects on lower shelves by seeing through the transparent shelf above.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Challenging Lighting Conditions&lt;/strong&gt;: The back of the fridge is dark; the front is bright. On top of the extreme contrast, bottle shadows fall on transparent shelves, and the system might mistake those "shadows" for actual objects.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7imp0yghx7d6yn8r7cvw.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7imp0yghx7d6yn8r7cvw.jpg" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The actual inside of the refrigerator&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;In other words: &lt;strong&gt;"If we can accurately segment water bottles in this harsh environment, most object recognition tasks in offices or factories will be a breeze."&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  System Architecture: $15 Edge Device Meets State-of-the-Art AI
&lt;/h2&gt;

&lt;p&gt;The setup is extremely simple and low-cost:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Edge (Eyes):&lt;/strong&gt; ESP32S3-CAM&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Brain:&lt;/strong&gt; Local PC Server (running Ryzen AI MAX+ 395 with Meta SAM 3)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Network:&lt;/strong&gt; WiFi (HTTP POST)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The ESP32 itself doesn't have the horsepower to run SAM 3. It functions purely as a capture device, sending images to the server. The server handles inference and returns results (inventory count, mask images, etc.).&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv4q50ps7t20acc4pdswh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv4q50ps7t20acc4pdswh.png" alt=" " width="800" height="427"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;System architecture diagram&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The Zero-Tuning Revelation: Fully AI-Driven Results
&lt;/h2&gt;

&lt;p&gt;The achievement I want to emphasize most from this PoC isn't the recognition accuracy itself—it's that &lt;strong&gt;we made zero environment-specific customizations&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Typically, for demos like this, there's a temptation to "cheat" (or "optimize") by adjusting lighting or fixing the camera at an angle that's easy to recognize. For this system, we eliminated all of that.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Refrigerator-agnostic:&lt;/strong&gt; Zero calibration for specific refrigerator models or shelf arrangements.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Object-agnostic:&lt;/strong&gt; No pre-training on specific water bottle brands or shapes (round vs. square).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Camera position-free:&lt;/strong&gt; No precise adjustment of camera placement or angle.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This means the transparent object recognition was achieved &lt;strong&gt;entirely through SAM 3's inference capabilities and automatic adjustments&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;In traditional image processing development, you'd need tedious parameter tuning (heuristic craftsmanship) like "for this refrigerator's lighting conditions, the binarization threshold should be around X..." &lt;/p&gt;

&lt;p&gt;But this time, we completely eliminated that human "overfitting to the environment." The fact that a pure foundation model demonstrated this level of environmental adaptability has huge implications for reducing deployment costs and accelerating rollout speed.&lt;/p&gt;

&lt;h2&gt;
  
  
  Results and Future Outlook
&lt;/h2&gt;

&lt;p&gt;The result: water bottles inside the refrigerator were segmented with remarkable accuracy, unfazed by transparent shelves or reflections.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8w22uytvggamb02vxevw.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8w22uytvggamb02vxevw.jpg" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Water bottles in the refrigerator with clean segmentation masks (color-coded)&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Now I might finally avoid the tragedy of having a packed fridge when I don't want a drink, and an empty one when I do... maybe.&lt;/p&gt;

&lt;p&gt;Through this PoC, we confirmed that &lt;strong&gt;foundation models like SAM 3 can be a powerful solution to the classic challenge of "transparent object recognition."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;As a startup, we're searching for the seeds of society-changing innovation through the accumulation of experiments like this—experiments that are relatable (and admittedly a bit ridiculous).&lt;/p&gt;

&lt;p&gt;If you have thoughts or feedback on this article, or if you're an engineer thinking "I want to try this with my fridge!"—drop a comment below!&lt;/p&gt;

</description>
      <category>iot</category>
      <category>ai</category>
      <category>showdev</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>Running SAM 3 on AMD Ryzen AI Max+ 395: A Complete Guide to Fixing the rocBLAS Error</title>
      <dc:creator>ARmedia</dc:creator>
      <pubDate>Thu, 18 Dec 2025 07:43:37 +0000</pubDate>
      <link>https://forem.com/sns_ar/running-sam-3-on-amd-ryzen-ai-max-395-a-complete-guide-to-fixing-the-rocblas-error-25io</link>
      <guid>https://forem.com/sns_ar/running-sam-3-on-amd-ryzen-ai-max-395-a-complete-guide-to-fixing-the-rocblas-error-25io</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;I've been battling with AI (Claude) for 14 hours a day. Couldn't be happier.&lt;/p&gt;

&lt;p&gt;— Akio Shiki (@ar_akio) &lt;a href="https://twitter.com/ar_akio/status/1980185472078070040?ref_src=twsrc%5Etfw" rel="noopener noreferrer"&gt;October 20, 2025&lt;/a&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Hi, I'm Akio, an engineer at an AI development startup. In my previous article, I introduced SAM 3. This time, I'll share the pitfalls I encountered when running SAM 3 on AMD hardware.&lt;/p&gt;

&lt;p&gt;We're constantly testing the latest AI models and hardware, and right now I have in my hands what can only be described as a monument to AMD engineering: the &lt;strong&gt;Ryzen AI Max+ 395&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhnhbtjx7xp2eoczn08sf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhnhbtjx7xp2eoczn08sf.png" alt=" " width="800" height="515"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.amd.com/ja/products/processors/laptop/ryzen/ai-300-series/amd-ryzen-ai-max-plus-395.html" rel="noopener noreferrer"&gt;AMD Official&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The specs on this machine are, frankly, insane. With high-bandwidth memory and a powerful iGPU, this device &lt;strong&gt;truly shines when running massive LLMs like OpenAI's gpt-oss-120b locally&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;But that's not what I'm doing today.&lt;/p&gt;

&lt;p&gt;Today, it's Meta's latest image segmentation model: &lt;strong&gt;SAM 3 (Segment Anything Model 3)&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ff3hq8fm3ncti4anu1e8p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ff3hq8fm3ncti4anu1e8p.png" alt=" " width="800" height="370"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://ai.meta.com/sam3/" rel="noopener noreferrer"&gt;Meta Official&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;"Wait, SAM 3? Isn't that lightweight? If you want inference speed, wouldn't an NVIDIA dGPU be a better fit?"&lt;/p&gt;

&lt;p&gt;You're absolutely right. No argument there.&lt;/p&gt;

&lt;p&gt;Running SAM 3 on a Ryzen AI Max+ 395 is, in a sense, &lt;strong&gt;using a sledgehammer to crack a nut&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;But you know what? I don't care. The reason is simple:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;"I just wanted to run the hottest new model on AMD's latest hardware."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This is a passion project, efficiency be damned. That said, the errors I encountered and the solutions I found should be universally valuable for AMD users. Consider this a definitive guide to conquering the rocBLAS error that virtually every Ryzen AI user will face.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Despair: No Answers Anywhere on the Web
&lt;/h2&gt;

&lt;p&gt;My setup: Windows 11, using AMD's AI stack &lt;strong&gt;ROCm (HIP SDK)&lt;/strong&gt; to run SAM 3 on PyTorch.&lt;/p&gt;

&lt;p&gt;Setup went smoothly. Time to run the inference script! ...And the moment I did, my terminal was flooded with merciless error logs.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;rocBLAS error: TensileLibrary.dat not found
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Ah yes, the classic AMD environment error. "TensileLibrary.dat not found." Translation: &lt;strong&gt;"I can't find the computation library for your GPU (gfx1151), so I can't do any calculations."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Because the Ryzen AI Max+ 395 uses the latest architecture, the official libraries haven't fully caught up with the path configurations... a common story with newly released hardware.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Standard "Environment Variable Spoofing" Trick Doesn't Work?
&lt;/h3&gt;

&lt;p&gt;Normally in AMD circles, when you hit this error, you use a &lt;strong&gt;workaround: spoofing the environment variable&lt;/strong&gt;. Since gfx1151 is highly compatible with the Radeon RX 7000 series (gfx1100), you can trick the system into thinking "I'm actually gfx1100."&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;$env:HSA_OVERRIDE_GFX_VERSION = "11.0.0"
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This should solve everything... or so I thought. &lt;strong&gt;But this time, it didn't work.&lt;/strong&gt; The error logs stubbornly insisted "I can't find the files for gfx1151" and kept looking in &lt;code&gt;site-packages\_rocm_sdk_libraries_gfx1151\bin&lt;/code&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Searching the Depths of the Internet, Finding Nothing
&lt;/h3&gt;

&lt;p&gt;Even after consulting various AI assistants for solutions, the final verdict was:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;🔴 Current Status: Local execution is technically impossible (as of December 2025)&lt;br&gt;
&lt;strong&gt;Local execution is technically impossible until AMD officially releases Tensile libraries for gfx1151.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;No way. There has to be a solution. I refused to give up.&lt;/p&gt;

&lt;p&gt;"rocBLAS error gfx1151," "Ryzen AI 300 PyTorch"... I searched Google with every keyword I could think of, dove deep into GitHub Issues and Reddit threads, but &lt;strong&gt;information was shockingly nonexistent&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;The Ryzen AI 300 series (Strix Point) is so new that apparently no one in the world had established a workaround for this error yet. Just as I was about to resign myself to using this as a dedicated LLM machine, I decided to go back to basics and dig through the library folders on my own PC.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Solution: The Files Were "Hidden" All Along
&lt;/h2&gt;

&lt;p&gt;If the web has no answers, look locally. The path indicated in the error logs indeed had no folder. However, when I thoroughly searched through &lt;code&gt;site-packages&lt;/code&gt;—the PyTorch (ROCm version) installation folder—I found an unfamiliar directory.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;_rocm_sdk_libraries_custom&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;"Custom...?" With a bad feeling, I opened it up and found something surprising.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdotdy35ta3ksdylq3jy3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdotdy35ta3ksdylq3jy3.png" alt=" " width="800" height="503"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;gfx1151's TensileLibrary.dat&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;(Screenshot: gfx1151-related files inside &lt;code&gt;_rocm_sdk_libraries_custom\bin\rocblas\library&lt;/code&gt;)&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;There it is! &lt;strong&gt;TensileLibrary_lazy_gfx1151.dat&lt;/strong&gt;!&lt;/p&gt;

&lt;p&gt;The RDNA 3.5 library files were included all along. But while PyTorch was &lt;strong&gt;looking for a folder named &lt;code&gt;_rocm_sdk_libraries_gfx1151&lt;/code&gt;&lt;/strong&gt;, the actual files were &lt;strong&gt;isolated deep within &lt;code&gt;_rocm_sdk_libraries_custom&lt;/code&gt;&lt;/strong&gt;. No wonder it couldn't find them.&lt;/p&gt;

&lt;p&gt;It makes sense why there was no information online. This wasn't a configuration error—it was a &lt;strong&gt;folder structure mismatch&lt;/strong&gt;, an extremely analog trap.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Complete Fix: Folder Transplant Surgery
&lt;/h2&gt;

&lt;p&gt;Once you know the cause, you just need to put the files where they belong. For AMD Ryzen AI users everywhere, here's the solution—possibly the first public documentation of this fix.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Rescue the Files from Their Hiding Place
&lt;/h3&gt;

&lt;p&gt;Open the following path in File Explorer (adjust the Python environment path for your setup):&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;...\site-packages\_rocm_sdk_libraries_custom\bin\rocblas\library
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Copy &lt;strong&gt;all files&lt;/strong&gt; in this directory (&lt;code&gt;.dat&lt;/code&gt; files, &lt;code&gt;.hsaco&lt;/code&gt; files, etc.).&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create the Correct Folder Structure
&lt;/h3&gt;

&lt;p&gt;Go back to the &lt;code&gt;site-packages&lt;/code&gt; root and create a new folder hierarchy that matches what PyTorch expects:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Create a folder named &lt;code&gt;_rocm_sdk_libraries_gfx1151&lt;/code&gt;
&lt;/li&gt;
&lt;li&gt;Inside it, create a folder named &lt;code&gt;bin&lt;/code&gt;
&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  Step 3: Place Files and Rename
&lt;/h3&gt;

&lt;p&gt;Paste all the files you copied in Step 1 into the &lt;code&gt;bin&lt;/code&gt; folder you just created.&lt;/p&gt;

&lt;p&gt;As an extra precaution, duplicate &lt;code&gt;TensileLibrary_lazy_gfx1151.dat&lt;/code&gt; and rename the copy to &lt;strong&gt;&lt;code&gt;TensileLibrary.dat&lt;/code&gt;&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjgoyf4ekgi4wd25zpoou.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjgoyf4ekgi4wd25zpoou.png" alt=" " width="800" height="503"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Files successfully moved to the newly created folder&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The Result: SAM 3 Running Blazingly Fast
&lt;/h2&gt;

&lt;p&gt;After fixing the folders, I ran the script again, fingers crossed.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F96j4yi3ucbptnbdnv9vz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F96j4yi3ucbptnbdnv9vz.png" alt=" " width="800" height="520"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Successful execution log (The "cuda" device label is just PyTorch being PyTorch—it sometimes shows this even for non-CUDA devices!)&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;It worked!&lt;/strong&gt; The error completely disappeared, and the integrated GPU was humming along running inference. VRAM usage: 7GB. Single image inference: about 8 seconds. Pretty lightweight performance, I'd say. Real-time video is out of the question, though. (Man, I really want a high-end NVIDIA GPU...)&lt;/p&gt;

&lt;p&gt;The Ryzen AI Max+ 395 is built for much heavier workloads, but there's something satisfying about watching it breeze through a lightweight model like SAM 3. Just confirming that "the latest image models can run on AMD hardware" is a win for today.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;The lesson from this troubleshooting adventure: &lt;strong&gt;"Don't just read the error logs—examine the actual folder structure too."&lt;/strong&gt; Basic stuff, right?&lt;/p&gt;

&lt;p&gt;Powerful hardware like the Ryzen AI Max+ 395 is in a transitional period where the software ecosystem (especially Windows ROCm) hasn't caught up with hardware evolution. However, as this case shows, there are many situations where "the files exist, but the paths aren't configured correctly." Don't give up—dig through those directories and you might find the solution.&lt;/p&gt;

&lt;p&gt;To all AMD users struggling with this same error: give this "folder transplant surgery" a try. Here's to comfortable (and slightly overpowered) local AI adventures!&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;If you have feedback on this article or requests for "truly heavy models" you'd like me to test on the Ryzen AI Max+ 395, drop a comment below!&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Next time, I'll be posting about combining SAM 3 with IoT cameras (ESP32-based), so stay tuned!&lt;/strong&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>SAM 3 Is Here: Meta's Latest Vision AI Can Now Understand Your Words</title>
      <dc:creator>ARmedia</dc:creator>
      <pubDate>Thu, 18 Dec 2025 07:43:27 +0000</pubDate>
      <link>https://forem.com/sns_ar/sam-3-is-here-metas-latest-vision-ai-can-now-understand-your-words-12ll</link>
      <guid>https://forem.com/sns_ar/sam-3-is-here-metas-latest-vision-ai-can-now-understand-your-words-12ll</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;I've been battling with AI (Claude) for 14 hours a day. Couldn't be happier.&lt;/p&gt;

&lt;p&gt;— Akio Shiki (@ar_akio) &lt;a href="https://twitter.com/ar_akio/status/1980185472078070040?ref_src=twsrc%5Etfw" rel="noopener noreferrer"&gt;October 20, 2025&lt;/a&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffogkw4y4r5g65561cgeq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffogkw4y4r5g65561cgeq.png" alt=" " width="800" height="581"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://ai.meta.com/sam3/" rel="noopener noreferrer"&gt;Meta Official Site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Hi there! I'm Akio, an engineer at an AI development startup.&lt;/p&gt;

&lt;p&gt;In November 2025, Meta quietly dropped &lt;strong&gt;SAM 3 (Segment Anything Model 3)&lt;/strong&gt;. Have you had a chance to try it yet?&lt;/p&gt;

&lt;p&gt;"Wait, didn't SAM 2 just come out?" "What's new this time?"&lt;/p&gt;

&lt;p&gt;If you're asking these questions, you're not alone. But here's the thing—SAM 3 isn't just an incremental update or minor accuracy improvement. It represents a fundamental leap toward &lt;strong&gt;true multimodal segmentation&lt;/strong&gt;, making the old "click to segment" workflow feel like ancient history.&lt;/p&gt;

&lt;p&gt;In this article, I'll break down what makes SAM 3 so impressive based on the official GitHub and Hugging Face releases. And in the second half, I'll give you a sneak peek at &lt;strong&gt;our successful local implementation using the AMD Ryzen AI Max+ 395&lt;/strong&gt;—complete with screenshots from our dev environment.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Quick Refresher: What Is SAM?
&lt;/h2&gt;

&lt;p&gt;Let's briefly look back at the Segment Anything Model lineage:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;SAM 1 (2023):&lt;/strong&gt; Introduced as a zero-shot model that could segment any object in an image with just clicks or bounding boxes. It revolutionized segmentation tasks overnight.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;SAM 2 (2024):&lt;/strong&gt; Extended capabilities to video, enabling object tracking across frames. This opened up new possibilities for video editing and analysis.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And now, we have &lt;strong&gt;SAM 3&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Makes SAM 3 Revolutionary: 3 Key Advancements
&lt;/h2&gt;

&lt;p&gt;After diving into the official repository (&lt;a href="https://github.com/facebookresearch/sam3" rel="noopener noreferrer"&gt;facebookresearch/sam3&lt;/a&gt;) and demos, the direction of evolution is crystal clear.&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Just Tell It What You Want: Open Vocabulary Segmentation
&lt;/h3&gt;

&lt;p&gt;This is the headline feature—and as an engineer, it's what excites me most.&lt;/p&gt;

&lt;p&gt;Previous SAM versions required you to specify &lt;em&gt;where&lt;/em&gt; to segment (via clicks or bounding boxes). SAM 3 natively understands &lt;strong&gt;text prompts&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;For example, given street footage:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Type &lt;strong&gt;"red car"&lt;/strong&gt; and it detects and masks every red car in the frame.&lt;/li&gt;
&lt;li&gt;Say &lt;strong&gt;"yellow school bus"&lt;/strong&gt; and it instantly identifies and tracks it.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This means detection, segmentation, and tracking are now fully unified. You no longer need to tell the AI &lt;em&gt;where&lt;/em&gt; something is—it understands &lt;em&gt;what&lt;/em&gt; you're describing and connects that to the visual information automatically.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1n1oj480j7wxr1ch2g59.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1n1oj480j7wxr1ch2g59.png" width="800" height="496"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Type "impala" and it segments only the impalas&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  2. A Unified Vision Foundation Across Images, Video, and 3D
&lt;/h3&gt;

&lt;p&gt;SAM 3 completely breaks down the barrier between still images and video.&lt;/p&gt;

&lt;p&gt;Using a shared vision backbone, it performs object detection on individual frames while maintaining consistent tracking across the temporal axis.&lt;/p&gt;

&lt;p&gt;Even more exciting is the &lt;strong&gt;3D reconstruction&lt;/strong&gt; capability. Sometimes called "SAM 3D," this feature enables not just 2D segmentation but also estimation of an object's three-dimensional shape from images or video. This opens up real possibilities for XR (AR/VR) development and robotics applications.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmzivc8zag7nh0j67t5wq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmzivc8zag7nh0j67t5wq.png" width="800" height="251"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;SAM 3's architecture integrating image, video, and prompt processing&lt;/em&gt;&lt;br&gt;
&lt;em&gt;(Source: &lt;a href="https://github.com/facebookresearch/sam3?tab=readme-ov-file" rel="noopener noreferrer"&gt;Meta AI GitHub&lt;/a&gt;)&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Impressive Efficiency and Optimization
&lt;/h3&gt;

&lt;p&gt;More features usually means heavier models, but SAM 3 bucks this trend with optimized inference efficiency. Benchmarks on Meta's "SA-Co Dataset" show it outperforming previous models in accuracy while being designed with edge device deployment in mind.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Startup Perspective: Why SAM 3 Matters
&lt;/h2&gt;

&lt;p&gt;For AI startups like ours, SAM 3 means &lt;strong&gt;dramatically faster development cycles&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Previously, detecting specific objects—say, a particular component in a factory or a specific crop variety—required collecting massive labeled datasets and fine-tuning dedicated models (like YOLO variants).&lt;/p&gt;

&lt;p&gt;With SAM 3's &lt;strong&gt;open vocabulary capability&lt;/strong&gt;, you can simply prompt "damaged component" or "ripe tomato" and get high-accuracy detection and segmentation with zero-shot inference—no additional training required.&lt;/p&gt;

&lt;p&gt;This has the potential to compress PoC timelines from months to &lt;strong&gt;days&lt;/strong&gt;. For startups where speed-to-proposal is everything, this is an incredibly powerful advantage.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sneak Peek: Our Local Implementation
&lt;/h2&gt;

&lt;p&gt;Now for the technical deep-dive. Our lab has already completed a local implementation of SAM 3.&lt;/p&gt;

&lt;p&gt;Running on-premises rather than through cloud APIs is crucial for security, latency, and cost considerations.&lt;/p&gt;

&lt;p&gt;Our hardware of choice: the beast that is the &lt;strong&gt;AMD Ryzen AI Max+ 395&lt;/strong&gt;.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;CPU:&lt;/strong&gt; 16-core Zen 5 (Strix Halo)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory:&lt;/strong&gt; 128GB LPDDR5x (8000MT/s)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Overall TOPS:&lt;/strong&gt; Up to 126 TOPS&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Conventional wisdom says running massive models like SAM 3 requires expensive GPU servers like the H100. However, by leveraging the Ryzen AI's unified memory architecture with its generous memory capacity, we've &lt;strong&gt;successfully run SAM 3 smoothly in a local environment without sending any data to the cloud&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;That said, the Ryzen AI Max+ 395's main strength is its unified memory architecture, making it ideal for &lt;strong&gt;memory-hungry workloads like running gpt-oss-120b locally at low cost&lt;/strong&gt;. For use cases like this one that require lightweight memory footprint plus speed, an NVIDIA GPU (including consumer-grade options) is probably the better fit.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvtea9l338f1griwvkbou.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvtea9l338f1griwvkbou.png" width="800" height="294"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Running SAM 3 locally on the Ryzen AI Max+ 395. Inference is impressively fast.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;We're currently developing a system that integrates this setup with IoT devices (edge cameras) for real-time "segment by description" detection.&lt;/p&gt;

&lt;p&gt;The detailed implementation guide for this &lt;strong&gt;SAM 3 × Ryzen AI Max+ 395 × IoT&lt;/strong&gt; stack—including code and benchmark results—will be covered in an upcoming article. Follow along to stay updated!&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Vision AI Enters the "Understanding" Phase
&lt;/h2&gt;

&lt;p&gt;SAM 3 isn't just a segmentation tool. It's a model that serves as "eyes" capable of understanding the world through language and perceiving it spatially.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;GitHub:&lt;/strong&gt; &lt;a href="https://github.com/facebookresearch/sam3" rel="noopener noreferrer"&gt;facebookresearch/sam3&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Hugging Face:&lt;/strong&gt; &lt;a href="https://huggingface.co/facebook/sam3" rel="noopener noreferrer"&gt;facebook/sam3&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Demo:&lt;/strong&gt; &lt;a href="https://aidemos.meta.com/segment-anything" rel="noopener noreferrer"&gt;Segment Anything Demo&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I encourage you to try the official demo yourself. Prepare to be impressed by the accuracy. AI is evolving faster than most of us realize.&lt;/p&gt;

&lt;p&gt;If you have thoughts, feedback, or specific requests like &lt;strong&gt;"I'd love to know more about X aspect of the Ryzen AI implementation!"&lt;/strong&gt;—drop a comment below. I'll take your input into account for the next article!&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;We're hiring!&lt;/strong&gt; We're looking for engineers who want to tackle real-world AI implementation with cutting-edge technology. Interested? Check out the link in my profile!&lt;/p&gt;

</description>
      <category>ai</category>
      <category>deeplearning</category>
      <category>news</category>
    </item>
  </channel>
</rss>
