<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Sachin Myadam</title>
    <description>The latest articles on Forem by Sachin Myadam (@sachin_myadam).</description>
    <link>https://forem.com/sachin_myadam</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3700449%2F049b8500-2fda-4858-9557-d86d494e57c9.jpg</url>
      <title>Forem: Sachin Myadam</title>
      <link>https://forem.com/sachin_myadam</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/sachin_myadam"/>
    <language>en</language>
    <item>
      <title>The Big News: Siri's New Smart Brain is Here!</title>
      <dc:creator>Sachin Myadam</dc:creator>
      <pubDate>Tue, 13 Jan 2026 05:50:11 +0000</pubDate>
      <link>https://forem.com/sachin_myadam/the-big-news-siris-new-smart-brain-is-here-amb</link>
      <guid>https://forem.com/sachin_myadam/the-big-news-siris-new-smart-brain-is-here-amb</guid>
      <description>&lt;p&gt;&lt;strong&gt;iOS 20: Siri Integrates Google Gemini in Hybrid AI Push&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Apple has fundamentally restructured Siri’s backend for iOS 20, confirming a shift from a strictly  Apple-owned stack to a hybrid architecture that integrates Google’s Gemini Large Language Model (LLM). This move marks a shift in Apple’s AI strategy, prioritizing deployment speed over vertical integration.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Architecture
&lt;/h2&gt;

&lt;p&gt;The new system operates on a tiered routing mechanism designed to balance latency with computational power. Apple defines this as a split-inference model.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Tier 1 (Local Processing):&lt;/strong&gt; For system controls, Personal Information Management (PIM) data, and low-latency tasks, the iPhone utilizes "Apple Foundation Models." These run entirely on-device or via Apple’s Private Cloud Compute, ensuring that sensitive user data remains within Apple’s controlled ecosystem.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Tier 2 (Cloud Inference):&lt;/strong&gt; When the system detects a query requiring complex reasoning—such as summarizing documents or multi-step orchestration—it routes the request to Google Gemini.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;This architecture leverages Gemini’s estimated 1.2 trillion parameters to handle heavy processing loads that Apple’s internal models cannot support. To bridge these two worlds, Apple employs a Privacy Buffer Layer.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This proxy strips request headers and anonymizes user identifiers before transmission, ensuring Google processes the data statelessly without logging or training on the session.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Strategic Analysis&lt;/strong&gt;&lt;br&gt;
Industry observers characterize this integration as a necessary "brain transplant" for Siri. For years, Siri operated within a "walled garden," relying on internal datasets that limited its ability to parse context or manage unstructured text. By licensing Gemini for an estimated $1 billion annually, Apple acknowledges the reality: its internal "AJ" models excel at device utility but lag behind competitors in generative reasoning.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;This partnership represents a pragmatic concession. Rather than waiting years for its internal R&amp;amp;D to match GPT-4o standards, Apple bought immediate competence. This allows Siri to finally deliver on functional promises, such as cross-application scripting and screen context parsing, features that previously stalled under legacy architecture.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;strong&gt;Market Impact&lt;/strong&gt;&lt;br&gt;
This alliance immediately reshapes the AI hardware market. With Apple projecting 250 million AI-capable devices in circulation by late 2025, Google gains extensive access to the premium mobile segment. For hardware competitors, this sets a new baseline for mobile AI: a hybrid approach where the device handles the basics, and a partner model handles the intelligence. While this creates a dependency on a direct competitor, it secures the iPhone's position as a viable terminal for advanced AI interactions.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>gemini</category>
      <category>ios</category>
      <category>news</category>
    </item>
    <item>
      <title>Running Llama 3 Locally with Apple MLX: A Guide to Unified Memory</title>
      <dc:creator>Sachin Myadam</dc:creator>
      <pubDate>Thu, 08 Jan 2026 19:33:45 +0000</pubDate>
      <link>https://forem.com/sachin_myadam/running-llama-3-locally-with-apple-mlx-a-guide-to-unified-memory-5dbb</link>
      <guid>https://forem.com/sachin_myadam/running-llama-3-locally-with-apple-mlx-a-guide-to-unified-memory-5dbb</guid>
      <description>&lt;h2&gt;
  
  
  I. Why Run Local LLMs?
&lt;/h2&gt;

&lt;p&gt;For years, running large AI models meant paying for cloud GPUs or worrying about data privacy. That has changed.&lt;/p&gt;

&lt;p&gt;Running LLMs locally on your own hardware solves three specific problems: latency, privacy, and cost. You don't need an internet connection, your data never leaves your device, and you stop paying per-token API fees. With Apple’s MLX framework, this is now practical on consumer hardware.&lt;/p&gt;

&lt;h2&gt;
  
  
  II. Context: From Cloud to Silicon
&lt;/h2&gt;

&lt;p&gt;Machine learning used to require massive dedicated clusters. When deep learning took over with TensorFlow and PyTorch, the reliance on NVIDIA GPUs became the standard.&lt;/p&gt;

&lt;p&gt;In 2020, Apple changed the architecture with the M1 chip and Unified Memory. In late 2023, they released MLX, a framework designed specifically for this architecture. It allows developers to run models efficiently without the overhead of traditional tools.&lt;/p&gt;

&lt;p&gt;Meta’s open-source Llama series accelerated this shift. We saw the release of Llama 3 in 2024, followed by the major Llama 4 release in April 2025. These models are now efficient enough to run on a laptop while outperforming older server-grade models.&lt;/p&gt;

&lt;h3&gt;
  
  
  III. The Hardware: M3 Ultra and Unified Memory
&lt;/h3&gt;

&lt;p&gt;The bottleneck in AI isn't always compute speed; often, it is memory bandwidth. Traditional PCs separate CPU RAM and GPU VRAM. To process a large model, you have to move data between them, which is slow.&lt;/p&gt;

&lt;p&gt;Apple’s Unified Memory architecture lets the CPU and GPU access the same memory pool. The M3 Ultra, for example, supports up to 128GB of Unified Memory with a bandwidth of 819 GB/s. This allows you to load massive 70B or even quantized 405B parameter models directly into RAM, something that is impossible on most consumer dedicated GPUs.&lt;/p&gt;

&lt;p&gt;January 2026 Update: While this guide focuses on the M3 Ultra for its massive memory bandwidth (819 GB/s), Apple’s new M5 architecture (released Oct 2025) brings even faster specialized Neural Accelerators. The good news? The same MLX principles and code provided below work perfectly on M5-powered MacBooks as well.&lt;/p&gt;

&lt;h2&gt;
  
  
  IV. Llama 3 Performance
&lt;/h2&gt;

&lt;p&gt;Llama 3 remains a solid baseline for local development. It handles reasoning, coding, and multi-language tasks well. While newer models exist, the 8B version of Llama 3 is the perfect starting point for testing local inference because it fits comfortably in 8GB or 16GB of RAM.&lt;/p&gt;

&lt;h2&gt;
  
  
  V. Cloud vs. Local
&lt;/h2&gt;

&lt;p&gt;Cloud GPUs like the H100 are still faster for massive training jobs. However, for inference (running the model), the MacBook Pro is surprisingly competitive. The main advantage is workflow: you can iterate on code, test prompts, and debug applications offline without waiting for server queues or managing API keys.&lt;/p&gt;

&lt;h2&gt;
  
  
  VI. Step-by-Step Tutorial
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Install Dependencies
&lt;/h3&gt;

&lt;p&gt;You need a Python environment (3.11+ is recommended). Open your terminal and run:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;pip &lt;span class="nb"&gt;install &lt;/span&gt;mlx-lm
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  2. Run from CLI
&lt;/h3&gt;

&lt;p&gt;To quickly test if the model works, use the command line interface. This pulls the 4-bit quantized version, which is optimized for speed and memory.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;python &lt;span class="nt"&gt;-m&lt;/span&gt; mlx_lm.generate &lt;span class="se"&gt;\&lt;/span&gt;
  &lt;span class="nt"&gt;--model&lt;/span&gt; mlx-community/Meta-Llama-3-8B-Instruct-4bit &lt;span class="se"&gt;\&lt;/span&gt;
  &lt;span class="nt"&gt;--prompt&lt;/span&gt; &lt;span class="s2"&gt;"Write a Python script to sort a list of dictionaries."&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  3. Run with Python
&lt;/h3&gt;

&lt;p&gt;For actual development, use this script:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;
&lt;span class="kn"&gt;from&lt;/span&gt; &lt;span class="n"&gt;mlx_lm&lt;/span&gt; &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;load&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;generate&lt;/span&gt;

&lt;span class="c1"&gt;# Load the model and tokenizer
&lt;/span&gt;&lt;span class="n"&gt;model&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;tokenizer&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nf"&gt;load&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;mlx-community/Meta-Llama-3-8B-Instruct-4bit&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;# Generate a response
&lt;/span&gt;&lt;span class="n"&gt;response&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nf"&gt;generate&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="n"&gt;model&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;tokenizer&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;prompt&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;Explain Unified Memory in one sentence.&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;verbose&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="bp"&gt;True&lt;/span&gt;
&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="nf"&gt;print&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;response&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  VII. What’s Next
&lt;/h2&gt;

&lt;p&gt;MLX is still evolving. We are seeing better integration with the Neural Engine and support for more complex quantization methods. The focus for 2026 is on "quantized models"—making these large networks smaller without losing accuracy, so they run faster on standard laptops.&lt;/p&gt;

&lt;h2&gt;
  
  
  VIII. Final Thoughts
&lt;/h2&gt;

&lt;p&gt;You no longer need a server farm to build AI applications. With Unified Memory and MLX, a MacBook is a legitimate platform for AI engineering. It’s cheaper, private, and capable of handling real-world production models.&lt;/p&gt;

</description>
      <category>python</category>
      <category>ai</category>
      <category>apple</category>
      <category>programming</category>
    </item>
  </channel>
</rss>
