<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Aditi Bindal</title>
    <description>The latest articles on Forem by Aditi Bindal (@aditi_b).</description>
    <link>https://forem.com/aditi_b</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2426053%2F079f8e5d-9e23-483c-8386-c68a45b5c3d0.jpg</url>
      <title>Forem: Aditi Bindal</title>
      <link>https://forem.com/aditi_b</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/aditi_b"/>
    <language>en</language>
    <item>
      <title>A Step-by-Step Guide to Install Qwen3-Next 80B</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 22 Sep 2025 07:15:58 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/a-step-by-step-guide-to-install-qwen3-next-80b-3dho</link>
      <guid>https://forem.com/nodeshiftcloud/a-step-by-step-guide-to-install-qwen3-next-80b-3dho</guid>
      <description>&lt;p&gt;If you're relentlessly following AI advancements, one thing can be clearly observed, the trend has been simple: go bigger. However, the new Qwen3-Next-80B series models challenges this paradigm by focusing on groundbreaking efficiency rather than just raw scale. This model represents a monumental leap forward, delivering the performance of a much larger model with a fraction of the computational cost. At its core is a revolutionary Hybrid Attention mechanism, to process ultra-long context lengths, natively supporting 262,144 tokens and extensible to over a million. This is paired with a High-Sparsity Mixture-of-Experts (MoE) architecture that keeps a staggering 80 billion total parameters on tap while only activating 3 billion at any given time. The result? Drastically reduced computational load, leading to inference speeds up to 10 times faster than its predecessors on long-context tasks. With additional enhancements like Multi-Token Prediction for accelerated performance and advanced stability optimizations, Qwen3-Next-80B proves its worth by outperforming models like Qwen3-32B with only 10% of the training cost and performing on par with models on key reasoning, coding, and alignment benchmarks.&lt;/p&gt;

&lt;p&gt;In this article we'll dive into the simple and straightforward walkthrough showing the installation, setup and usage of this model.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 2x H200s or 4x H100s&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 1TB+ (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 160GB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install Qwen3-Next-80B-A3B Locally
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by NodeShift since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 2x H200 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 5 TB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F78ex5301m0jmnmar6gbe.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F78ex5301m0jmnmar6gbe.png" alt="Image-step3-1" width="800" height="271"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 2x H200 140GB GPU node with 192vCPUs/504GB RAM/5TB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3cbh4j66eg49118de0mi.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3cbh4j66eg49118de0mi.png" alt="Image-step4-1" width="800" height="560"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk810i78g0piq7z2jxu8j.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk810i78g0piq7z2jxu8j.png" alt="Image-step5-3" width="800" height="397"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmvw5wtv572xfkv01hoyp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmvw5wtv572xfkv01hoyp.png" alt="Image-step6-1" width="800" height="326"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n qwen python=3.11 -y &amp;amp;&amp;amp; conda activate qwen
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flriuy01pg5znv68e8o5c.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flriuy01pg5znv68e8o5c.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Install required dependencies.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121 
pip install git+https://github.com/huggingface/transformers.git@main
pip install git+https://github.com/huggingface/accelerate
pip install huggingface_hub
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe7x6myqz7f0exmlawxik.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe7x6myqz7f0exmlawxik.png" alt="Image-step7-2" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Login to Hugging Face with HF READ token.&lt;/p&gt;

&lt;p&gt;This is a gated model, make sure to get access granted from the model card.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;hf auth login
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq8ft3orv1ku3qro1jdpn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq8ft3orv1ku3qro1jdpn.png" alt="Image-step7-3" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;4) Install and run jupyter notebook.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda install -c conda-forge --override-channels notebook -y
conda install -c conda-forge --override-channels ipywidgets -y
jupyter notebook --allow-root
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;5) If you’re on a remote machine (e.g., NodeShift GPU), you’ll need to do SSH port forwarding in order to access the jupyter notebook session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_PORT&amp;gt;&lt;/code&gt; with the PORT allotted to your remote server (For the NodeShift server – you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;ssh -L 8888:localhost:8888 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnb4ojg5ic1gib6uigxtb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnb4ojg5ic1gib6uigxtb.png" alt="Image-step7-4" width="800" height="231"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After this copy the URL you received in your remote server:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz74xo2mne7xx5flisla6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz74xo2mne7xx5flisla6.png" alt="Image-step7-5" width="800" height="267"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And paste this on your local browser to access the Jupyter Notebook session.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Download and Run the model
&lt;/h3&gt;

&lt;p&gt;1) Open a Python notebook inside Jupyter.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhjeb8u2ttf96pxi3enag.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhjeb8u2ttf96pxi3enag.png" alt="Image-step8-1" width="800" height="447"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Download the model checkpoints.&lt;/p&gt;

&lt;p&gt;To download the thinking model, just replace the model_name value with &lt;code&gt;"Qwen/Qwen3-Next-80B-A3B-Thinking"&lt;/code&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "Qwen/Qwen3-Next-80B-A3B-Instruct"

# load the tokenizer and the model
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    dtype="auto",
    device_map="auto",
)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgsnbi60mg9ihs7qyy6k2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgsnbi60mg9ihs7qyy6k2.png" alt="Image-step8-2" width="800" height="445"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Run the model for inference.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;# prepare the model input
prompt = "Give me a short introduction to large language model."
messages = [
    {"role": "user", "content": prompt},
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True,
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

# conduct text completion
generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=16384,
)
output_ids = generated_ids[0][len(model_inputs.input_ids[0]):].tolist() 

content = tokenizer.decode(output_ids, skip_special_tokens=True)

print("content:", content)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;And for inferencing with the thinking model, use the following snippet:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;# prepare the model input
prompt = "Give me a short introduction to large language model."
messages = [
    {"role": "user", "content": prompt},
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True,
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

# conduct text completion
generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=32768,
)
output_ids = generated_ids[0][len(model_inputs.input_ids[0]):].tolist() 

# parsing thinking content
try:
    # rindex finding 151668 (&amp;lt;/think&amp;gt;)
    index = len(output_ids) - output_ids[::-1].index(151668)
except ValueError:
    index = 0

thinking_content = tokenizer.decode(output_ids[:index], skip_special_tokens=True).strip("\n")
content = tokenizer.decode(output_ids[index:], skip_special_tokens=True).strip("\n")

print("thinking content:", thinking_content) # no opening &amp;lt;think&amp;gt; tag
print("content:", content)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx2tngy25r69t5d6tkd11.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx2tngy25r69t5d6tkd11.png" alt="Image-step8-3" width="800" height="419"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;The Qwen3-Next-80B model represents a shift in AI development, prioritizing efficiency over raw scale through its Hybrid Attention and High-Sparsity Mixture-of-Experts (MoE) architecture. This allows it to achieve high performance with a fraction of the computational load, enabling it to handle massive context lengths and accelerate inference speeds. NodeShift Cloud plays a crucial role in making this advanced technology accessible and practical by providing a cost-effective, secured platform for deploying and running such compute-intensive models. By offering affordable GPU resources, NodeShift Cloud democratizes access to state-of-the-art AI, allowing developers and businesses to leverage the power of models like Qwen3-Next-80B without the prohibitive costs and infrastructure management typically associated with large-scale AI.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>tutorial</category>
      <category>opensource</category>
      <category>qwen</category>
    </item>
    <item>
      <title>Generate Expressive, Long Form Multi-Speaker Audios &amp; Podcasts with Microsoft's VibeVoice</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Wed, 03 Sep 2025 16:48:39 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/generate-expressive-long-form-multi-speaker-audios-podcasts-with-microsofts-vibevoice-4hfp</link>
      <guid>https://forem.com/nodeshiftcloud/generate-expressive-long-form-multi-speaker-audios-podcasts-with-microsofts-vibevoice-4hfp</guid>
      <description>&lt;p&gt;If you're looking for an open-source text-to-speech system that can generate podcasts, audiobooks, or multi-speaker conversations that actually sound real, Microsoft’s VibeVoice is a model you’ll want to try. Unlike traditional TTS systems that often feel robotic, inconsistent, or restricted to short clips, VibeVoice is designed from the ground up to produce expressive, long-form, multi-speaker audio with remarkable naturalness and flow. It can synthesize speech lasting up to 90 minutes and seamlessly handle up to four distinct speakers, an impressive upgrade over most existing models that struggle to maintain quality beyond a few minutes or across more than two voices. What makes this possible is its continuous speech tokenizers (acoustic and semantic) that operate at a very low frame rate (7.5 Hz), preserving audio richness while drastically reducing computation. On top of this, the model uses a next-token diffusion framework, powered by a Qwen2.5-based LLM, to understand dialogue context and generate nuanced turn-taking, while a lightweight diffusion head ensures high-fidelity acoustic detail. The result: smooth, consistent, and lifelike conversations that feel like they were recorded, not generated.&lt;/p&gt;

&lt;p&gt;In this guide, we have covered a simple and step-by-step walkthrough of how to get this model up and running locally or in GPU-accelerated environments.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 1x RTX 4090 or 1x RTX A6000&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 50GB (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 16GB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install and run VibeVoice
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by &lt;a href="https://nodeshift.com" rel="noopener noreferrer"&gt;NodeShift&lt;/a&gt; since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 1x A100 SXM4 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 100GB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" alt="Image-step3-1" width="800" height="227"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 1x RTXA6000 GPU node with 64vCPUs/63GB RAM/200GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" alt="Image-step4-1" width="800" height="394"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In our previous blogs, we used pre-built images from the Templates tab when creating a Virtual Machine. However, for running a CUDA dependent application like VibeVoice, we need a more customized environment with full CUDA development capabilities. That’s why, in this case, we switched to the Custom Image tab and selected a specific Docker image that meets all runtime and compatibility requirements.&lt;/p&gt;

&lt;p&gt;We chose the following image:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;nvidia/cuda:12.1.1-devel-ubuntu22.04
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This image is essential because it includes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Full CUDA toolkit (including nvcc)&lt;/li&gt;
&lt;li&gt;Proper support for building and running GPU-based applications&lt;/li&gt;
&lt;li&gt;Compatibility with CUDA 12.1.1 required by certain model operations&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  Launch Mode
&lt;/h4&gt;

&lt;p&gt;We selected:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Interactive shell server
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This gives us SSH access and full control over terminal operations — perfect for installing dependencies, running benchmarks, and launching models.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frhol4bqjn2f6zi2sv4ab.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frhol4bqjn2f6zi2sv4ab.png" alt="Image-step5-2" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h4&gt;
  
  
  Docker Repository Authentication
&lt;/h4&gt;

&lt;p&gt;We left all fields &lt;strong&gt;empty&lt;/strong&gt; here.&lt;/p&gt;

&lt;p&gt;Since the Docker image is publicly available on Docker Hub, no login credentials are required.&lt;/p&gt;

&lt;h4&gt;
  
  
  Identification
&lt;/h4&gt;

&lt;p&gt;Template Name:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;nvidia/cuda:12.1.1-devel-ubuntu22.04
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwt7p0dw5c44p7u3axrs9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwt7p0dw5c44p7u3axrs9.png" alt="Image-step5-3" width="800" height="376"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That’s it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-4" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" alt="Image-step5-5" width="800" height="389"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnqwto9145nkb27vd3d71.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnqwto9145nkb27vd3d71.png" alt="Image-step6-1" width="800" height="360"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n vibe python=3.11 -y &amp;amp;&amp;amp; conda activate vibe
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa44mazey6p4kary0n39p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa44mazey6p4kary0n39p.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Clone the official repository and move inside the project directory.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;git clone https://github.com/microsoft/VibeVoice.git
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fguwl3vseqa9vadrmvjkg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fguwl3vseqa9vadrmvjkg.png" alt="Image-step7-2" width="800" height="153"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Install required dependencies.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install -e .
pip install flash-attn --no-build-isolation
apt update &amp;amp;&amp;amp; apt install ffmpeg -y
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyux491lk91d70ve2hqir.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyux491lk91d70ve2hqir.png" alt="Image-step7-3" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;4) Launch the Gradio demo. This will automatically download the model checkpoints as well.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;python demo/gradio_demo.py --model_path microsoft/VibeVoice-1.5B --share
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fib4ai6y8vtdm1v9nfz9u.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fib4ai6y8vtdm1v9nfz9u.png" alt="Image-step7-4" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flhf2qf7pmjiw0eu590fy.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flhf2qf7pmjiw0eu590fy.png" alt="Image-step7-5" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;5) If you're on a remote machine (e.g., NodeShift GPU), you'll need to do SSH port forwarding in order to access the Gradio session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_PORT&amp;gt;&lt;/code&gt; with the PORT allotted to your remote server (For the NodeShift server - you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;ssh -L 7860:localhost:7860 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;After this copy the URL you received in your remote server: &lt;a href="http://0.0.0.0:7860" rel="noopener noreferrer"&gt;http://0.0.0.0:7860&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And paste this on your local browser to access the Gradio session.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Run the model
&lt;/h3&gt;

&lt;p&gt;1) Once you access the Gradio interface, it will look like this:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv99jk5854w87o84sfqiw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv99jk5854w87o84sfqiw.png" alt="Image-step8-1" width="800" height="484"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Generate podcast from script.&lt;/p&gt;

&lt;p&gt;Put any script of your chocie, we’re using one of the &lt;a href="https://github.com/microsoft/VibeVoice/blob/main/demo/text_examples/4p_climate_45min.txt" rel="noopener noreferrer"&gt;example scripts&lt;/a&gt; given in the official repo.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffgm3ovx013se97c9o6nl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffgm3ovx013se97c9o6nl.png" alt="Image-step8-2" width="800" height="554"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) The app has started to stream generated podcast audio in real-time.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4pit993yejnqhe0y45kj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4pit993yejnqhe0y45kj.png" alt="Image-step8-3" width="800" height="690"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;VibeVoice stands out as a groundbreaking open-source TTS framework that combines continuous speech tokenizers, a Qwen2.5-powered LLM, and a diffusion head to deliver expressive, long-form, multi-speaker audio that feels astonishingly real. Its ability to generate up to 90 minutes of consistent, multi-voice speech makes it a powerful tool for creators and researchers alike. And while running it locally is a great way to get started, NodeShift makes the experience even smoother by providing GPU-accelerated environments, simplified deployment, and scalability out of the box, so you can focus on exploring and scaling with the model’s capabilities without worrying about complex infrastructure setup.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>microsoft</category>
      <category>podcast</category>
      <category>ai</category>
      <category>opensource</category>
    </item>
    <item>
      <title>A Step-by-Step Guide to Install DeepSeek V3.1</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 01 Sep 2025 18:21:46 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/a-step-by-step-guide-to-install-deepseek-v31-3dgg</link>
      <guid>https://forem.com/nodeshiftcloud/a-step-by-step-guide-to-install-deepseek-v31-3dgg</guid>
      <description>&lt;p&gt;DeepSeek has once again pushed the boundaries of what’s possible in open-source AI with the release of DeepSeek-V3.1, a next-generation hybrid model that seamlessly supports both thinking and non-thinking modes. Building on the foundation of its powerful V3 base checkpoint, this version introduces smarter tool calling, faster reasoning efficiency, and a more versatile chat template design that adapts effortlessly to different use cases. Its post-training optimization dramatically boosts performance in agent tasks and tool usage, making it a strong choice for developers working on automation, research assistance, and coding agents. Moreover, the model’s ability to process extended contexts has been expanded through a two-phase long context extension approach: a massive 10x increase in the 32K token phase to 630B tokens and a 3.3x increase in the 128K token phase to 209B tokens. Combined with training on the cutting-edge UE8M0 FP8 data format, DeepSeek-V3.1 not only ensures efficiency and scalability but also guarantees compatibility with modern microscaling data pipelines.&lt;/p&gt;

&lt;p&gt;Deploying a model of this caliber locally might seem daunting at first due to its substantial 671 billion parameters. However, Unsloth has made it entirely feasible. Unsloth has used selective quantization techniques to reduce the model's size without any significant loss of accuracy by targeting specific layers, such as the Mixture-of-Experts (MoE) layers, while preserving the precision of attention and other critical layers.&lt;/p&gt;

&lt;p&gt;In the following guide, we'll walk you through the step-by-step process of installing and running DeepSeek-V3.1 locally using LLaMA.cpp and Unsloth's dynamic quants, ensuring you can access its full potential efficiently and effectively.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The system requirements for running DeepSeek-V3.1 are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: Multiple H100s or H200s (count may vary across different bits)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 1TB+ (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Nvidia Cuda installed.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Disk Space requirements depending on the type of model are as follows:&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7rtfv01ia0jnqggux4nb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7rtfv01ia0jnqggux4nb.png" alt="Image-prequisites" width="800" height="426"&gt;&lt;/a&gt;&lt;br&gt;
Source: Unsloth&lt;/p&gt;

&lt;p&gt;We recommend you to take a screenshot of this chart and save it somewhere to quickly look up to the disk space prerequisites before trying a specific bit quantized version.&lt;/p&gt;

&lt;p&gt;For this article, we’ll download the 2.71-bit version (recommended).&lt;/p&gt;
&lt;h2&gt;
  
  
  Step-by-step process to install DeepSeek-V3.1 Locally
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by NodeShift since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;
&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;
&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 1x H200 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 200 GB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F78ex5301m0jmnmar6gbe.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F78ex5301m0jmnmar6gbe.png" alt="Image-step3-1" width="800" height="271"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 1x H100 SXM 80GB GPU node with 192vCPUs/80GB RAM/200GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffs8a0r7ibcqiyeu9v2oy.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffs8a0r7ibcqiyeu9v2oy.png" alt="Image-step4-1" width="800" height="560"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk810i78g0piq7z2jxu8j.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk810i78g0piq7z2jxu8j.png" alt="Image-step5-3" width="800" height="397"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa8bldddlk57tshry6ngf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa8bldddlk57tshry6ngf.png" alt="Image-step6-1" width="800" height="332"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Install and build LLaMA.cpp
&lt;/h3&gt;

&lt;p&gt;&lt;code&gt;llama.cpp&lt;/code&gt; is a C++ library for running LLaMA and other large language models efficiently on GPUs, CPUs and edge devices.&lt;/p&gt;

&lt;p&gt;We’ll first install &lt;code&gt;llama.cpp&lt;/code&gt; as we’ll use it to install and run DeepSeek-V3-0324.&lt;/p&gt;

&lt;p&gt;1) Start by creating a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n deepseek python=3.11 -y &amp;amp;&amp;amp; conda activate deepseek
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5hwf3q01mgtlwhxavyue.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5hwf3q01mgtlwhxavyue.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Once inside the environment, update the Ubuntu package source-list for fetching the latest repository updates and patches.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;br&gt;
apt-get update&lt;br&gt;
&lt;/code&gt;`&lt;/p&gt;

&lt;p&gt;3) Install dependencies for llama.cpp.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;&lt;code&gt;&lt;br&gt;
apt-get install pciutils build-essential cmake curl libcurl4-openssl-dev -y&lt;br&gt;
&lt;/code&gt;&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgotwcz9o08de7837bj33.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgotwcz9o08de7837bj33.png" alt="Image-step7-2" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;4) Clone the official repository of llama.cpp.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;&lt;code&gt;&lt;br&gt;
git clone https://github.com/ggml-org/llama.cpp&lt;br&gt;
&lt;/code&gt;&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frpg8mmhvabkqtol1ldex.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frpg8mmhvabkqtol1ldex.png" alt="Image-step7-3" width="800" height="167"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;5) Compile &lt;code&gt;llama.cpp&lt;/code&gt;‘s build files.&lt;/p&gt;

&lt;p&gt;In the below command, keep &lt;code&gt;-DGGML_CUDA=OFF&lt;/code&gt; if you’re running it on a non-GPU system. However, it’s recommended to keep it OFF, even if you’re on a GPU-based system, as it will allow llama.cpp’s compilation process to occur through CPU, which is faster in this case as compared to GPU-based compilation. In addition to being slow, compiling llama.cpp through GPU can sometimes throw unwanted errors.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;&lt;code&gt;&lt;br&gt;
cmake llama.cpp -B llama.cpp/build \&lt;br&gt;
    -DBUILD_SHARED_LIBS=OFF -DGGML_CUDA=OFF -DLLAMA_CURL=ON&lt;br&gt;
&lt;/code&gt;&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpobvl0sy0x4vp1ivg5h1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpobvl0sy0x4vp1ivg5h1.png" alt="Image-step7-4" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;6) Build &lt;code&gt;llama.cpp&lt;/code&gt; from the build directory.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;&lt;code&gt;&lt;br&gt;
cmake --build llama.cpp/build --config Release -j --clean-first --target llama-quantize llama-cli llama-gguf-split&lt;br&gt;
&lt;/code&gt;&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fym4abt71d49jjhwb13s6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fym4abt71d49jjhwb13s6.png" alt="Image-step7-5" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;7) Finally, we’ll copy all the executables from llama.cpp/build/bin/ that start with llama- into the llama.cpp directory.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;&lt;code&gt;&lt;br&gt;
cp llama.cpp/build/bin/llama-* llama.cpp&lt;br&gt;
&lt;/code&gt;&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Download the Model Files
&lt;/h3&gt;

&lt;p&gt;We’ll download the model files from Hugging Face using a Python script.&lt;/p&gt;

&lt;p&gt;1) To do that, let’s first install the Hugging Face Python packages.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;&lt;code&gt;&lt;br&gt;
pip install huggingface_hub hf_transfer&lt;br&gt;
&lt;/code&gt;&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;&lt;code&gt;huggingface_hub&lt;/code&gt;&lt;/strong&gt; – Provides an interface to interact with the Hugging Face Hub, allowing you to download, upload, and manage models, datasets, and other resources.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;&lt;code&gt;hf_transfer&lt;/code&gt;&lt;/strong&gt; – A tool optimized for faster uploads and downloads of large files (e.g., LLaMA, DeepSeek models) from the Hugging Face Hub using a more efficient transfer protocol.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;`&lt;br&gt;
Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcgdd0nqckkule24fogxo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcgdd0nqckkule24fogxo.png" alt="Image-step8-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Run the model installation script with Python.&lt;/p&gt;

&lt;p&gt;The script below will download all the specifical quant’s checkpoints from &lt;a href="https://huggingface.co/unsloth/DeepSeek-V3.1" rel="noopener noreferrer"&gt;unsloth/DeepSeek-V3.1&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;`&lt;br&gt;
python -c "import os; os.environ['HF_HUB_ENABLE_HF_TRANSFER']='0'; from huggingface_hub import snapshot_download; snapshot_download(repo_id='unsloth/DeepSeek-V3.1-GGUF', local_dir='unsloth/DeepSeek-V3.1-GGUF', allow_patterns=['*UD-Q2_K_XL*'])"&lt;br&gt;
`&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd8jthrsf9njkvdi9zyb3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd8jthrsf9njkvdi9zyb3.png" alt="Image-step8-2" width="800" height="242"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Depending on your GPU configuration, the download process can be slow and take some time. The installation might also seem stuck at some points, which is normal, so do not interrupt or kill the installation in between.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 9: Run the model for Inference
&lt;/h3&gt;

&lt;p&gt;Finally, once all checkpoints are downloaded, we can proceed to the inference part.&lt;/p&gt;

&lt;p&gt;In the below command, we’ll run the model with a prompt given inside a formatted template which will be run through LLaMA.cpp’s LLaMA-CLI tool. The prompt will ask the model to create a Flappy Bird game in Python with all the interface, logic, and controls.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;`&lt;br&gt;
./llama.cpp/llama-cli \&lt;br&gt;
    --model unsloth/DeepSeek-V3.1-GGUF/UD-Q2_K_XL/DeepSeek-V3.1-UD-Q2_K_XL-00001-of-00006.gguf \&lt;br&gt;
    --cache-type-k q4_0 \&lt;br&gt;
    --threads -1 \&lt;br&gt;
    --n-gpu-layers 99 \&lt;br&gt;
    --prio 3 \&lt;br&gt;
    --temp 0.6 \&lt;br&gt;
    --top_p 0.95 \&lt;br&gt;
    --min_p 0.01 \&lt;br&gt;
    --ctx-size 16384 \&lt;br&gt;
    --seed 3407 \&lt;br&gt;
    -ot ".ffn_.*_exps.=CPU" \&lt;br&gt;
    -no-cnv \&lt;br&gt;
    --prompt "&amp;lt;｜User｜&amp;gt;Create a Flappy Bird game in Python. You must include these things:\n1. You must use pygame.\n2. The background color should be randomly chosen and is a light shade. Start with a light blue color.\n3. Pressing SPACE multiple times will accelerate the bird.\n4. The bird's shape should be randomly chosen as a square, circle or triangle. The color should be randomly chosen as a dark color.\n5. Place on the bottom some land colored as dark brown or yellow chosen randomly.\n6. Make a score shown on the top right side. Increment if you pass pipes and don't hit them.\n7. Make randomly spaced pipes with enough space. Color them randomly as dark green or light brown or a dark gray shade.\n8. When you lose, show the best score. Make the text inside the screen. Pressing q or Esc will quit the game. Restarting is pressing SPACE again.\nThe final game should be inside a markdown section in Python. Check your code for errors and fix them before the final markdown section.&amp;lt;｜Assistant｜&amp;gt;"&lt;br&gt;
`&lt;/code&gt;&lt;br&gt;
Output:&lt;/p&gt;

&lt;p&gt;The model has started generating the code as shown below:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flb15xhpd0wel4j1kt2mx.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flb15xhpd0wel4j1kt2mx.png" alt="Image-step9-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Once the process is complete, it may end the output like this:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2ma3y5j4w9b5fxm6nh1r.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2ma3y5j4w9b5fxm6nh1r.png" alt="Image-step9-2" width="800" height="353"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As we run the code for the Flappy Bird game generated by DeepSeek-V3.1 through VSCode Editor, it opens a game panel as shown below (Note: Install pygame in your editor before running the code):&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhvduyozxu1wf5th8n9ak.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhvduyozxu1wf5th8n9ak.png" alt=" " width="655" height="1024"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;You can see the live demonstration of the game in the video attached on the original article &lt;a href="https://nodeshift.cloud/blog/a-step-by-step-guide-to-install-deepseek-v3-1?utm_source=devto&amp;amp;utm_medium=blog&amp;amp;utm_campaign=content_share" rel="noopener noreferrer"&gt;here&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;In this guide, we explored how DeepSeek-V3.1 elevates open-source AI with its hybrid thinking modes, smarter tool calling, faster reasoning, and extended long-context capabilities, all supported by efficient training techniques like FP8 scaling and Unsloth’s dynamic quantization. While deploying such a massive model locally with LLaMA.cpp is now more accessible, it still demands considerable compute resources. This is where NodeShip Cloud steps in, offering a seamless alternative with scalable, cost-effective GPU and compute infrastructure. By offloading deployment to NodeShip’s intuitive cloud platform, developers can unlock the full potential of DeepSeek-V3.1 without the burden of managing heavy local infrastructure, making experimentation, scaling, and production use both faster and simpler.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>deepseek</category>
      <category>ai</category>
      <category>llm</category>
      <category>opensource</category>
    </item>
    <item>
      <title>A Complete Setup Guide to Powerful AI Image Editing with Qwen-Image-Edit</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 01 Sep 2025 16:52:19 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/a-complete-setup-guide-to-powerful-ai-image-editing-with-qwen-image-edit-3kg1</link>
      <guid>https://forem.com/nodeshiftcloud/a-complete-setup-guide-to-powerful-ai-image-editing-with-qwen-image-edit-3kg1</guid>
      <description>&lt;p&gt;Image editing has always required a delicate balance between precision and creativity, and that’s exactly what Qwen-Image-Edit delivers. Built on the robust 20B Qwen-Image model, this cutting-edge tool takes image editing to the next level by combining semantic control (powered by Qwen2.5-VL) with appearance control (via its VAE Encoder). This dual-system approach allows users to seamlessly perform both low-level edits, like adding or removing objects while keeping the rest of the image untouched, and high-level transformations, such as rotating objects, transferring artistic styles, or even creating new concepts entirely. What truly sets Qwen-Image-Edit apart, however, is its precise text editing capability, enabling direct modification of text in English and Chinese while preserving the original font, size, and style. &lt;/p&gt;

&lt;p&gt;If you’re looking for an image editing model that’s powerful, versatile, and incredibly easy to use, Qwen-Image-Edit is a must-try. Let's see how to get it up and running on your machine.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 1x H100&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 50 GB (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 64 GB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install and run Qwen Image Edit
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by &lt;a href="https://nodeshift.com" rel="noopener noreferrer"&gt;NodeShift&lt;/a&gt; since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 1x H100 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 200 GB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2nwqblqu9dtn5vbnnpvm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2nwqblqu9dtn5vbnnpvm.png" alt="Image-step3-1" width="800" height="277"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 1x H100 SXM 80GB GPU node with 192vCPUs/80GB RAM/200GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F33ctuz0kf0n28kilc7zj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F33ctuz0kf0n28kilc7zj.png" alt="Image-step4-1" width="800" height="356"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Filngygf82xfgk2o7lyxv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Filngygf82xfgk2o7lyxv.png" alt="Image-step5-3" width="800" height="397"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqckok7vzis7m6g0pecxw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqckok7vzis7m6g0pecxw.png" alt="Image-step6-1" width="800" height="378"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n qwen python=3.11 -y &amp;amp;&amp;amp; conda activate qwen
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffytezgl8rh29jhqc9dk8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffytezgl8rh29jhqc9dk8.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Install required dependencies.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121
pip install git+https://github.com/huggingface/diffusers
pip install transformers accelerate gradio pillow
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fluaqs2mc3pgm6b65qusv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fluaqs2mc3pgm6b65qusv.png" alt="Image-step7-2" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Install and run jupyter notebook.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda install -c conda-forge --override-channels notebook -y
conda install -c conda-forge --override-channels ipywidgets -y
jupyter notebook --allow-root
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;4) If you’re on a remote machine (e.g., NodeShift GPU), you’ll need to do SSH port forwarding in order to access the jupyter notebook session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;` with the PORT allotted to your remote server (For the NodeShift server – you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;&lt;code&gt;&lt;br&gt;
ssh -L 8888:localhost:8888 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;br&gt;
&lt;/code&gt;&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2r8j1owgltt9aq3dt6yq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2r8j1owgltt9aq3dt6yq.png" alt="Image-step7-3" width="800" height="231"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After this copy the URL you received in your remote server:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmfvxbys411rncwon4x0s.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmfvxbys411rncwon4x0s.png" alt="Image-step7-4" width="800" height="267"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And paste this on your local browser to access the Jupyter Notebook session.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Download and Run the model
&lt;/h3&gt;

&lt;p&gt;1) Open a Python notebook inside Jupyter.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5tiohhqac3l7fo2zgy0n.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5tiohhqac3l7fo2zgy0n.png" alt="Image-step8-1" width="800" height="447"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Download the model checkpoints.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;`&lt;br&gt;
import os&lt;br&gt;
from PIL import Image&lt;br&gt;
import torch&lt;/p&gt;

&lt;p&gt;from diffusers import QwenImageEditPipeline&lt;/p&gt;

&lt;p&gt;pipeline = QwenImageEditPipeline.from_pretrained("Qwen/Qwen-Image-Edit")&lt;br&gt;
print("pipeline loaded")&lt;br&gt;
pipeline.to(torch.bfloat16)&lt;br&gt;
pipeline.to("cuda")&lt;br&gt;
pipeline.set_progress_bar_config(disable=None)&lt;br&gt;
image = Image.open("./cat.jpg").convert("RGB")&lt;br&gt;
prompt = "Add a pillow under cat's head and cover it with a blanket."&lt;br&gt;
inputs = {&lt;br&gt;
    "image": image,&lt;br&gt;
    "prompt": prompt,&lt;br&gt;
    "generator": torch.manual_seed(0),&lt;br&gt;
    "true_cfg_scale": 4.0,&lt;br&gt;
    "negative_prompt": " ",&lt;br&gt;
    "num_inference_steps": 50,&lt;br&gt;
}&lt;/p&gt;

&lt;p&gt;with torch.inference_mode():&lt;br&gt;
    output = pipeline(**inputs)&lt;br&gt;
    output_image = output.images[0]&lt;br&gt;
    output_image.save("output_image_edit.png")&lt;br&gt;
    print("image saved at", os.path.abspath("output_image_edit.png"))&lt;br&gt;
`&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F021cu0h7kfedcuvtv2pe.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F021cu0h7kfedcuvtv2pe.png" alt="Image-step8-2" width="800" height="160"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Original Image:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv8o8vgqmw74kjunq1i7n.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv8o8vgqmw74kjunq1i7n.png" alt="Image-step8-3" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Edited Image:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk26adm8z6p71vx7tn68a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk26adm8z6p71vx7tn68a.png" alt="Image-step8-4" width="786" height="524"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Qwen-Image-Edit stands out as a next-generation image editing model, seamlessly blending semantic intelligence with appearance precision to enable everything from subtle object adjustments to bold creative transformations, all while offering unmatched text editing capabilities. By running it on Nodeshift Cloud, you gain a frictionless way to harness this power, eliminating complex setup hurdles and ensuring a smooth, scalable environment for experimentation. Together, Qwen-Image-Edit and Nodeshift Cloud make advanced image editing not just possible, but practical and accessible for creators, developers, and enterprises alike.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>qwen</category>
      <category>opensource</category>
      <category>ai</category>
      <category>productivity</category>
    </item>
    <item>
      <title>Get Started with MiniCPM-v4: The Next-Gen Multimodal AI Model by OpenBMB</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 01 Sep 2025 16:24:49 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/get-started-with-minicpm-v4-the-next-gen-multimodal-ai-model-by-openbmb-4occ</link>
      <guid>https://forem.com/nodeshiftcloud/get-started-with-minicpm-v4-the-next-gen-multimodal-ai-model-by-openbmb-4occ</guid>
      <description>&lt;p&gt;Multimodal AI is rapidly evolving, MiniCPM-V 4.0 by OpenBMB emerges as a game-changer, combining cutting-edge visual understanding with unprecedented efficiency. Built on SigLIP2-400M and MiniCPM4-3B, this compact yet powerful model packs 4.1B parameters, but consistently punches above its weight. It not only inherits the strong single-image, multi-image, and video comprehension capabilities of its predecessor (MiniCPM-V 2.6), but also surpasses it with remarkable efficiency. Benchmark results on OpenCompass demonstrate this leap. MiniCPM-V 4.0 achieves a 69.0 average score, outperforming models like GPT-4.1-mini-20250414, MiniCPM-V 2.6 (8.1B), and Qwen2.5-VL-3B-Instruct, proving that smaller can indeed be smarter. What makes it even more exciting is its real-world usability: the model runs seamlessly on end devices, delivering under 2s first-token delay and over 17 tokens/s decoding on iPhone 16 Pro Max, all without heating issues, making on-device multimodal AI finally practical. With easy integration across frameworks like llama.cpp, Ollama, vLLM, SGLang, LLaMA-Factory, and even a native iOS app, MiniCPM-V 4.0 isn’t just another AI model, it’s a versatile, efficient, and deployment-ready multimodal powerhouse.&lt;/p&gt;

&lt;p&gt;In this article, we're going to see a step-by-step process to install and run this model locally or in GPU-accelerated environments.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 1x RTX 4090 or 1x RTX A6000&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 50GB (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 16GB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install and run MiniCPM-v4
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by &lt;a href="https://nodeshift.com" rel="noopener noreferrer"&gt;NodeShift&lt;/a&gt; since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 1x A100 SXM4 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 100GB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" alt="Image-step3-1" width="800" height="227"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 1x RTXA6000 GPU node with 64vCPUs/63GB RAM/200GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" alt="Image-step4-1" width="800" height="394"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" alt="Image-step5-3" width="800" height="389"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zxg4ejw14rh5vwg7z7e.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zxg4ejw14rh5vwg7z7e.png" alt="Image-step6-1" width="800" height="370"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n minicpm python=3.11 -y &amp;amp;&amp;amp; conda activate minicpm
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe1g2ubh3ivrkyj8f4pyo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe1g2ubh3ivrkyj8f4pyo.png" alt="Image-step7-1" width="800" height="485"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Install required dependencies.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121 
pip install einops timm pillow
pip install git+https://github.com/huggingface/transformers
pip install git+https://github.com/huggingface/accelerate
pip install git+https://github.com/huggingface/diffusers
pip install huggingface_hub
pip install sentencepiece bitsandbytes protobuf decord numpy
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Falkb0c52i63ff1t01k0d.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Falkb0c52i63ff1t01k0d.png" alt="Image-step7-2" width="800" height="485"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Install and run jupyter notebook.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda install -c conda-forge --override-channels notebook -y
conda install -c conda-forge --override-channels ipywidgets -y
jupyter notebook --allow-root
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;4) If you're on a remote machine (e.g., NodeShift GPU), you'll need to do SSH port forwarding in order to access the jupyter notebook session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_PORT&amp;gt;&lt;/code&gt; with the PORT allotted to your remote server (For the NodeShift server - you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;ssh -L 8888:localhost:8888 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" alt="Image-step7-3" width="800" height="231"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After this copy the URL you received in your remote server:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" alt="Image-step7-4" width="800" height="267"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And paste this on your local browser to access the Jupyter Notebook session.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Download and Run the model
&lt;/h3&gt;

&lt;p&gt;1) Open a Python notebook inside Jupyter.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" alt="Image-step8-1" width="800" height="447"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Download the model checkpoints.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;from PIL import Image
import torch
from transformers import AutoModel, AutoTokenizer

model_path = 'openbmb/MiniCPM-V-4'
model = AutoModel.from_pretrained(model_path, trust_remote_code=True,
                                  # sdpa or flash_attention_2, no eager
                                  attn_implementation='sdpa', torch_dtype=torch.bfloat16)
model = model.eval().cuda()
tokenizer = AutoTokenizer.from_pretrained(
    model_path, trust_remote_code=True)



image = Image.open('./landform.jpg').convert('RGB')

# First round chat 
question = "What is the landform in the picture?"
msgs = [{'role': 'user', 'content': [image, question]}]

answer = model.chat(
    msgs=msgs,
    image=image,
    tokenizer=tokenizer
)
print(answer)


# Second round chat, pass history context of multi-turn conversation
msgs.append({"role": "assistant", "content": [answer]})
msgs.append({"role": "user", "content": [
            "What should I pay attention to when traveling here?"]})

answer = model.chat(
    msgs=msgs,
    image=None,
    tokenizer=tokenizer
)
print(answer)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb4wbgse6j4bjxhe2vidk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb4wbgse6j4bjxhe2vidk.png" alt="Image-step8-2" width="800" height="379"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Here’s the image we used to testing the model:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjqmmtouc957t3wd0uycp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjqmmtouc957t3wd0uycp.png" alt="Image-step8-3" width="200" height="300"&gt;&lt;/a&gt;&lt;br&gt;
Picsum ID: 866&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpgztxszybzjktbxkxa6a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpgztxszybzjktbxkxa6a.png" alt="Image-step8-4" width="800" height="183"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;To wrap up, MiniCPM-V 4.0 clearly demonstrates how multimodal AI is becoming more efficient, accessible, and deployment-ready, setting a new benchmark in balancing compact design with powerful visual and reasoning capabilities. From its ability to outperform larger models on benchmarks to its seamless real-world usability on devices like the iPhone 16 Pro Max, it proves that high performance no longer requires massive scale. At the same time, Nodeshift Cloud makes experimenting with and deploying such state-of-the-art models far more practical, offering GPU-accelerated environments, simple setup workflows, and flexible scaling to match your needs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>machinelearning</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>How to Install &amp; Run Qwen Image</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 01 Sep 2025 16:10:21 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/how-to-install-run-qwen-image-1pjd</link>
      <guid>https://forem.com/nodeshiftcloud/how-to-install-run-qwen-image-1pjd</guid>
      <description>&lt;p&gt;Imagine transforming a simple text prompt into a high-quality image with just a few lines of code. Qwen-Image makes this possible by combining advanced image generation with precise text rendering, whether you’re working in English or Chinese. It handles everything from photorealistic scenes and impressionist-style paintings to clean, minimalist designs, adapting its output to your needs. On top of that, Qwen-Image offers powerful editing features: you can insert or remove objects, fine-tune colours and details, edit text directly within an image, and even adjust human poses—all through clear, natural-language commands. Behind the scenes, it also performs tasks like object detection, semantic segmentation, depth estimation and super-resolution, giving you a complete toolkit for creating and refining images with ease.&lt;/p&gt;

&lt;p&gt;Getting started is simple. In the next section, you’ll see exactly how to install Qwen-Image and run your first prompt in minutes.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 1x H100&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 50 GB (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 64 GB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install and run Qwen Image
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by &lt;a href="https://nodeshift.com" rel="noopener noreferrer"&gt;NodeShift&lt;/a&gt; since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 1x H200 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 200 GB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2nwqblqu9dtn5vbnnpvm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2nwqblqu9dtn5vbnnpvm.png" alt="Image-step3-1" width="800" height="277"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 1x H100 SXM 80GB GPU node with 192vCPUs/80GB RAM/200GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F33ctuz0kf0n28kilc7zj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F33ctuz0kf0n28kilc7zj.png" alt="Image-step4-1" width="800" height="356"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Filngygf82xfgk2o7lyxv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Filngygf82xfgk2o7lyxv.png" alt="Image-step5-3" width="800" height="397"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqckok7vzis7m6g0pecxw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqckok7vzis7m6g0pecxw.png" alt="Image-step6-1" width="800" height="378"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n qwen-img python=3.11 -y &amp;amp;&amp;amp; conda activate qwen-img
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zy0g2h5aopcr1jenciq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zy0g2h5aopcr1jenciq.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Install required dependencies.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121 
pip install einops timm pillow
pip install git+https://github.com/huggingface/transformers
pip install git+https://github.com/huggingface/accelerate
pip install git+https://github.com/huggingface/diffusers
pip install huggingface_hub
pip install sentencepiece bitsandbytes protobuf decord numpy
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq57nt1dltxdp38396921.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq57nt1dltxdp38396921.png" alt="Image-step7-2" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Install and run jupyter notebook.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda install -c conda-forge --override-channels notebook -y
conda install -c conda-forge --override-channels ipywidgets -y
jupyter notebook --allow-root
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;4) If you’re on a remote machine (e.g., NodeShift GPU), you’ll need to do SSH port forwarding in order to access the jupyter notebook session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;` with the PORT allotted to your remote server (For the NodeShift server – you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;&lt;code&gt;&lt;br&gt;
ssh -L 8888:localhost:8888 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;br&gt;
&lt;/code&gt;&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2r8j1owgltt9aq3dt6yq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2r8j1owgltt9aq3dt6yq.png" alt="Image-step7-3" width="800" height="231"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After this copy the URL you received in your remote server:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmfvxbys411rncwon4x0s.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmfvxbys411rncwon4x0s.png" alt="Image-step7-4" width="800" height="267"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And paste this on your local browser to access the Jupyter Notebook session.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Download and Run the model
&lt;/h3&gt;

&lt;p&gt;1) Open a Python notebook inside Jupyter.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5tiohhqac3l7fo2zgy0n.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5tiohhqac3l7fo2zgy0n.png" alt="Image-step8-1" width="800" height="447"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Download the model checkpoints.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&lt;/code&gt;`&lt;br&gt;
from diffusers import DiffusionPipeline&lt;br&gt;
import torch&lt;/p&gt;

&lt;p&gt;model_name = "Qwen/Qwen-Image"&lt;/p&gt;

&lt;h1&gt;
  
  
  Load the pipeline
&lt;/h1&gt;

&lt;p&gt;if torch.cuda.is_available():&lt;br&gt;
    torch_dtype = torch.bfloat16&lt;br&gt;
    device = "cuda"&lt;br&gt;
else:&lt;br&gt;
    torch_dtype = torch.float32&lt;br&gt;
    device = "cpu"&lt;/p&gt;

&lt;p&gt;pipe = DiffusionPipeline.from_pretrained(model_name, torch_dtype=torch_dtype)&lt;br&gt;
pipe = pipe.to(device)&lt;/p&gt;

&lt;p&gt;positive_magic = {&lt;br&gt;
    "en": "Ultra HD, 4K, cinematic composition." # for english prompt,&lt;br&gt;
    "zh": "超清，4K，电影级构图" # for chinese prompt,&lt;br&gt;
}&lt;/p&gt;

&lt;h1&gt;
  
  
  Generate image
&lt;/h1&gt;

&lt;p&gt;prompt = '''A coffee shop entrance features a chalkboard sign reading "Qwen Coffee 😊 $2 per cup," with a neon light beside it displaying "通义千问". Next to it hangs a poster showing a beautiful Chinese woman, and beneath the poster is written "π≈3.1415926-53589793-23846264-33832795-02384197". Ultra HD, 4K, cinematic composition'''&lt;/p&gt;

&lt;p&gt;negative_prompt = " " # using an empty string if you do not have specific concept to remove&lt;/p&gt;

&lt;h1&gt;
  
  
  Generate with different aspect ratios
&lt;/h1&gt;

&lt;p&gt;aspect_ratios = {&lt;br&gt;
    "1:1": (1328, 1328),&lt;br&gt;
    "16:9": (1664, 928),&lt;br&gt;
    "9:16": (928, 1664),&lt;br&gt;
    "4:3": (1472, 1140),&lt;br&gt;
    "3:4": (1140, 1472),&lt;br&gt;
    "3:2": (1584, 1056),&lt;br&gt;
    "2:3": (1056, 1584),&lt;br&gt;
}&lt;/p&gt;

&lt;p&gt;width, height = aspect_ratios["16:9"]&lt;/p&gt;

&lt;p&gt;image = pipe(&lt;br&gt;
    prompt=prompt + positive_magic["en"],&lt;br&gt;
    negative_prompt=negative_prompt,&lt;br&gt;
    width=width,&lt;br&gt;
    height=height,&lt;br&gt;
    num_inference_steps=50,&lt;br&gt;
    true_cfg_scale=4.0,&lt;br&gt;
    generator=torch.Generator(device="cuda").manual_seed(42)&lt;br&gt;
).images[0]&lt;/p&gt;

&lt;p&gt;image.save("example.png")&lt;br&gt;
`&lt;code&gt;&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F704sfgamozztlieecl2r.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F704sfgamozztlieecl2r.png" alt="Image-step8-2" width="800" height="445"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F71h42kbar10oiyz8qz71.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F71h42kbar10oiyz8qz71.png" alt="Image-step8-3" width="800" height="446"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;You’ve seen how Qwen-Image turns simple text prompts into stunning, high-fidelity images, whether photorealistic, painterly or minimal, and offers intuitive editing for objects, colour, text and even human poses, all backed by robust image-understanding capabilities like segmentation and super-resolution. Equally straightforward is getting up and running: a few commands installs the model via diffusers and, in minutes, you’re generating your first visuals. By pairing Qwen-Image with NodeShift Cloud, you gain instant access to scalable GPU instances, automated deployment of your inference pipeline and managed versioning, so you can focus on creativity while NodeShift ensures performance, reliability and easy integration into your existing workflows.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>qwen</category>
      <category>ai</category>
      <category>genai</category>
      <category>opensource</category>
    </item>
    <item>
      <title>A Step-By-Step Guide to Install Qwen3 30B Locally</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 11 Aug 2025 13:43:54 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/a-step-by-step-guide-to-install-qwen3-30b-locally-o7j</link>
      <guid>https://forem.com/nodeshiftcloud/a-step-by-step-guide-to-install-qwen3-30b-locally-o7j</guid>
      <description>&lt;p&gt;The Qwen3-30B-A3B-Instruct-2507 is an advanced iteration of the Qwen3 series, marking a significant leap forward in the landscape of causal language models. Boasting an impressive 30.5 billion parameters with 3.3 billion actively engaged, this model excels across a diverse array of capabilities such as instruction following, complex logical reasoning, text comprehension, mathematics, and science. Its robust coding proficiency, demonstrated by high scores in benchmarks such as MultiPL-E and LiveCodeBench, makes it particularly attractive to developers and researchers. The model also excels in multilingual contexts and handles extensive 256K token contexts effortlessly, making it ideal for intricate, lengthy tasks. Furthermore, its refined alignment with user preferences in subjective and open-ended scenarios ensures that interactions feel natural, intuitive, and highly personalised.&lt;/p&gt;

&lt;p&gt;In this article, we guide you step-by-step on installing Qwen3-30B locally or in GPU-accelerated environment.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 1x H200&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 50 GB (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 64 GB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install and run Qwen3-30B
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by NodeShift since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 1x H200 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 200 GB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4eg1srvuvt289uey0baa.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4eg1srvuvt289uey0baa.png" alt="Image-step3-1" width="800" height="277"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 1x H100 SXM 80GB GPU node with 192vCPUs/80GB RAM/200GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffuuphp4rmseb1dr3c6qt.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffuuphp4rmseb1dr3c6qt.png" alt="Image-step4-1" width="800" height="356"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk810i78g0piq7z2jxu8j.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk810i78g0piq7z2jxu8j.png" alt="Image-step5-3" width="800" height="397"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F15qhftsr1k75orubyj15.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F15qhftsr1k75orubyj15.png" alt="Image-step6-1" width="800" height="378"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n qwen python=3.11 -y &amp;amp;&amp;amp; conda activate qwen
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5hwf3q01mgtlwhxavyue.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5hwf3q01mgtlwhxavyue.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Once you’re inside the environment, install vllm with dependencies.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install --upgrade vllm
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8l3uska8dmuptv4jyb96.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8l3uska8dmuptv4jyb96.png" alt="Image-step7-2" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Also, open a second terminal, connect to remote server with SSH and install open-webui.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install open-webui
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 8: Download and Run the model
&lt;/h3&gt;

&lt;p&gt;1) Download model with vllm and host the endpoint at 8000.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;vllm serve Qwen/Qwen3-30B-A3B-Instruct-2507 --max-model-len 32768 --gpu-memory-utilization 0.95
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9b7fry6idimqoi1sl2ur.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9b7fry6idimqoi1sl2ur.png" alt="Image-step8-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) In the second terminal connected with the GPU host with ssh, serve the open-webui frontend endpoint.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;open-webui serve --port 3000
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhszbpe7ynff5vyqgx36w.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhszbpe7ynff5vyqgx36w.png" alt="Image-step8-2" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjhk910xjcstcgte8buly.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjhk910xjcstcgte8buly.png" alt="Image-step8-3" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Forward both the ports and tunnel them to access in the local browser.&lt;/p&gt;

&lt;p&gt;If you’re on a remote machine (e.g., NodeShift GPU), you’ll need to do SSH port forwarding in order to access the both vllm and open-webui session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_PORT&amp;gt;&lt;/code&gt; with the PORT allotted to your remote server (For the NodeShift server – you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;ssh -L 3000:localhost:3000 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;In another local terminal run forward the port for vllm endpoint:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;ssh -L 8000:localhost:8000 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 9: Run the model via Open WebUI Interface
&lt;/h3&gt;

&lt;p&gt;Once ports are forwarded, you can simply access the model via Open WebUI interface and chat with it.&lt;/p&gt;

&lt;p&gt;1) Before running the model, connect the webui with vllm API endpoint in the settings.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe2aoj18gjzcu8pvrori0.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe2aoj18gjzcu8pvrori0.png" alt="Image-step9-1" width="800" height="512"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Select the Qwen3-30B model in the chat page and run the prompt.&lt;/p&gt;

&lt;p&gt;For e.g., we’re testing the following prompt:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;1. Summarize the following passage in 3 bullet points.
2. Then, extract 3 key insights and explain their implications.
3. Finally, write a Python function that could analyze similar passages for sentiment.

---
Passage:

"The rapid advancement of AI technologies has transformed industries across the globe. In healthcare, AI models are diagnosing diseases earlier and more accurately. In finance, algorithmic trading and risk modeling are becoming more sophisticated. Yet, as AI grows more powerful, ethical questions around bias, privacy, and job displacement remain urgent. Policymakers and technologists must collaborate to create guardrails that ensure innovation benefits society as a whole."
---

Give your response in clearly separated sections.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frtl4170rm5axn0mcm3ek.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frtl4170rm5axn0mcm3ek.png" alt="Image-step9-2" width="800" height="436"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdxdp9zrtimcx7ageygde.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdxdp9zrtimcx7ageygde.png" alt="Image-step9-3" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fskm5ssf4mesaqw432140.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fskm5ssf4mesaqw432140.png" alt="Image-step9-4" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Installing Qwen3-30B-A3B-Instruct-2507 locally equips developers and researchers with a cutting-edge language model, renowned for its powerful reasoning, extensive multilingual support, and exceptional handling of long-context tasks. Pariring it with NodeShift GPUs further enhances this experience, providing streamlined deployment, efficient resource management, and scalable infrastructure. Together, these tools empower users to harness advanced AI capabilities effectively, bridging innovation with accessibility and performance.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>qwen</category>
      <category>ai</category>
      <category>llm</category>
      <category>opensource</category>
    </item>
    <item>
      <title>How AI is Saving Sales Teams 1,000+ Hours Annually - Securely and at Scale</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 11 Aug 2025 12:20:12 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/how-ai-is-saving-sales-teams-1000-hours-annually-securely-and-at-scale-3anc</link>
      <guid>https://forem.com/nodeshiftcloud/how-ai-is-saving-sales-teams-1000-hours-annually-securely-and-at-scale-3anc</guid>
      <description>&lt;p&gt;Sales teams are under immense pressure. Quarter after quarter, they’re expected to hit ambitious revenue targets, respond faster than ever, and deliver personalized experiences across every touchpoint. However, there’s a hidden roadblock that no one talks about, i.e., time. From drafting repetitive outreach emails, chasing follow-ups, and customizing proposals to filling out RFPs and updating CRM entries, a massive portion of a sales representative’s day is spent in "not selling". In fact, studies show that over 60% of a representative’s time is consumed by non-revenue-generating admin work. That’s not just frustrating, it’s expensive.&lt;/p&gt;

&lt;p&gt;For mid-sized sales teams, this amounts to over 1,000 hours lost per year, often by the highest-paid and most skilled employees. And while traditional sales automation tools offer some relief, they rarely deliver the deep contextual understanding or security that today’s enterprise environments demand.&lt;/p&gt;

&lt;p&gt;That’s where NodeShift’s sovereign AI platform steps in, offering a private, self-hosted solution that seamlessly integrates with your existing CRM tools like HubSpot, Apollo, or Salesforce and more. Powered by state-of-the-art open-source models like LLaMA, DeepSeek, and Mistral, NodeShift delivers intelligent sales copilots that work just like ChatGPT, only fully within your infrastructure.&lt;/p&gt;

&lt;h2&gt;
  
  
  Practical Ways Sales Teams Save Time with AI
&lt;/h2&gt;

&lt;p&gt;Sales teams today are overwhelmed by admin-heavy workflows that slow down pipeline momentum and erase selling time. With NodeShift’s sovereign AI platform, sales orgs can deploy high-performing, private AI agents that embed directly into their tools and workflows, turning weeks of manual work into minutes of intelligent automation.&lt;/p&gt;

&lt;p&gt;Here are five real-world applications where sales teams are regaining control of their time and focus:&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;1) Auto-Generate Personalized Emails &amp;amp; Outreach&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Writing personalized prospecting emails at scale is time-consuming, and using templates often leads to low engagement. NodeShift-hosted models, such as LLaMA 3 or Mistral-7B, can be fine-tuned on your past outreach, ICP definitions, and product positioning. Once integrated with your CRM (e.g., HubSpot, Apollo via MCPs or APIs), the AI can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Auto-draft hyper-personalized cold emails for each lead based on firmographic and behavioral data&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Tailor follow-ups dynamically based on engagement history&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Localize content (e.g., for multilingual markets) using in-house translation support&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Impact:&lt;/strong&gt; Sales reps save 6–10 hours/week on outbound efforts, while improving open and reply rates by 20–35%.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;2) Instantly Generate Proposals, Sales Decks &amp;amp; Security Responses&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Creating customized proposals, tailored pitch decks, and responses to procurement or security questionnaires is a time-consuming process. Reps and solution engineers often spend hours assembling assets from multiple documents, past deals, and stakeholder inputs, often under tight deadlines.&lt;/p&gt;

&lt;p&gt;NodeShift’s sovereign AI platform acts as your always-on sales assistant. By connecting securely to your internal document repository, past proposals, slide decks, pricing sheets, and product documentation, your AI copilot can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Generate fully customized sales proposals and SoWs (Statements of Work) based on customer profile, industry, and deal stage&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Auto-build branded presentation decks with editable slides, based on opportunity data, product fit, and competitor positioning&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Complete security questionnaires and RFPs by retrieving previously approved answers and formatting them to match submission requirements&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Pull insights from your CRM, knowledge base, and compliance docs without ever sending data to any external third-party service or cloud&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The system outputs editable content in formats like PDF, Google Slides, Microsoft PowerPoint, and Word, making review and refinement effortless.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Impact:&lt;/strong&gt; Sales teams reduce proposal cycles from 3–5 days to less than 3 hours, reclaiming dozens of hours per month while improving consistency and response quality.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;3) Call Summarization &amp;amp; Deal Insights&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Valuable insights from discovery calls, demos, and follow-ups often get buried in meeting recordings or never make it into the CRM. &lt;/p&gt;

&lt;p&gt;Once connected to tools like Zoom or Google Meet transcripts, NodeShift’s AI:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Summarizes key takeaways, objections, and action items&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Automatically updates the CRM (Salesforce, HubSpot) with notes&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Suggests next steps or personalized recap emails&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Impact:&lt;/strong&gt; Reps save ~5 hours/week on call logging, while managers gain full visibility into deal health without micromanaging.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;4) Intelligent CRM Hygiene &amp;amp; Lead Prioritization&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;CRMs are often cluttered, outdated, or underutilized, making it hard to prioritize the right deals.&lt;/p&gt;

&lt;p&gt;By integrating NodeShift's AI models with your CRM and third-party data sources (via APIs or the Model Context Protocol), AI agents can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Auto-enrich lead profiles with public data (e.g., LinkedIn, Clearbit, Apollo)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Flag duplicate or stale records&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Score leads based on buying intent and account fit&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Recommend the next best action (call, email, nurture)&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Impact:&lt;/strong&gt; Sales operations save 2–4 hours/week, while reps spend less time guessing and more time engaging.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;5) Conversational Team Onboarding &amp;amp; Enablement&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Onboarding new sales reps can take 1–2 months of shadowing, reading, and repetitive Q&amp;amp;A.&lt;/p&gt;

&lt;p&gt;For this, you can deploy a custom AI copilot on NodeShift, which is trained on:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Product FAQs and battlecards&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Objection handling scripts&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Past deals and CRM notes&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Competitive positioning documents&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;New hires can interact with the AI like a live trainer:&lt;/p&gt;

&lt;p&gt;“How do we position our product against [Competitor] for [Healthcare CTOs]?”&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Impact:&lt;/strong&gt; Onboarding time drops by 30–50%, and reps are deal-ready weeks faster.&lt;/p&gt;

&lt;h2&gt;
  
  
  NodeShift's Sovereign AI - Designed for Security and Scale
&lt;/h2&gt;

&lt;p&gt;NodeShift offers a fundamentally different approach, a sovereign AI platform purpose-built for privacy-conscious organizations that want the power of open-source large language models (LLMs) without sacrificing data ownership or operational control.&lt;/p&gt;

&lt;p&gt;By securely deploying models like Mistral, LLaMA 3, DeepSeek, and others directly within your infrastructure, and integrating seamlessly with CRM systems like HubSpot, Apollo, Salesforce, and internal knowledge bases,  NodeShift enables organizations to build intelligent sales copilots that are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Deeply contextualized with your product, customer, and market knowledge&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Fully under your control (on-prem, VPC, or private sovereign cloud)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Capable of automating hours of manual tasks across the entire sales funnel&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If your team is spending more time documenting deals than closing them, it’s time to rethink what AI can do for your sales operations.&lt;/p&gt;

&lt;h2&gt;
  
  
  Ready to See NodeShift AI in Action?
&lt;/h2&gt;

&lt;p&gt;Sales performance is no longer just about hiring more reps; it’s about removing the friction that slows them down.&lt;/p&gt;

&lt;p&gt;With NodeShift’s sovereign AI platform, your organization can harness the power of cutting-edge open-source models like Mistral, LLaMA, and DeepSeek, securely deployed inside your infrastructure, connected to your CRM, and tailored to your workflows.&lt;/p&gt;

&lt;p&gt;Now that you’ve seen how sales teams can reclaim 1,000+ hours annually, it’s time to explore what that transformation looks like for your team. Our team is ready to help you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Assess integration points and automation gaps&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Deploy the right open-source AI model in your own infrastructure&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Launch in under a week with full ownership and compliance&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>automation</category>
      <category>startup</category>
      <category>productivity</category>
    </item>
    <item>
      <title>How to Install &amp; Run Qwen3-Thinking</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 04 Aug 2025 16:48:42 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/how-to-install-run-qwen3-thinking-42gi</link>
      <guid>https://forem.com/nodeshiftcloud/how-to-install-run-qwen3-thinking-42gi</guid>
      <description>&lt;p&gt;In the world of open-source AI, very few models come close to rivaling the intellectual firepower of proprietary giants, until now. Introducing Qwen3-235B-A22B-Thinking-2507, a frontier model in the realm of thinking-capable language models. Engineered by Alibaba Cloud, this 235B-parameter behemoth, 22B of which are actively used per inference, excels in high-level reasoning, mathematical problem solving, scientific logic, and advanced coding tasks. With its unprecedented 256K context length, this model is built not just for chat, but for deep, extended reasoning across massive documents and chains of logic. From dominating benchmarks like AIME25 and HMMT25 to outperforming Claude Opus in reasoning-heavy scenarios, Qwen Thinking isn’t just another LLM, it’s a state-of-the-art brain built for intellectual rigor. And the best part? You can now run it locally.&lt;/p&gt;

&lt;p&gt;Let's walk through how to install and harness this thinking model right from your own machine.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 8x H200&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 1 TB (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 1 TB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install and run Qwen Thinking
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by NodeShift since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 8x H200 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 1 TB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgly6qkzpqv51516doe07.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgly6qkzpqv51516doe07.png" alt="Image-step2-3" width="800" height="272"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 8x H200 140 GB GPU node with 384vCPUs/2TB RAM/200GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi69b5gfnkjz0h73sylw6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi69b5gfnkjz0h73sylw6.png" alt="Image-step4-1" width="800" height="382"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fed4vdnyb08thflcpa2nc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fed4vdnyb08thflcpa2nc.png" alt="Image-step5-3" width="800" height="388"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm63iskjch3q6digrgaik.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm63iskjch3q6digrgaik.png" alt="Image-step6-1" width="800" height="306"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n qwen-thinking python=3.11 -y &amp;amp;&amp;amp; conda activate qwen-thinking
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fh5di06b2nzhrps7yayxz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fh5di06b2nzhrps7yayxz.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Once you're inside the environment, install necessary dependencies to run the model.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121
pip install --upgrade transformers accelerate einops
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvn22juw19oekhs965trk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvn22juw19oekhs965trk.png" alt="Image-step7-2" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Install PyTorch, transformers and other python packages.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install torch torchvision torchaudio 
pip install einops timm pillow
pip install transformers==4.47.0 git+https://github.com/huggingface/accelerate
pip install git+https://github.com/huggingface/diffusers
pip install huggingface_hub
pip install sentencepiece bitsandbytes protobuf decord numpy ffmpeg
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;4) Install and run jupyter notebook.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda install -c conda-forge --override-channels notebook -y
conda install -c conda-forge --override-channels ipywidgets -y
jupyter notebook --allow-root
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;5) If you're on a remote machine (e.g., NodeShift GPU), you'll need to do SSH port forwarding in order to access the jupyter notebook session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_PORT&amp;gt;&lt;/code&gt; with the PORT allotted to your remote server (For the NodeShift server - you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;ssh -L 8888:localhost:8888 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" alt="Image-step7-3" width="800" height="231"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After this copy the URL you received in your remote server:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" alt="Image-step7-4" width="800" height="267"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And paste this on your local browser to access the Jupyter Notebook session.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Download and Run the model
&lt;/h3&gt;

&lt;p&gt;1) Open a Python notebook inside Jupyter.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" alt="Image-step8-1" width="800" height="447"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Download the model checkpoints.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "Qwen/Qwen3-235B-A22B-Thinking-2507"

# load the tokenizer and the model
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)

# prepare the model input
prompt = "Give me a short introduction to large language model."
messages = [
    {"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True,
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

# conduct text completion
generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=32768
)
output_ids = generated_ids[0][len(model_inputs.input_ids[0]):].tolist() 

# parsing thinking content
try:
    # rindex finding 151668 (&amp;lt;/think&amp;gt;)
    index = len(output_ids) - output_ids[::-1].index(151668)
except ValueError:
    index = 0

thinking_content = tokenizer.decode(output_ids[:index], skip_special_tokens=True).strip("\n")
content = tokenizer.decode(output_ids[index:], skip_special_tokens=True).strip("\n")

print("thinking content:", thinking_content)
print("content:", content)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2qdr0t8ls70s4vx3z6aw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2qdr0t8ls70s4vx3z6aw.png" alt="Image-step8-3" width="800" height="442"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9tekpwxcn52nw11ncs2n.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9tekpwxcn52nw11ncs2n.png" alt="Image-step8-4" width="800" height="442"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flcff8bxz74q03gbwyau8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flcff8bxz74q03gbwyau8.png" alt="Image-step8-5" width="800" height="289"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmh5q7yhastuw4rftidxz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmh5q7yhastuw4rftidxz.png" alt="Image-step8-6" width="800" height="247"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Installing Qwen3-235B-A22B-Thinking-2507 locally isn’t just a technical feat; it’s a gateway to unlocking one of the most advanced open-source reasoning models available today. In this guide, we explored what makes this model a powerhouse for logical reasoning, coding, and long-context understanding, and how its “thinking mode” elevates it far beyond conventional LLMs. NodeShift Cloud played a pivotal role by simplifying the deployment process, offering the compute muscle and flexible infrastructure needed to run such a massive model seamlessly. Whether you’re experimenting, building, or benchmarking, NodeShift makes it easier than ever to bring cutting-edge AI capabilities right to your fingertips.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>qwen</category>
      <category>ai</category>
      <category>coding</category>
      <category>opensource</category>
    </item>
    <item>
      <title>How to Install &amp; Run Higgs Audio v2 Locally</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 04 Aug 2025 16:12:38 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/how-to-install-run-higgs-audio-v2-locally-3ppg</link>
      <guid>https://forem.com/nodeshiftcloud/how-to-install-run-higgs-audio-v2-locally-3ppg</guid>
      <description>&lt;p&gt;Imagine an audio generation model so expressive, it can narrate a story with human-like cadence, speak in your cloned voice with melodic voice, or conduct a natural conversation between two completely different speakers, all in multiple languages and without any fine-tuning. That’s exactly what Higgs Audio v2 delivers. Pretrained on over 10 million hours of meticulously annotated audio and text, this powerful open-source audio foundation model pushes the boundaries of what's possible in text-to-speech (TTS) and audio synthesis. Built on top of Llama 3.2-3B and enhanced with a novel DualFFN audio adapter architecture, Higgs Audio v2 combines the deep language understanding of LLMs with a cutting-edge discretized audio tokenizer capable of capturing both semantic and acoustic detail at just 25 fps. It excels in zero-shot prosody adaptation, multilingual translation, multi-speaker dialogue generation, and even simultaneous background music and speech synthesis. With state-of-the-art results on Seed-TTS Eval, ESD, and EmergentTTS-Eval, and win rates of up to 75.7% over GPT-4o-mini-TTS in emotion-rich generations, this model is not just a technical marvel, it's an invitation to explore the future of voice AI.&lt;/p&gt;

&lt;p&gt;If you're ready to build with the next generation of audio intelligence, this guide will walk you through installing Higgs Audio v2 locally, unlocking everything from high-fidelity narration to real-time multilingual voice cloning, right on your machine.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 1x RTX 4090 or 1x RTX A6000&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 50GB (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 16GB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install and run Higgs Audio v2
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by NodeShift since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 1x A100 SXM4 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 100GB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" alt="Image-step3-1" width="800" height="227"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 2x A100 80GB GPU node with 32vCPUs/131GB RAM/100GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" alt="Image-step4-1" width="800" height="394"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" alt="Image-step5-3" width="800" height="389"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zxg4ejw14rh5vwg7z7e.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zxg4ejw14rh5vwg7z7e.png" alt="Image-step6-1" width="800" height="370"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n higgs python=3.11 -y &amp;amp;&amp;amp; conda activate higgs
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw8hbhqft6yydyippha53.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw8hbhqft6yydyippha53.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Once you’re inside the environment, clone the official repository.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;git clone https://github.com/boson-ai/higgs-audio.git
cd higgs-audio
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl3mv27z3tbc2phnvcq8v.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl3mv27z3tbc2phnvcq8v.png" alt="Image-step7-2" width="800" height="182"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Install required dependencies.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install -r requirements.txt
pip install -e .
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;4) Install PyTorch, transformers and other python packages.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install torch torchvision torchaudio 
pip install einops timm pillow
pip install transformers==4.47.0 git+https://github.com/huggingface/accelerate
pip install git+https://github.com/huggingface/diffusers
pip install huggingface_hub
pip install sentencepiece bitsandbytes protobuf decord numpy ffmpeg
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;5) Install and run jupyter notebook.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda install -c conda-forge --override-channels notebook -y
conda install -c conda-forge --override-channels ipywidgets -y
jupyter notebook --allow-root
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;6) If you're on a remote machine (e.g., NodeShift GPU), you'll need to do SSH port forwarding in order to access the jupyter notebook session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_PORT&amp;gt;&lt;/code&gt; with the PORT allotted to your remote server (For the NodeShift server - you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;ssh -L 8888:localhost:8888 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" alt="Image-step7-3" width="800" height="231"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After this copy the URL you received in your remote server:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" alt="Image-step7-4" width="800" height="267"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And paste this on your local browser to access the Jupyter Notebook session.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Download and Run the model
&lt;/h3&gt;

&lt;p&gt;1) Open a Python notebook inside Jupyter.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" alt="Image-step8-1" width="800" height="447"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Download model checkpoints and run the model for inference.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;from boson_multimodal.serve.serve_engine import HiggsAudioServeEngine, HiggsAudioResponse
from boson_multimodal.data_types import ChatMLSample, Message, AudioContent

import torch
import torchaudio
import time
import click

MODEL_PATH = "bosonai/higgs-audio-v2-generation-3B-base"
AUDIO_TOKENIZER_PATH = "bosonai/higgs-audio-v2-tokenizer"

system_prompt = (
    "Generate audio following instruction.\n\n&amp;lt;|scene_desc_start|&amp;gt;\nAudio is recorded from a quiet room.\n&amp;lt;|scene_desc_end|&amp;gt;"
)

messages = [
    Message(
        role="system",
        content=system_prompt,
    ),
    Message(
        role="user",
        content="The sun rises in the east and sets in the west. This simple fact has been observed by humans for thousands of years.",
    ),
]
device = "cuda" if torch.cuda.is_available() else "cpu"

serve_engine = HiggsAudioServeEngine(MODEL_PATH, AUDIO_TOKENIZER_PATH, device=device)

output: HiggsAudioResponse = serve_engine.generate(
    chat_ml_sample=ChatMLSample(messages=messages),
    max_new_tokens=1024,
    temperature=0.3,
    top_p=0.95,
    top_k=50,
    stop_strings=["&amp;lt;|end_of_text|&amp;gt;", "&amp;lt;|eot_id|&amp;gt;"],
)
torchaudio.save(f"output.wav", torch.from_numpy(output.audio)[None, :], output.sampling_rate)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmv11819gpgs796sha0ny.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmv11819gpgs796sha0ny.png" alt="Image-step8-2" width="800" height="444"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Higgs Audio v2 showcases the cutting edge of expressive audio generation, from zero-shot multilingual TTS to realistic multi-speaker dialogues, all powered by innovations like DualFFN architecture, a unified audio tokenizer, and training on 10 million hours of diverse audio. Installing it locally opens the door to these advanced capabilities for developers, researchers, and creatives alike. Powered by NodeShift Cloud, the deployment process becomes even more seamless, offering scalable compute, fast storage, and integrated tooling that accelerates experimentation and production workflows.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>opensource</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>Transform Clinical Research with Microsoft's MediPhi-Instruct</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 04 Aug 2025 15:32:16 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/transform-clinical-research-with-microsofts-mediphi-instruct-1ma</link>
      <guid>https://forem.com/nodeshiftcloud/transform-clinical-research-with-microsofts-mediphi-instruct-1ma</guid>
      <description>&lt;p&gt;In an era where medical language understanding is fast becoming indispensable, Microsoft’s MediPhi-Instruct stands out as a game-changing clinical AI model that combines precision, efficiency, and modular design. Built on the Phi-3.5-mini-instruct foundation, MediPhi isn’t just one model, it’s a symphony of seven finely tuned experts, each trained on distinct medical corpora such as PubMed, medical guidelines, and clinical documents. These specialists are smartly fused using SLERP and BreadCrumbs techniques to retain both depth and generality, culminating in the final MediPhi-Instruct model, aligned using Microsoft’s massive MediFlow dataset. The result? A compact, 3.8B-parameter powerhouse optimized for clinical NLP tasks, from parsing medical codes and literature to assisting in healthcare research, all while running efficiently in memory-constrained or low-latency environments. With its clinical alignment, and modular architecture, MediPhi is not just a research tool, it’s an invitation to revolutionize how we interact with medical language models.&lt;/p&gt;

&lt;p&gt;If you're a medical researcher, NLP engineer, or healthcare innovator, MediPhi-Instruct is remarkably easy to install and tailor to your own clinical use cases, let’s dive in and get it running.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 1x RTX 4090 or 1x RTX A6000&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 50GB (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 16GB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install and run Gemma 3n
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by NodeShift since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 1x A100 SXM4 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 100GB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" alt="Image-step3-1" width="800" height="227"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 2x A100 80GB GPU node with 32vCPUs/131GB RAM/100GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" alt="Image-step4-1" width="800" height="394"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" alt="Image-step5-3" width="800" height="389"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zxg4ejw14rh5vwg7z7e.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zxg4ejw14rh5vwg7z7e.png" alt="Image-step6-1" width="800" height="370"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n med python=3.11 -y &amp;amp;&amp;amp; conda activate med
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs2xp66y4v4j0a8rfaxrk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs2xp66y4v4j0a8rfaxrk.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Once you're inside the environment, install necessary dependencies to run the model.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121
pip install einops timm pillow
pip install git+https://github.com/huggingface/transformers
pip install git+https://github.com/huggingface/accelerate
pip install huggingface_hub
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0wqfibm9bty9zblvcjx7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0wqfibm9bty9zblvcjx7.png" alt=" " width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Install and run jupyter notebook.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda install -c conda-forge --override-channels notebook -y
conda install -c conda-forge --override-channels ipywidgets -y
jupyter notebook --allow-root
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;4) If you're on a remote machine (e.g., NodeShift GPU), you'll need to do SSH port forwarding in order to access the jupyter notebook session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_PORT&amp;gt;&lt;/code&gt; with the PORT allotted to your remote server (For the NodeShift server - you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;ssh -L 8888:localhost:8888 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" alt="Image-step7-3" width="800" height="231"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After this copy the URL you received in your remote server:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" alt="Image-step7-4" width="800" height="267"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And paste this on your local browser to access the Jupyter Notebook session.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Download and Run the model
&lt;/h3&gt;

&lt;p&gt;1) Open a Python notebook inside Jupyter.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" alt="Image-step8-1" width="800" height="447"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Download model checkpoints and run the model for inference.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline

torch.random.manual_seed(0)

model_name = "microsoft/MediPhi-Instruct"
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    device_map="cuda",
    torch_dtype="auto",
    trust_remote_code=True,
)
tokenizer = AutoTokenizer.from_pretrained(model_name)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0gehjnbhmuz74pyrynwv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0gehjnbhmuz74pyrynwv.png" alt="Image-step8-2" width="800" height="443"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Run the model for your inference.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;prompt = "Operative Report:\nPerformed: Cholecystectomy\nOperative Findings: The gallbladder contained multiple stones and had thickening of its wall. Mild peritoneal fluid was noted."

messages = [
    {"role": "system", "content": "Extract medical keywords from this operative notes focus on anatomical, pathological, or procedural vocabulary."},
    {"role": "user", "content": prompt},
]

pipe = pipeline(
    "text-generation",
    model=model,
    tokenizer=tokenizer,
)

generation_args = {
    "max_new_tokens": 500,
    "return_full_text": False,
    "temperature": 0.0,
    "do_sample": False,
}

output = pipe(messages, **generation_args)
print(output[0]['generated_text'])
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdteg6ztwcmzgvnzehinv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdteg6ztwcmzgvnzehinv.png" alt="Image-step8-3" width="800" height="306"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;MediPhi-Instruct brings together the power of modular fine-tuning, expert model merging, and clinical alignment to deliver a lightweight yet robust solution for medical language processing. From decoding complex medical literature to enabling real-time clinical assistance in low-resource settings, it represents a significant leap in accessible and efficient healthcare AI. Powered by Microsoft’s deep domain expertise, it’s designed not only to be technically advanced but also practical for real-world use. And with NodeShift, deploying and experimenting with MediPhi becomes even easier, offering a seamless, GPU-backed environment where researchers can skip the infrastructure hassles and focus on what matters most: unlocking insights and driving innovation in medical AI.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>medical</category>
      <category>microsoft</category>
      <category>ai</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Combine the Power of ASR &amp; LLM with NVIDIA's Canary-Qwen-2.5B</title>
      <dc:creator>Aditi Bindal</dc:creator>
      <pubDate>Mon, 04 Aug 2025 14:37:36 +0000</pubDate>
      <link>https://forem.com/nodeshiftcloud/combine-the-power-of-asr-llm-with-nvidias-canary-qwen-25b-24bb</link>
      <guid>https://forem.com/nodeshiftcloud/combine-the-power-of-asr-llm-with-nvidias-canary-qwen-25b-24bb</guid>
      <description>&lt;p&gt;If you’ve been looking for a way to bring powerful, reliable speech recognition to your local environment, without relying on external APIs, NVIDIA’s new Canary-Qwen-2.5B might be exactly what you need. With 2.5 billion parameters under the hood, this model doesn’t just transcribe English speech with near state-of-the-art accuracy, it does so with punctuation, capitalization, and ultra fast speed (418 RTFx). Canary-Qwen stands out with its two-in-one nature: in ASR mode, it delivers high-quality transcriptions; in LLM mode, it can go further, summarizing, answering questions, or post-processing transcripts using full language understanding. It's fast, it's flexible, and it’s designed for real-world, production-grade use.&lt;/p&gt;

&lt;p&gt;In this article, we’ll walk you through how to get Canary-Qwen-2.5B up and running locally or in GPU-acclerated environments in minutes.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;The minimum system requirements for running this model are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;GPU: 1x RTX 4090 or 1x RTX A6000&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storage: 50GB (preferable)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;VRAM: at least 16GB&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda installed&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Step-by-step process to install and run Canary-Qwen-2.5B
&lt;/h2&gt;

&lt;p&gt;For the purpose of this tutorial, we’ll use a GPU-powered Virtual Machine by NodeShift since it provides high compute Virtual Machines at a very affordable cost on a scale that meets GDPR, SOC2, and ISO27001 requirements. Also, it offers an intuitive and user-friendly interface, making it easier for beginners to get started with Cloud deployments. However, feel free to use any cloud provider of your choice and follow the same steps for the rest of the tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Setting up a NodeShift Account
&lt;/h3&gt;

&lt;p&gt;Visit &lt;a href="https://app.nodeshift.com/sign-up" rel="noopener noreferrer"&gt;app.nodeshift.com&lt;/a&gt; and create an account by filling in basic details, or continue signing up with your Google/GitHub account.&lt;/p&gt;

&lt;p&gt;If you already have an account, &lt;a href="http://app.nodeshift.com" rel="noopener noreferrer"&gt;login&lt;/a&gt; straight to your dashboard.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu3p61u5r46mrb6vcsiqr.png" alt="Image-step1-1" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Create a GPU Node
&lt;/h3&gt;

&lt;p&gt;After accessing your account, you should see a dashboard (see image), now:&lt;/p&gt;

&lt;p&gt;1) Navigate to the menu on the left side.&lt;/p&gt;

&lt;p&gt;2) Click on the &lt;strong&gt;GPU Nodes&lt;/strong&gt; option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fokdraa5tkg40fzgkn7fo.png" alt="Image-step2-1" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Click on &lt;strong&gt;Start&lt;/strong&gt; to start creating your very first GPU node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyfhk9s2i1dfe211zgfev.png" alt="Image-step2-2" width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These GPU nodes are GPU-powered virtual machines by NodeShift. These nodes are highly customizable and let you control different environmental configurations for GPUs ranging from H100s to A100s, CPUs, RAM, and storage, according to your needs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Selecting configuration for GPU (model, region, storage)
&lt;/h3&gt;

&lt;p&gt;1) For this tutorial, we’ll be using 1x A100 SXM4 GPU, however, you can choose any GPU as per the prerequisites.&lt;/p&gt;

&lt;p&gt;2) Similarly, we’ll opt for 100GB storage by sliding the bar. You can also select the region where you want your GPU to reside from the available ones.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6dbmr98w4nv70agy6bq7.png" alt="Image-step3-1" width="800" height="227"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Choose GPU Configuration and Authentication method
&lt;/h3&gt;

&lt;p&gt;1) After selecting your required configuration options, you’ll see the available GPU nodes in your region and according to (or very close to) your configuration. In our case, we’ll choose a 2x A100 80GB GPU node with 32vCPUs/131GB RAM/100GB SSD.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqs3xs9imbvtkd43jv002.png" alt="Image-step4-1" width="800" height="394"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Next, you'll need to select an authentication method. Two methods are available: Password and SSH Key. We recommend using SSH keys, as they are a more secure option. To create one, head over to our &lt;a href="https://docs.nodeshift.com/gpus/create-gpu-deployment" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fchyrp5ijzlmevkc7puaf.png" alt="Image-step4-2" width="800" height="278"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Choose an Image
&lt;/h3&gt;

&lt;p&gt;The final step is to choose an image for the VM, which in our case is &lt;strong&gt;Nvidia Cuda&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnm3gwe0tprkoeqnx5x51.png" alt="Image-step5-1" width="800" height="282"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That's it! You are now ready to deploy the node. Finalize the configuration summary, and if it looks good, click &lt;strong&gt;Create&lt;/strong&gt; to deploy the node.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F647pyrcdxwtp6gz0tieb.png" alt="Image-step5-2" width="800" height="107"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpbg0gd8w65m7lkncav5p.png" alt="Image-step5-3" width="800" height="389"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Connect to active Compute Node using SSH
&lt;/h3&gt;

&lt;p&gt;1) As soon as you create the node, it will be deployed in a few seconds or a minute. Once deployed, you will see a status &lt;strong&gt;Running&lt;/strong&gt; in green, meaning that our Compute node is ready to use!&lt;/p&gt;

&lt;p&gt;2) Once your GPU shows this status, navigate to the three dots on the right, click on &lt;strong&gt;Connect with SSH&lt;/strong&gt;, and copy the SSH details that appear.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zxg4ejw14rh5vwg7z7e.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8zxg4ejw14rh5vwg7z7e.png" alt="Image-step6-1" width="800" height="370"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As you copy the details, follow the below steps to connect to the running GPU VM via SSH:&lt;/p&gt;

&lt;p&gt;1) Open your terminal, paste the SSH command, and run it.&lt;/p&gt;

&lt;p&gt;2) In some cases, your terminal may take your consent before connecting. Enter ‘yes’.&lt;/p&gt;

&lt;p&gt;3) A prompt will request a password. Type the SSH password, and you should be connected.&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7307nybljxnshe9dm4p2.png" alt="Image-step6-2" width="800" height="311"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, If you want to check the GPU details, run the following command in the terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;!nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Step 7: Set up the project environment with dependencies
&lt;/h3&gt;

&lt;p&gt;1) Create a virtual environment using &lt;a href="https://nodeshift.com/blog/set-up-anaconda-on-ubuntu-22-04-in-minutes-simplify-your-ai-workflow" rel="noopener noreferrer"&gt;Anaconda&lt;/a&gt;.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda create -n canary python=3.11 -y &amp;amp;&amp;amp; conda activate canary
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe1sgx53f9r79vgt0uud3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe1sgx53f9r79vgt0uud3.png" alt="Image-step7-1" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Once you're inside the environment, install necessary dependencies to run the model.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install torch&amp;gt;=2.6.0 torchvision torchaudio 
pip install einops timm pillow
pip install git+https://github.com/huggingface/transformers
pip install git+https://github.com/huggingface/accelerate
pip install git+https://github.com/huggingface/diffusers
pip install huggingface_hub
pip install sentencepiece bitsandbytes protobuf decord numpy
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5jse1jf4m6npm21vmbq2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5jse1jf4m6npm21vmbq2.png" alt="Image-step7-2" width="800" height="499"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3) Install NVIDIA NeMo.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;python -m pip install "nemo_toolkit[asr,tts] @ git+https://github.com/NVIDIA/NeMo.git"
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;4) Install some other required dependencies for audio processing.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;sudo apt-get update
sudo apt-get install ffmpeg
pip install sacrebleu
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;5) Install and run jupyter notebook.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;conda install -c conda-forge --override-channels notebook -y
conda install -c conda-forge --override-channels ipywidgets -y
jupyter notebook --allow-root
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;6) If you're on a remote machine (e.g., NodeShift GPU), you'll need to do SSH port forwarding in order to access the jupyter notebook session on your local browser.&lt;/p&gt;

&lt;p&gt;Run the following command in your local terminal after replacing:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_PORT&amp;gt;&lt;/code&gt; with the PORT allotted to your remote server (For the NodeShift server - you can find it in the deployed GPU details on the dashboard).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;PATH_TO_SSH_KEY&amp;gt;&lt;/code&gt; with the path to the location where your SSH key is stored.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;&amp;lt;YOUR_SERVER_IP&amp;gt;&lt;/code&gt; with the IP address of your remote server.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;ssh -L 8888:localhost:8888 -p &amp;lt;YOUR_SERVER_PORT&amp;gt; -i &amp;lt;PATH_TO_SSH_KEY&amp;gt; root@&amp;lt;YOUR_SERVER_IP&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1vq35fq4cbfchpj5wc90.png" alt="Image-step7-3" width="800" height="231"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After this copy the URL you received in your remote server:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz3eexinsdfsvredbkcd.png" alt="Image-step7-4" width="800" height="267"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And paste this on your local browser to access the Jupyter Notebook session.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 8: Download and Run the model
&lt;/h3&gt;

&lt;p&gt;1) Open a Python notebook inside Jupyter.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3jb500tp8lgtfj4pmh1m.png" alt="Image-step8-1" width="800" height="447"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2) Download the model checkpoints, load it to GPU and configure prompt and load audio file.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;import torch
import torchaudio
from nemo.collections.speechlm2.models import SALM

# Load the model
model = SALM.from_pretrained('nvidia/canary-qwen-2.5b')

# Load and preprocess the audio
waveform, sample_rate = torchaudio.load("speech.wav")

expected_sr = model.perception.preprocessor.featurizer.sample_rate

if sample_rate != expected_sr:
    resampler = torchaudio.transforms.Resample(orig_freq=sample_rate, new_freq=expected_sr)
    waveform = resampler(waveform)

if waveform.dim() == 2:
    waveform = waveform.mean(dim=0)

waveform = waveform.unsqueeze(0)
audio_lens = torch.tensor([waveform.shape[1]])

prompt = [[{"role": "user", "content": f"Transcribe the following: {model.audio_locator_tag}"}]]

answer_ids = model.generate(
    prompts=prompt,
    audios=waveform,
    audio_lens=audio_lens,
    max_new_tokens=128,
)

# Decode result
text = model.tokenizer.ids_to_text(answer_ids[0].tolist())
print(text)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Here’s the output for the given audio file:&lt;/p&gt;

&lt;p&gt;Audio:&lt;br&gt;
&lt;a href="https://drive.google.com/file/d/1u9583BT8pvQB_pmxHhrDlNDfVi8BTG4_/view?usp=sharing" rel="noopener noreferrer"&gt;https://drive.google.com/file/d/1u9583BT8pvQB_pmxHhrDlNDfVi8BTG4_/view?usp=sharing&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuj6db2f7ijg2650rbvix.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuj6db2f7ijg2650rbvix.png" alt="Image-step8-2" width="800" height="291"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs91wy4o6j35wht39mva0.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs91wy4o6j35wht39mva0.png" alt="Image-step8-3" width="800" height="280"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;We’ve covered everything from installing dependencies and loading the 2.5 B‑parameter Canary‑Qwen model locally or on GPU‑accelerated infrastructure, to managing audio preprocessing, transcription, and optional LLM post‑processing. What makes this setup rock-solid is pairing Canary‑Qwen with NodeShift AI infra. NodeShift offers affordable, scalable on‑demand GPU instances across global regions, automated deployment via Terraform or GitHub Actions, and enterprise‑grade compliance, so you can spin up an A100 or H100‑backed VM in minutes, run your transcription workflows and other AI workflows, and scale them securely and cost‑effectively.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For more information about NodeShift:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.nodeshift.com/?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Docs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.linkedin.com/company/nodeshift/?%0Aref=blog.nodeshift.com" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://x.com/nodeshiftai?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;X&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://discord.gg/4dHNxnW7p7?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;Discord&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://app.daily.dev/nodeshift?ref=blog.nodeshift.com" rel="noopener noreferrer"&gt;daily.dev&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>tutorial</category>
      <category>opensource</category>
    </item>
  </channel>
</rss>
