<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Abhinav Srivastav</title>
    <description>The latest articles on Forem by Abhinav Srivastav (@abhinav_srivastav_5a243fd).</description>
    <link>https://forem.com/abhinav_srivastav_5a243fd</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3678495%2F63f66254-8a07-42c1-bd46-f0855b93a105.jpg</url>
      <title>Forem: Abhinav Srivastav</title>
      <link>https://forem.com/abhinav_srivastav_5a243fd</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/abhinav_srivastav_5a243fd"/>
    <language>en</language>
    <item>
      <title>What Actually Slows Down PyTorch Training? I Surveyed ML Engineers</title>
      <dc:creator>Abhinav Srivastav</dc:creator>
      <pubDate>Thu, 25 Dec 2025 15:50:22 +0000</pubDate>
      <link>https://forem.com/abhinav_srivastav_5a243fd/what-actually-slows-down-pytorch-training-i-surveyed-ml-engineers-2m8j</link>
      <guid>https://forem.com/abhinav_srivastav_5a243fd/what-actually-slows-down-pytorch-training-i-surveyed-ml-engineers-2m8j</guid>
      <description>&lt;p&gt;I surveyed ML engineers about their training bottlenecks. The results were eye-opening.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Setup
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;69% work with NLP/Transformers&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;84% have training runs lasting 1+ hours&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;Most train on local GPUs (77%) or multi-GPU setups (46%)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These aren't quick experiments; performance matters.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Problems
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Top 3 pain points:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;GPU Out-of-Memory: 62%&lt;/strong&gt; - The nightmare scenario&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Slow dataloader: 39%&lt;/strong&gt; - Classic CPU bottleneck
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Low GPU utilization: 31%&lt;/strong&gt; - Expensive GPU sitting idle&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;77% of engineers hit OOM errors at least occasionally.&lt;/strong&gt; This isn't rare, it's a regular frustration.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Issue
&lt;/h2&gt;

&lt;p&gt;When I asked what is slowing down their training:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;46% said "I don't know"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Let that sink in. Nearly half of ML engineers can't identify their bottlenecks.&lt;/p&gt;

&lt;p&gt;The other 31% pointed to forward pass, dataloader, or batch size issues. Only 8% had it figured out.&lt;/p&gt;

&lt;h2&gt;
  
  
  Current Tools Aren't Enough
&lt;/h2&gt;

&lt;p&gt;What people use today:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;PyTorch Profiler (55%)&lt;/li&gt;
&lt;li&gt;TensorBoard (45%)&lt;/li&gt;
&lt;li&gt;Custom print statements (36%)&lt;/li&gt;
&lt;li&gt;WandB (27%)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The problem? These tools fall into two camps:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Heavy profilers&lt;/strong&gt; (PyTorch Profiler): Great detail, but 10-50% overhead&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Aggregate monitoring&lt;/strong&gt; (TensorBoard, WandB): Shows overall metrics, not layer-level bottlenecks&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What Engineers Actually Want
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Most requested features:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Slowest layers and timings: 58%&lt;/li&gt;
&lt;li&gt;Per-layer memory usage: 50%
&lt;/li&gt;
&lt;li&gt;CPU/GPU utilization: 50%&lt;/li&gt;
&lt;li&gt;Dataloader breakdown: 42%&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The pattern is clear: &lt;strong&gt;people want layer-level visibility without killing performance.&lt;/strong&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why I Built TraceML
&lt;/h2&gt;

&lt;p&gt;The gap is obvious—we need lightweight, always-on profiling that shows:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Which layers are slow (forward and backward)
&lt;/li&gt;
&lt;li&gt;Real-time updates during training
&lt;/li&gt;
&lt;li&gt;Minimal overhead (1-2% measured on NVIDIA T4)
&lt;/li&gt;
&lt;li&gt;Layer-level memory tracking&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The dashboard shows you in real-time:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Which layer takes 40% of your training time&lt;/li&gt;
&lt;li&gt;Whether your dataloader is actually the bottleneck&lt;/li&gt;
&lt;li&gt;Where to optimize first&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;No guessing. Just data.&lt;/p&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;GitHub:&lt;/strong&gt; &lt;a href="https://github.com/traceopt-ai/traceml/" rel="noopener noreferrer"&gt;https://github.com/traceopt-ai/traceml/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;If you've ever wondered why your training is slow or hit mysterious OOM errors, give it a try. Would love your feedback.&lt;/p&gt;

&lt;p&gt;⭐ &lt;strong&gt;Star on GitHub&lt;/strong&gt; if you find it useful&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>machinelearning</category>
      <category>developer</category>
    </item>
  </channel>
</rss>
