<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Truc Nguyen</title>
    <description>The latest articles on Forem by Truc Nguyen (@trng28).</description>
    <link>https://forem.com/trng28</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3681115%2Fad5c53ca-b95f-4484-aae4-5ae9fe2b7c6e.jpg</url>
      <title>Forem: Truc Nguyen</title>
      <link>https://forem.com/trng28</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/trng28"/>
    <language>en</language>
    <item>
      <title>[Boost]</title>
      <dc:creator>Truc Nguyen</dc:creator>
      <pubDate>Wed, 14 Jan 2026 07:55:00 +0000</pubDate>
      <link>https://forem.com/trng28/-59m2</link>
      <guid>https://forem.com/trng28/-59m2</guid>
      <description>&lt;div class="ltag__link"&gt;
  &lt;a href="/prema_ananda" class="ltag__link__link"&gt;
    &lt;div class="ltag__link__pic"&gt;
      &lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3203633%2Fd35ed6f5-9b36-4268-aafb-17660b97308e.png" alt="prema_ananda"&gt;
    &lt;/div&gt;
  &lt;/a&gt;
  &lt;a href="https://dev.to/prema_ananda/building-a-multi-agent-ai-brand-monitoring-system-with-n8n-and-brightdata-oe0" class="ltag__link__link"&gt;
    &lt;div class="ltag__link__content"&gt;
      &lt;h2&gt;Building a Multi-Agent AI Brand Monitoring System with n8n and BrightData&lt;/h2&gt;
      &lt;h3&gt;Prema Ananda ・ Aug 30 '25&lt;/h3&gt;
      &lt;div class="ltag__link__taglist"&gt;
        &lt;span class="ltag__link__tag"&gt;#devchallenge&lt;/span&gt;
        &lt;span class="ltag__link__tag"&gt;#n8nbrightdatachallenge&lt;/span&gt;
        &lt;span class="ltag__link__tag"&gt;#ai&lt;/span&gt;
        &lt;span class="ltag__link__tag"&gt;#webdev&lt;/span&gt;
      &lt;/div&gt;
    &lt;/div&gt;
  &lt;/a&gt;
&lt;/div&gt;


</description>
      <category>devchallenge</category>
      <category>n8nbrightdatachallenge</category>
      <category>ai</category>
      <category>webdev</category>
    </item>
    <item>
      <title>Flexora: Flexible Low-Rank Adaptation for Large Language Models</title>
      <dc:creator>Truc Nguyen</dc:creator>
      <pubDate>Tue, 13 Jan 2026 22:22:29 +0000</pubDate>
      <link>https://forem.com/trng28/flexora-flexible-low-rank-adaptation-for-large-language-models-3m13</link>
      <guid>https://forem.com/trng28/flexora-flexible-low-rank-adaptation-for-large-language-models-3m13</guid>
      <description>&lt;h3&gt;
  
  
  1. Current Problem
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Background:&lt;/strong&gt; Fine-tuning Large Language Models (LLMs) is highly resource-intensive. &lt;strong&gt;LoRA (Low-Rank Adaptation)&lt;/strong&gt; was introduced to address this by freezing the base model and training only a small number of additional parameters.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Limitation of LoRA:&lt;/strong&gt; Despite its efficiency, LoRA often suffers from &lt;strong&gt;overfitting&lt;/strong&gt;-performing well on training data but generalizing poorly to real-world tasks. Existing mitigation methods typically rely on manual tuning or lack flexibility across different tasks.&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  2. Proposed Solution: Flexora
&lt;/h3&gt;

&lt;p&gt;The authors introduce &lt;strong&gt;Flexora&lt;/strong&gt;, a novel method that &lt;strong&gt;automatically selects the most important layers&lt;/strong&gt; of a model to fine-tune, instead of tuning all layers or selecting them heuristically.&lt;/p&gt;




&lt;h3&gt;
  
  
  3. Mechanism (Three Stages)
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqkstwhvbdp7zqgzdud6l.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqkstwhvbdp7zqgzdud6l.png" alt=" " width="800" height="257"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Flexora formulates layer selection as a &lt;strong&gt;Hyperparameter Optimization (HPO)&lt;/strong&gt; problem. The process consists of three stages:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;a. Initialization Stage&lt;/strong&gt;&lt;br&gt;
A scalar weight parameter (denoted as $\alpha$) is attached to the LoRA modules of each layer in the model.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;b. Flexible Layer Selection Stage&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A small validation set is used to train the $\alpha$ weights via &lt;strong&gt;Unrolled Differentiation&lt;/strong&gt;.&lt;/li&gt;
&lt;li&gt;The system automatically learns which layers contribute most to the final performance.&lt;/li&gt;
&lt;li&gt;Layers with high scores are retained, while low-scoring layers are pruned.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;c. Fine-tuning Stage&lt;/strong&gt;&lt;br&gt;
Only the selected important layers are fine-tuned, while all other layers remain frozen. This significantly reduces computational cost and focuses learning on the most impactful components.&lt;/p&gt;




&lt;h3&gt;
  
  
  4. Results and Effectiveness
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Higher performance:&lt;/strong&gt; Flexora outperforms standard LoRA and other baseline methods on multiple benchmarks (HellaSwag, PIQA, RACE, etc.).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Reduced overfitting:&lt;/strong&gt; By eliminating redundant parameters, the model generalizes better-learning meaningful patterns rather than memorizing data.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Parameter efficiency:&lt;/strong&gt; Flexora typically uses only about &lt;strong&gt;50% of the parameters&lt;/strong&gt; required by LoRA while achieving superior performance.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Scalability:&lt;/strong&gt; The method generalizes well across different model families (LLaMA, Mistral, ChatGLM, etc.).&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  5. Key Insights from the Study
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Not all layers in an LLM are equally important for a given task.&lt;/li&gt;
&lt;li&gt;Flexora tends to prioritize &lt;strong&gt;early (input)&lt;/strong&gt; and &lt;strong&gt;late (output)&lt;/strong&gt; layers for fine-tuning, as these layers capture critical information related to input representation and output generation.&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;&lt;strong&gt;Summary:&lt;/strong&gt;&lt;br&gt;
Flexora is an intelligent extension of LoRA that automatically &lt;strong&gt;selects only the most valuable layers to learn&lt;/strong&gt;, resulting in models that are more efficient, more robust, and less prone to overfitting.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Paper:&lt;/strong&gt; &lt;a href="https://aclanthology.org/2025.acl-long.713.pdf" rel="noopener noreferrer"&gt;https://aclanthology.org/2025.acl-long.713.pdf&lt;/a&gt;&lt;/p&gt;

</description>
      <category>llm</category>
      <category>nlp</category>
      <category>lora</category>
    </item>
  </channel>
</rss>
