HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

GIFT-SW: Gaussian noise Injected Fine-Tuning of Salient Weights for LLMs

Maxim Zhelnin Viktor Moskvoretskii Egor Shvetsov Egor Venediktov Mariya Krylova Aleksandr Zuev Evgeny Burnaev

GIFT-SW: Gaussian noise Injected Fine-Tuning of Salient Weights for LLMs

Abstract

Parameter Efficient Fine-Tuning (PEFT) methods have gained popularity anddemocratized the usage of Large Language Models (LLMs). Recent studies haveshown that a small subset of weights significantly impacts performance. Basedon this observation, we introduce a novel PEFT method, called Gaussian noiseInjected Fine Tuning of Salient Weights (GIFT-SW). Our method updates onlysalient columns, while injecting Gaussian noise into non-salient ones. Toidentify these columns, we developeda generalized sensitivity metric thatextends and unifies metrics from previous studies. Experiments with LLaMAmodels demonstrate that GIFT-SW outperforms full fine-tuning and modern PEFTmethods under the same computational budget. Moreover, GIFT-SW offers practicaladvantages to recover performance of models subjected to mixed-precisionquantization with keeping salient weights in full precision.

Code Repositories

On-Point-RND/GIFT_SW
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
parameter-efficient-fine-tuning-on-boolqLLaMA2-7b
Accuracy (% ): 82.63
parameter-efficient-fine-tuning-on-hellaswagLLaMA2-7b
Accuracy (% ): 76.68
parameter-efficient-fine-tuning-on-winograndeLLaMA2-7b
Accuracy (% ): 70.80

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
GIFT-SW: Gaussian noise Injected Fine-Tuning of Salient Weights for LLMs | Papers | HyperAI