Principal Singular Values and Singular Vectors Adaptation of Large Language Models

View the PDF file from the paper entitled Pissa: the main individual values and single single air conditioning of the large language models, by Fanxu Ming and 2 other authors
PDF HTML (experimental) view
a summary:To the PEFT linguistic models (PEFT) effectively, the lower adjustment method (Lora) is almost changed the form of the form $ \ Delta w \ in \ mathbb {r}^{M \ Times N} $ through a two -product product $ A \ in \ mathb {r} \ mathbb {r}^{r \ Times N} $, where $ R \ l \ min (M, N) is $, $ A $ with Gaussian noise, $ B $ B with zeros. Lora freezes the original form $ W and update the “Noise & Zero” adapter, which may lead to slow rapprochement. To overcome this restriction, we offer the main individual values and single single adaptation (Pissa). Pissa shares the same architecture as Lora, but it is to prepare the transformer matrix $ A $ A and $ B with the main components of the original matrix $ w, and puts the remaining ingredients in the remaining matrix $ w^{RES} \ in \ mathbb compared to Lora, Pissa updates the main ingredients with freezing the “remaining” parts, which allows rapprochement The fastest and performance of Mohsen. Pissa and Lora’s comparative experiments reveal across 12 different models, ranging from 184 meters to 70b, including 5 NLG and 8 NLU tasks, that Pissa is constantly outperforming Lora under identical experimental settings. On the GSM8K Standard, Mistral-7B is set with a 72.86 % Pissa, exceeding 67.7 % of Lora by 5.16 %. Due to the same architecture, Pissa is also compatible with quantitative measurement to reduce memory requirements for control. Compared to QLora, QPissa displays smaller quantity errors in the initial stages. Llama-3-70B polishing on GSM8K, QPissa gets 86.05 %, and QLora performance exceeds 81.73 %. Take advantage of the fast SVD technology, Pissa can be prepared in just a few seconds, which provides a small cost to move from Lora to Pissa. A code available in this URL https.
The application date
From: Mang Vanxo [view email]
[v1]
Wed, April 3, 2024 15:06:43 UTC (419 KB)
[v2]
Sun, April 14, 2024 15:24:10 UTC (457 KB)
[v3]
Tuesday, 28 May 2024 14:19:33 UTC (5562 KB)
[v4]
Wed, April 9, 2025 06:54:20 UTC (5,592 KB)
Don’t miss more hot News like this! AI/" target="_blank" rel="noopener">Click here to discover the latest in AI news!
2025-04-10 04:00:00