LasQ: Largest Singular Components Fine-Tuning for LLMs with Quantization

  • Xiang Zhao
  • , Beining Lin
  • , You Song*
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Large language models (LLMs) have demonstrated strong capabilities in various industries, but as the model parameters increase, the computational cost of fine-tuning the entire model becomes extremely high. To address this challenge, we focus on applying quantization and LoRA fine-tuning together in pre-training scenarios and propose an efficient parameter fine-tuning (PEFT) method, the LasQ (Largest Singular Components Fine-tuning for LLMs with Quantization) framework. Performs singular value decomposition on the pre-trained weights after quantization, using high-order singular value components to initialize the low rank adapter. We evaluate our method in natural language understanding, question answering, summarization, and natural language generation tasks. The experiments show that our method can significantly outperform existing methods with fewer training parameters. Compared with LoftQ and QLoRA methods, it has a 2%–15% improvement, and it can even achieve equivalent LoRA fine-tuning effects and full parameter fine-tuning effects.

Original languageEnglish
Title of host publicationNatural Language Processing and Chinese Computing - 13th National CCF Conference, NLPCC 2024, Proceedings
EditorsDerek F. Wong, Zhongyu Wei, Muyun Yang
PublisherSpringer Science and Business Media Deutschland GmbH
Pages42-54
Number of pages13
ISBN (Print)9789819794362
DOIs
StatePublished - 2025
Event13th CCF International Conference on Natural Language Processing and Chinese Computing, NLPCC 2024 - Hangzhou, China
Duration: 1 Nov 20243 Nov 2024

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume15361 LNAI
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference13th CCF International Conference on Natural Language Processing and Chinese Computing, NLPCC 2024
Country/TerritoryChina
CityHangzhou
Period1/11/243/11/24

Keywords

  • Finetune
  • LLMs
  • Quantization

Fingerprint

Dive into the research topics of 'LasQ: Largest Singular Components Fine-Tuning for LLMs with Quantization'. Together they form a unique fingerprint.

Cite this