liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Fast Randomized Low-Rank Adaptation of Pre-trained Language Models with PAC Regularization
Hong Kong Baptist Univ, Peoples R China.
Linköping University, Department of Computer and Information Science, The Division of Statistics and Machine Learning. Linköping University, Faculty of Science & Engineering.
Hong Kong Baptist Univ, Peoples R China.
2024 (English)In: FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, ASSOC COMPUTATIONAL LINGUISTICS-ACL , 2024, p. 5236-5249Conference paper, Published paper (Refereed)
Abstract [en]

Low-rank adaptation (LoRA) achieves parameter efficient fine-tuning for large language models (LLMs) by decomposing the model weight update into a pair of low-rank projection matrices. Yet, the memory overhead restricts it to scale up when the model size increases. We propose Randomized LoRA (RLoRA) which adopts Randomized Walsh-Hadamard Transform to achieve significant reduction in the size of trainable parameters compared to LoRA. At the same time, it allows a PAC-Bayes regularizer to be efficiently incorporated to improve generalization. We evaluate the effectiveness of RLoRA on LLMs RoBERTa, GPT-2 and LLaMA-7B using GLUE, E2E and math reasoning benchmarks. With a much lower memory requirement, RLoRA can give similar performance as the SOTA low-rank adaptation methods for these three tasks and significantly better performance under few-shot settings.

Place, publisher, year, edition, pages
ASSOC COMPUTATIONAL LINGUISTICS-ACL , 2024. p. 5236-5249
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:liu:diva-212057ISI: 001356731805022ISBN: 9798891760998 (print)OAI: oai:DiVA.org:liu-212057DiVA, id: diva2:1942648
Conference
62nd Annual Meeting of the Association-for-Computational-Linguistics (ACL) / Student Research Workshop (SRW), Bangkok, THAILAND, aug 11-16, 2024
Note

Funding Agencies|Hong Kong Government [RMGS2021_8_06]

Available from: 2025-03-06 Created: 2025-03-06 Last updated: 2025-03-06

Open Access in DiVA

No full text in DiVA

Search in DiVA

By author/editor
Qian, Dong
By organisation
The Division of Statistics and Machine LearningFaculty of Science & Engineering
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 50 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf