Advances in Parameter-Efficient Fine-Tuning of Large Language Models

The field of large language models is moving towards more efficient and effective fine-tuning methods. Recent developments have focused on parameter-efficient fine-tuning (PEFT) techniques, which reduce the number of trainable parameters while maintaining or improving performance. These methods have shown promising results in various applications, including speech recognition and natural language understanding. Notably, techniques such as low-rank adaptation (LoRA) and its variants have been widely adopted, and new methods like structure-learnable adapters and vector-based random tensor networks have emerged. These innovations enable more flexible and efficient adaptation of large language models to downstream tasks, while also providing insights into the underlying mechanisms of these models.

Some noteworthy papers in this area include: LobRA, which introduces a framework for joint fine-tuning of LoRA adapters with heterogeneous resource usages and parallel configurations, achieving significant reductions in GPU seconds required for joint fine-tuning. SSVD, which presents a comprehensive integration and benchmarking of PEFT methods within ESPnet and introduces structured SVD-guided fine-tuning for robust domain adaptation. TeRA, which proposes a vector-based random tensor network for high-rank adaptation of large language models, achieving high-rank weight updates while retaining parameter efficiency. IPA, which introduces a feature-aware projection framework that explicitly preserves information in the reduced hidden space, improving performance over LoRA and DoRA. L1RA, which dynamically assigns rank to LoRA adapters during fine-tuning, optimizing resource utilization and achieving comparable or better performance than other LoRA variants. OLieRA, which introduces Lie group theory into LLM fine-tuning, preserving parameter geometry while enforcing orthogonality constraints to task subspaces, and achieving state-of-the-art results on the Standard CL benchmark.

Sources

LobRA: Multi-tenant Fine-tuning over Heterogeneous Data

SSVD: Structured SVD for Parameter-Efficient Fine-Tuning and Benchmarking under Domain Shift in ASR

Structure-Learnable Adapter Fine-Tuning for Parameter-Efficient Large Language Models

TeRA: Vector-based Random Tensor Network for High-Rank Adaptation of Large Language Models

IPA: An Information-Preserving Input Projection Framework for Efficient Foundation Model Adaptation

L1RA: Dynamic Rank Assignment in LoRA Fine-Tuning

A Lightweight Framework for Trigger-Guided LoRA-Based Self-Adaptation in LLMs

Orthogonal Low-rank Adaptation in Lie Groups for Continual Learning of Large Language Models

Behind the Scenes: Mechanistic Interpretability of LoRA-adapted Whisper for Speech Emotion Recognition

Built with on top of