Next Article in Journal
Mean-Field Analysis with Random Perturbations to Detect Gliders in Cellular Automata
Next Article in Special Issue
Efficient Federated Learning with Pre-Trained Large Language Model Using Several Adapter Mechanisms
Previous Article in Journal
A Contemporary Approach of Integral Khan-Type Multivalued Contractions with Generalized Dynamic Process and an Application
Previous Article in Special Issue
Breast Cancer Diagnosis Using a Novel Parallel Support Vector Machine with Harris Hawks Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning

Command and Control Engineering College, Army Engineering University of PLA, Nanjing 210007, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(20), 4317; https://doi.org/10.3390/math11204317
Submission received: 3 September 2023 / Revised: 8 October 2023 / Accepted: 10 October 2023 / Published: 17 October 2023

Abstract

:
With the growing scale of pre-trained language models (PLMs), full parameter fine-tuning becomes prohibitively expensive and practically infeasible. Therefore, parameter-efficient adaptation techniques for PLMs have been proposed to learn through incremental updates of pre-trained weights, such as in low-rank adaptation (LoRA). However, LoRA relies on heuristics to select the modules and layers to which it is applied, and assigns them the same rank. As a consequence, any fine-tuning that ignores the structural information between modules and layers is suboptimal. In this work, we propose structure-aware low-rank adaptation (SaLoRA), which adaptively learns the intrinsic rank of each incremental matrix by removing rank-0 components during training. We conduct comprehensive experiments using pre-trained models of different scales in both task-oriented (GLUE) and task-agnostic (Yelp and GYAFC) settings. The experimental results show that SaLoRA effectively captures the structure-aware intrinsic rank. Moreover, our method consistently outperforms LoRA without significantly compromising training efficiency.

1. Introduction

With the scaling of model and corpus size [1,2,3,4,5], large language models (LLMs) have demonstrated an ability for in-context learning  [1,6,7] in various natural language processing (NLP) tasks, that is, learning from a few examples within context. Although in-context learning is now the prevalent paradigm for using LLMs, fine-tuning still outperforms it in task-specific settings. In such scenarios, a task-specific model is exclusively trained on a dataset comprising input–output examples specific to the target task. However, full parameter fine-tuning, which updates and stores all the parameters for different tasks, becomes impractical when dealing with large-scale models.
In fact, LLMs with billions of parameters can be effectively fine-tuned by optimizing only a few parameters [8,9,10]. This has given rise to a branch of parameter-efficient fine-tuning (PEFT) techniques [11,12,13,14,15,16] for model tuning. These techniques optimize a small fraction of the model parameters while keeping the rest fixed, thereby significantly reducing computational and storage costs. For example, LoRA [15] introduces trainable low-rank decomposition matrices into LLMs, enabling the model to adapt to a new task while preserving the integrity of the original LLMs and retaining the acquired knowledge. Fundamentally, this approach is built upon the assumption that updates to the weights of the pre-trained language model have a lower rank during adaptation to specific downstream tasks [8,9]. Thus, by reducing the rank of the incremental matrices, LoRA optimizes less than 0.5 % of the additional trainable parameters. Remarkably, this optimization achieves comparable or even superior performance to that of full parameter fine-tuning.
However, despite its advantages, LoRA also comes with certain limitations that warrant consideration. One limitation lies in LoRA’s reliance on heuristics to select the modules and layers to which it is applied. Though heuristics can be effective under specific circumstances, their lack of generalizability is a concern. This lack of generalizability can result in suboptimal performance, or even complete failure, when applied to new data. Another limitation is the assignment of the same rank to incremental matrices across different modules and layers. This tends to oversimplify the complex structural relationships and important disparities that exist within neural networks. This phenomenon is illustrated in Figure 1.
In this paper, we propose a novel approach called structure-aware low-rank adaptation (SaLoRA), which adaptively learns the intrinsic rank of each incremental matrix by removing rank-0 components. As shown in Figure 2, we introduce a diagonal gate matrix G = diag ( g 1 , , g r ) for each incremental matrix. The modified incremental matrix can be represented as Δ W = B G A . The incremental matrix is divided into triplets, where each triplet T i contains the i-th column of B , the i-th gate mask of G and the i-th row of A . Here, g i represents the binary “gate” that indicates the presence or absence of the i-th triplet. Although incorporating the active triplet count as a penalty term in the learning objective is unfeasible, we employ a differentiable relaxation method to selectively remove non-critical triplets by considering the L 0 norm [17,18]. The L 0 norm is equal to the number of non-zero triplets and encourages the model to deactivate less essential triplets. This strategy assigns a higher rank to crucial incremental matrices to capture task-specific information. Conversely, less significant matrices are pruned to possess a lower ranks preventing overfitting. However, A and B are not orthogonal, implying potential dependence among the triplets. Removing these triplets can result in a more significant deviation from the original matrix. To enhance training stability and generalization, we introduce orthogonality regularization for B and A . Furthermore, we integrate a density constraint and leverage Lagrangian relaxation [19] to control the number of valid parameters.
We conduct extensive experiments on a wide range of tasks and models to evaluate the effectiveness of SaLoRA. Specifically, we conduct experiments on the General Language Understanding Evaluation [20] benchmark in a task-oriented setting to assess the model’s performance. In addition, we evaluate the model’s performance in a task-agnostic setting by fine-tuning LLaMA-7B with a 50K cleaned instruction-following dataset [21], and then perform zero-shot task inference on two text style transfer tasks: sentiment transfer [22] and formality transfer [23]. The experimental results demonstrate that SaLoRA consistently outperforms LoRA without significantly compromising training efficiency.

2. Backgound

Transformer Architecture. The Transformer [24] is primarily constructed using two key submodules: a multi-head self-attention (MHA) layer and a fully connected feed-forward (FFN) layer. The MHA is defined as follows:
M H A ( Q , K , V ) = C o n c a t ( h e a d 1 , , h e a d h ) W O , h e a d i = A t e n t i o n ( Q W i Q , K W i K , V W i V )
where Q , K , V R n × d are input-embedding matrices; W O R d × d is an output projection; W i Q , W i K , W i V R d × d k are query, key and value projections of head i, respectively; n is sequence length; d is the embedding dimension; h is the number of heads and d k = d / h is the hidden dimension of the projection subspaces. The FFN consists of two linear transformations separated by a ReLU activation:
F F N ( x ) = R e L U ( x W U + b U ) W D + b D
where W U R d × d m and W D R d m × d .
Parameter-Efficient Fine-Tuning. With the growing size of models, recent works have developed three main categories of parameter-efficient fine-tuning (PEFT) techniques. These techniques optimize a small fraction of model parameters while keeping the rest fixed, thereby significantly reducing computational and storage costs [10]. For example, addition-based methods [11,12,13,25,26] introduce additional trainable modules or parameters that are not part of the original model or process. Specifcation-based methods [14,27,28] specify certain parameters within the original model or process as trainable, whereas the others remain frozen. Reparameterization-based methods [15,16,29], including LoRA, reparameterize existing parameters into a parameter-efficient form by transformation. In this study, we focus on reparameterization-based methods, with particular emphasis on LoRA.
Low-Rank Adaptation. LoRA, as introduced in the work of Hu et al. [15], represents a typical example of a reparameterization-based method. In LoRA, some pre-trained weights of LLMs’ dense layers are reparameterized by injecting trainable low-rank incremental matrices. This reparameterization only allows low-rank matrices to be updated, while keeping the original pre-trained weights frozen. By reducing the rank of these matrices, LoRA effectively reduces the number of parameters during the fine-tuning process of LLMs. Consider a pre-trained weight matrix W R d × k , accompanied by a low-rank incremental matrix Δ W = BA . For h = Wx , the modified forward pass is as follows:
h = Wx + α r Δ Wx = Wx + α r BAx
where B R d × r , A R r × k , with the rank r m i n ( d , k ) , and α is a constant scale hyperparameter. The matrix A adopts a random zero-mean Gaussian initialization, while the matrix B is initialized as a zero matrix. Consequently, the product Δ W = BA is initially set to zero at the beginning of training. Let B j and A j denote the j-th column of B and the j-th row of A , respectively. Using this notation, Δ W can be expressed as Δ W = j = 1 r B j A j .

3. Method

In this section, we will first give a brief introduction to parameter-efficient fine-tuning, and then discuss our proposed model based on the problem definition.

3.1. Problem Formalization

We consider the general problem of efficiently fine-tuning LLMs for specific downstream tasks. Firstly, let us introduce some notations. Consider a training corpus D = ( x i , y i ) i = 1 N , where N represents the number of samples. Each sample consists an input, x i , and its corresponding output, y i . We use the index i to refer to the incremental matrix, i.e., Δ W i = B i A i for i = 1 , , K , where K is the number of incremental matrices. However, LoRA’s assumption of identical ranks for each incremental matrix overlooks structural relationships and the varying importance of weight matrices across different modules and layers during fine-tuning. This oversight can potentially impact overall model performance. Our objective is to determine the optimal { r a n k Δ W i } i = 1 K on the fly. The optimization objective can be formulated as follows:
min W R ( W ) 1 N ( i = 1 N L ( f ( x i ; W ) , y i ) ) s . t . r a n k Δ W i r , k = 1 , , K .
where W = { Δ W i , , Δ W K } represents the sets of trainable parameters and L corresponds to a loss function, such as cross-entropy for classification. Note that r a n k Δ W i { 0 , 1 , , r } is an unknown parameter that needs to be optimized.

3.2. Structure-Aware Intrinsic Rank Using L 0 Norm

To find the optimal { r a n k Δ W i } i = 1 K on the fly, with minimal computational overhead during training, we introduce a gate matrix G to define the structure-aware intrinsic rank:
Δ W = B G A = j = 1 r g j B j A j
where the g j { 0 , 1 } serves as a binary “gate”, indicating the presence or absence of the j-th rank. The gate matrix G = diag ( g 1 , , g r ) is a diagonal matrix consisting of the pruning variables. By learning the variable g j , we can control the rank of each incremental matrix individually, rather than applying the same rank to all matrices. To deactivate non-critical rank-0 components, the ideal approach would be to apply L 0 norm regularization to the gate matrix G :
| | G | | 0 = j = 1 r g j
where r is the rank of incremental matrices. The L 0 norm measures the number of non-zero triplets; thus, optimizing L 0 would encourage the model to deactivate less important incremental matrices.
Unfortunately, the optimization objective involving | | G | | 0 is computationally intractable due to its non-differentiability, making it impossible to directly incorporate it as a regularization term in the objective function. Instead, we use a stochastic relaxation approach, where the gate variables g are treated as continuous variables distributed within the interval [ 0 , 1 ] . We leverage the reparameterization trick [30,31] to ensure that g remains differentiable. Following prior studies [17,19], we adopt the Hard-Concrete (HC) distribution as a continuous surrogate for random variables g, illustrated in Figure 3. The HC distribution applies a hard-sigmoid rectification to s, which can easily be sampled by first sampling u U ( 0 , 1 ) and then computing as follows:
s = S i g m o d log u 1 u + log θ τ × ζ γ + γ g = min 1 , max 0 , s
where θ is the trainable parameter of the distribution and τ is the temperature. The interval ( γ , ζ ) , with γ < 0 and ζ > 1 , enables the distribution to concentrate probability mass at the edge of the support. The final outputs g are rectified into [ 0 , 1 ] . By summing up the probabilities of the gates being non-zero, the L 0 norm regularization can be computed via a closed form, as follows:
E [ | | G | | 0 ] = j = 1 r E g j > 0 = j = 1 r S i g m o d log θ j τ log γ ζ
As g now represents the output of the parameterized HC distribution function and serves as an intermediate representation for the neural network, gradient-based optimization methods can perform gradient updates for θ = { θ 1 , , θ r } . For each training batch, we sample the gate mask and then share it across the training examples within the batch to enhance sampling efficiency.

3.3. Enhanced Stability Using Orthogonal Regularization

In deep networks, orthogonality plays a crucial role in preserving the norm of the original matrix during multiplication, preventing signal vanishing or exploding [32]. However, in LoRA, where B and A are not orthogonal, the dependence can lead to larger variations when removing certain columns or rows through L 0 regularization. This, in turn, leads to training instability and the potential for negative effects on generalization [16]. For this, we turn to orthogonal regularization, which enforces the orthogonality condition:
R o r t h ( B , A ) = | | B T B I | | F 2 + | | A A T I | | F 2
where I is the identity matrix.
Now, let us substitute Equations (8) and (9) into Equation (4) to derive the new training objective:
min W , Θ R ( W , Θ ) 1 N ( i = 1 N L ( f ( x i ; W ) , y i ) ) + λ i = 1 K E [ | | G i | | 0 ] + β i = 1 K R o r t h ( B i , A i )
where Θ = { θ i , , θ K } represents the sets of trainable parameters, and λ and β are two constant hyperparameters.

3.4. Controlled Budget Using Lagrangian Relaxation

If we only rely on Equation (10) to learn the intrinsic rank for each incremental matrix, the resulting parameter budget cannot be directly controlled. This limitation becomes problematic in many real-world applications that require a specific model size or parameter budget. To address this issue, we further introduce an additional density constraint on R W , Θ to guide the network towards achieving a specific desired budget.
min W R W 1 N ( i = 1 N L ( f ( x i ; W ) , y i ) ) + β i = 1 K R o r t h ( B i , A i ) s . t . C ( Θ ) i = 1 K E | | G i | | 0 × d i + k i # B i + # A i = b
where b represents the target density and # ( x ) counts the total number of parameters in matrix x. Δ W i = B i A i , where B i is of d i × r i , and A i is of r i × k i . However, lowering the density constraint poses a challenging and (not necessarily strictly) constrained optimization problem. To tackle this challenge, we leverage Lagrangian relaxation as an alternative approach, along with the corresponding min-max game:
max λ min W , Θ L ( W , Θ , λ ) R W , Θ + λ C ( Θ ) b 2
where λ R is the Lagrangian multiplier, which is jointly updated during training. The updates to λ would increase the training loss unless the equality constraint is satisfied, resulting in the desired parameter budget. We optimize the Lagrangian relaxation by simultaneously performing gradient descent on ( W , Θ ) and projected gradient ascent (to R + ) on λ , as demonstrated in previous works [19,33]. During the experiments, we observed that the term λ C ( Θ ) b 2 converged quickly. To enhance training efficiency, we only optimize ( Θ , λ ) between T s t a r t and T e n d time steps. We provide a summarized algorithm in Algorithm 1.
Algorithm 1 SaLoRA
  • Input: Dataset D ; total iterations T; target density b; hyperparameters τ , γ , ζ , β , η p , η c .
  • Output: The fine-tuned parameters { W , Θ }.
  •    for  t = 1 , , T  do
  •        Sample a mini-batch from D
  •        if  T s t a r t t < T e n d  then
  •           Sample a gate mask set G from HC distribution and share it across the mini-batch
  •           Compute the gradient L ( W , Θ , λ )
  •           Update W ( t + 1 ) = W ( t ) η p W L ( W ( t ) , Θ ( t ) , λ ( t ) )
  •           Update Θ ( t + 1 ) = Θ ( t ) η c Θ L ( W ( t ) , Θ ( t ) , λ ( t ) )
  •           Update λ ( t + 1 ) = λ ( t ) + η c λ ( t ) L ( W ( t ) , Θ ( t ) , λ ( t ) )
  •       else
  •           Compute the gradient L ( W )
  •           Update W ( t + 1 ) = W ( t ) η p W L ( W ( t ) )
  •       end if
  • end for
  • return The fine-tuned parameters { W ( T ) , Θ ( T ) , λ ( T ) }.

3.5. Inference

During training, the gate mask g i is a random variable drawn from the HC distribution. At inference time, we first calculate the expected value of each g i in G . If the value of g i is greater than 0, we retain the corresponding i-th low-rank triplet. This procedure enables us to obtain the deterministic matrices B and A .

4. Experiments

We evaluated the effectiveness of the proposed SaLoRA on RoBERTa [34] and LLaMA-7B in both task-oriented and task-agnostic settings.
Baselines. We compared SaLoRA with the following methods:
  • Fine-tuning (FT) is the most common approach for adaptation. To establish an upper bound for the performance of our proposed method, we fine-tuned all parameters within the model.
  • Adapting tuning, as proposed by Houlsby et al. [25], incorporates adapter layers between the self-attention module (and the MLP module) and the subsequent residual connection. Each adapter module consists of two fully connected layers with biases and a nonlinearity in between. This original design is referred to as AdapterH. Recently, Pfeiffer et al. [11] introduced a more efficient approach, applying the adapter layer only after the MLP module and following a LayerNorm. We call it AdapterP.
  • Prefix-tuning (Prefix) [12] prepends a sequence of continuous task-specific activations to the input. During tuning, prefix-tuning freezes the model parameters and only backpropagates the gradient to the prefix activations.
  • Prompt-tuning (Prompt) [13] is a simplified version of prefix-tuning, allowing the additional k tunable tokens per downstream task to be prepended to the input text.
  • LoRA, introduced by Hu et al. [15], is a state-of-the-art method for parameter-efficient fine-tuning. The original implementation of LoRA applied the method solely to query and value projections. However, empirical studies [16,35] have shown that extending LoRA to all matrices, including W Q , W K , W V , W O , W U and W D , can further improve its performance. Therefore, we compare our approach with this generalized LoRA configuration to maximize its effectiveness.
  • AdaLoRA, proposed by Zhang et al. [16], utilizes singular value decomposition (SVD) to adaptively allocate the parameter budget among weight matrices based on their respective importance scores. However, this baseline involves computationally intensive operations, especially for large matrices. The training cost can be significant, making it less efficient for resource-constrained scenarios.

4.1. Task-Oriented Performance

Models and Datasets. We evaluated the performance of different adaptive methods on the GLUE benchmark [20] using pre-trained RoBERTa-base (125M) and RoBERTa-large (355M) [34] models from the HuggingFace Transformers library [36]. See Appendix A for additional details on the datasets we used.
Implementation Details. For running all the baselines, we utilized a publicly available implementation [37]. We evaluated the performance of LoRA, AdaLoRA and SaLoRA at r = 8 . To maintain a controlled parameter budget, we set the desired budget ratio (b) to 0.50 for both SaLoRA and AdaLoRA. During training, we used the AdamW optimizer [38], along with the linear learning rate scheduler. During our experiments, we observed that using a larger learning rate ( η c ) significantly improved the learning process for both the gate matrices and Lagrange multiplier. Therefore, we set η c to 0.01 for all conducted experiments. We fine-tuned all models using an NVIDIA A100 (40 GB) GPU. Additional details can be found in Appendix B.
Main Results. We compared SaLoRA with the baseline methods under different model scale settings, and the experimental results on the GLUE development set are presented in Table 1. We can see that SaLoRA consistently achieved better or comparable performance compared with existing approaches for all datasets. Moreover, it even outperformed the FT method. SaLoRA’s superiority was particularly striking when compared with LoRA, despite both models having a similar parameter count of 1.33 M/3.54 M for base/large model scales. After training, SaLoRA effectively utilized only 0.5 × 1.33   M / 0.5 × 3.54 parameters, yet still attained superior performance. This observation emphasizes the effectiveness of our method in learning the intrinsic rank for incremental matrices.

4.2. Task-Agnostic Performance

Models and Datasets. We present the experiments conducted to evaluate the performance of the self-instruct tuned LLaMA-7B models on instruction-following data [21]. Our objective was to assess their capability in comprehending and executing instructions for arbitrary tasks. We evaluated model performance on two text style transfer datasets: Yelp [22] and GYAFC [23]). Text style transfer refers to the task of changing the style of a sentence to the desired style while preserving the style-independent content. The prompts used in these experiments can be found in Appendix C.
Implementation Details. We tuned the learning rate η p from { 8 × 10 5 , 3 × 10 5 , 1 × 10 4 , 3 × 10 4 , 8 × 10 4 , 1 × 10 3 } and kept the following hyperparameters fixed: r = 8 , b = 0.5 , η c = 0.01 , β = 0.1 , τ = 1.0 , γ = 0.1 and ζ = 1.1 . All models were fine-tuned on an NVIDIA A800 (80 G) GPU.
Furthermore, we compared the performance of SaLoRA with dataset-specific style transfer models, including StyTrans [15], StyIns [16] and TSST [17]. In contrast to SaLoRA, these models were trained on a specific dataset. To evaluate the performance of style transfer models, we used the following metrics: (1) Transfer accuracy (ACC) using a fine-tuned BERT-base [39] classifier with each dataset. (2) Semantic similarity to human references via BLEU [40] score. (3) Sentence fluency (PPL) via perplexity, as measured by KenLM [41].
Main Results. Table 2 presents our experimental results on the Yelp and GYAFC datasets. Compared with LoRA, our method SaloRA achieved better or comparable performance across all directions on both datasets. This demonstrates the effectiveness of our method. In the negative-to-positive transfer direction, though SaloRA’s transfer accuracy was lower than the dataset-specific models (e.g., StyIns achieved 92.40 compared with SaloRA’s 71), it still aligned with the human reference accuracy of 64.60. Furthermore, SaloRA exhibited a lower perplexity (PPL) compared with dataset-specific models. These results show that SaLoRA (including LoRA) aligns more closely with human writing tendencies. In the formal-to-informal transfer direction, we also observed that our transfer accuracy was lower than dataset-specific models. This disparity may be attributed to the inherent bias of a large model for generating more formal outputs. This can be verified from the fact that SaLoRA exhibited a significant improvement in the transfer accuracy compared with dataset-specific models.

4.3. Analysis

The Effect of Rank r. Figure 4 illustrates the experimental results of fine-tuning RoBERTa-large across different ranks. We see that the rank r significantly influenced the model’s performance. Both large and small values of r led to suboptimal results. This observation emphasizes that selecting the optimal value for r through heuristic approaches is not always feasible. Notably, SaLoRA consistently improved performance across all ranks when compared with the baseline LoRA. This suggests that SaLoRA effectively captured the “intrinsic rank” of the incremental matrix.
The Effect of Sparsity b. Figure 5 shows the experimental results of fine-tuning RoBERTa-large across various levels of sparsity. Remarkably, SaLoRA consistently exhibited enhanced performance across all sparsity levels compared with the baseline. This result suggests that SaLoRA’s modifications facilitated the acquisition of the “intrinsic rank” of the incremental matrix under different sparsities. It is noteworthy that SaLoRA’s performance even surpassed the results of LoRA under low sparsity conditions ( 0.125 ). The fact that SaLoRA can outperform LoRA even under low sparsity conditions highlights its capacity to capture and leverage parameters with a constrained budget. Consequently, SaLoRA’s efficacy could be expanded on a limited budget, making it a versatile method with a broader range of applications.
Ablation Study. We investigated the impact of Lagrangian relaxation and orthogonal regularization in SaLoRA. Specifically, we compared SaLoRA with the following variants: (i) SaLoRA λ = 0 : SaLoRA without Lagrangian relaxation; (ii) SaLoRA β = 0 : SaLoRA without orthogonal regularization. These variations involved the fine-tuning of the RoBERTa-base model on the CoLA, STS-B, and MRPC datasets. The target sparsity was set to 0.5 by default. SPS represented the expected sparsity of the incremental matrix. From Table 3, we see that:
  • Without Lagrangian relaxation, the parameter budget was uncontrollable, being 0.37 , 0.42 and 0.43 on the three datasets, respectively. Such results highlight the pivotal role that Lagrangian relaxation plays in controlling the allocation of the parameter budget. Nonetheless, it is worth noting that omitting Lagrange relaxation may lead to slight enhancements in performance. However, given the emphasis on control over the parameter budget, this incremental enhancement should be disregarded.
  • Without orthogonal regularization, the performance of SaLoRA degenerated. These results validate that incorporating orthogonal regularization into SaLoRA ensures the independence of doublets from one another, leading to a significant enhancement in its performance.
Visualization of Four Components. We plotted the visualization of expected sparsity b ^ , the Lagrangian multiplier λ and | | A T A I | | F 2 and | | B T B I | | F 2 to show whether these four components were regularized by Lagrangian relaxation and orthogonal regularization, respectively. Specifically, we fine-tuned the RoBERTa-base using SaLoRA on the CoLA, STS-B and MRPC datasets. The initial Lagrangian multiplier λ was 0 and the target sparsity b was 0.5. From Figure 6, we see that:
  • The expected sparsity b ^ decreased from 0.92 to about 0.50, and the Lagrangian multiplier λ kept increasing during training. The results indicate that the SaLoRA algorithm placed more emphasis on satisfying the constraints, eventually reaching a trade-off between satisfying the constraints and optimizing the objective function.
  • The values of | | A T A I | | F 2 and | | B T B I | | F 2 could be optimized to a highly negligible level (e.g., 0.001). Therefore, this optimization process enforced orthogonality upon both matrices A and B, guaranteeing the independence of doublets from one another.
Comparison of Training Efficiency. We analyzed the efficiency of SaLoRA in terms of memory and computational efficiency, as shown in Table 4. Specifically, we selected two scales of the RoBERTa model, that is, R o B b a s e and R o B l a r g e , and measured the peak GPU memory and training time under different batch sizes on an NVIDIA A100 (40 GB) GPU. From Table 4, we see that:
  • The GPU memory usages of both methods were remarkably similar. Such results demonstrate that SaLoRA does not impose significant memory overhead. The reason behind this is that SaLoRA only introduces gate matrices in contrast to LoRA. The total number of parameters was r × L × M . In this experiment, r denotes the rank of the incremental matrix (set at 8), L corresponds to the number of layers within the model (12 for R o B b a s e and 24 for R o B l a r g e ) and M stands for the number of modules in each layer (set at 6).
  • The training time of SaLoRA increased by 11% when using a batch size of 32 compared with LoRA. This suggests that the additional computational requirements introduced by SaLoRA are justified by its notable gains in performance. This is because SaLoRA is only utilized during a specific training phase ( T s t a r t to T e n d ) comprising 30% of the overall training time. With the remaining 70% being equivalent to LoRA, the overall impact on training time remains manageable.
The Resulting Rank Distribution. Figure 7 shows the resulting rank of each incremental matrix obtained from fine-tuning RoBERTa-base with SaLoRA. We observed that SaLoRA always assigned higher ranks to modules ( W U , W O and W V ) and layers (4, 5, 6 and 7). This aligns with the empirical results shown in Figure 1, indicating that modules ( W U , W O and W V ) and layers (4, 5, 6 and 7) play a more important role in model performance. Hence, these findings not only validate SaLoRA’s effective prioritization of critical modules and layers, but also emphasizes its capacity to learn the structure-aware intrinsic rank of the incremental matrix.

5. Conclusions

In this paper, we present SaLoRA, a structure-aware low-rank adaptation method that adaptively learns the intrinsic rank of each incremental matrix. In SaLoRA, we introduced a diagonal gate matrix to adjust the rank of the incremental matrix by penalizing the L 0 norm based on the count of activated gates. To enhance training stability and model generalization, we orthogonally regularized B and A . Furthermore, we integrated a density constraint and employed Lagrangian relaxation to control the number of valid ranks. In our experiments, we demonstrated that SaLoRA effectively captures the structure-aware intrinsic rank and consistently outperforms LoRA without significantly compromising training efficiency.

Author Contributions

Conceptualization, Y.H. and Z.P.; methodology, Y.H. and M.C.; validation, Y.H. and M.C.; formal analysis, Y.H. and Y.X.; investigation, Y.X.; resources, Y.X.; data curation, Y.H.; writing—original draft preparation, Y.H.; visualization, Y.H. and T.W.; supervision, Z.P.; project administration, Z.P.; funding acquisition, Z.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 62076251).

Data Availability Statement

Data sharing is not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
PLMsPre-trained language models
LLMsLarge language models
NLPNatural language process
LoRALow-rank adaptation
MHAMulti-head self-attention
FFNFeed-forward network
FTFine-tuning
PEFTParameter-efficient fine-tuning
HCHard-concrete distribution

Appendix A. Description of Datasets

Table A1. Description of datasets.
Table A1. Description of datasets.
DatasetDescriptionTrainValidTestMetrics
GLUE Benchmark
MNLIInference393.0k20.0k20.0kAccuracy
SST-2Sentiment analysis7.0k1.5k1.4kAccuracy
MRPCParaphrase detection3.7k4081.7kAccuracy
CoLALinguistic acceptability8.5k1.0k1.0kMatthews correlation
QNLIInference108.0k5.7k5.7kAccuracy
QQPQuestion answering364.0k40.0k391kAccuracy
RTEInference2.5k2763.0kAccuracy
STS-BTextual similarity7.0k1.5k1.4kPearson correlation
Text Style Transfer
Yelp-NegativeNegative reviews of restaurants and businesses17.7k2.0k500Accuracy
Similarity
Fluency
Yelp-PositivePositive reviews of restaurants and businesses26.6k2.0k500Accuracy
Similarity
Fluency
GYAFC-InformalInformal sentences from the Family and Relationships domain5.2k2.2k1.3kAccuracy
Similarity
Fluency
GYAFC-FormalFormal sentences from the Family and Relationships domain5.2k2.8k1.0kAccuracy
Similarity
Fluency

Appendix B. Training Details

We tuned the learning rate η p from { 5 × 10 5 , 7 × 10 5 , 9 × 10 5 , 2 × 10 4 , 3 × 10 4 , 4 × 10 4 , 5 × 10 4 , 7 × 10 4 } and selected the best learning rate.
Table A2. The hyperparameters we used for RoBERTa on the GLUE benchmark.
Table A2. The hyperparameters we used for RoBERTa on the GLUE benchmark.
ModelMNLISST-2CoLAQQPQNLIRTEMRPCSTS-B
# Epoch RoB base 1520202015404030
RoB large 1520202015404030
η p RoB base 9 × 10 5 4 × 10 4 5 × 10 4 4 × 10 4 5 × 10 4 4 × 10 4 4 × 10 4 7 × 10 4
RoB large 9 × 10 5 5 × 10 4 4 × 10 4 4 × 10 4 4 × 10 4 5 × 10 4 2 × 10 4 4 × 10 4
T s t a r t = 0.2 × # Epochs, T e n d = 0.5 × # Epochs. r = 8 , b = 0.5 , α = 16 , η c = 0.01 , β = 0.1 , τ = 1.0 , γ = 0.1 , ζ = 1.1 .

Appendix C. Prompts

Table A3. The prompts used in text style transfer.
Table A3. The prompts used in text style transfer.
Yelp: Negative → Positive
“Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{Please change the sentiment of the following sentence to be more positive.}
### Input:
{$Sentence}
### Response:”
Yelp: Positive → Negative
“Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{Please change the sentiment of the following sentence to be more negative.}
### Input:
{$Sentence}
### Response:”
GYAFC: Informal → Formal
“Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{Please rewrite the following sentence to be more formal.}
### Input:
{$Sentence}
### Response:”
GYAFC: Formal → Informal
“Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{Please rewrite the following sentence to be more informal.}
### Input:
{$Sentence}
### Response:”

References

  1. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–12 December 2020; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 1877–1901. [Google Scholar]
  2. Zeng, A.; Liu, X.; Du, Z.; Wang, Z.; Lai, H.; Ding, M.; Yang, Z.; Xu, Y.; Zheng, W.; Xia, X.; et al. Glm-130b: An open bilingual pre-trained model. arXiv 2022, arXiv:2210.02414. [Google Scholar]
  3. Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. LLaMA: Open and Efficient Foundation Language Models. arXiv 2023, arXiv:2302.13971. [Google Scholar] [CrossRef]
  4. OpenAI. GPT-4 Technical Report. arXiv 2023, arXiv:2303.08774. [Google Scholar] [CrossRef]
  5. Pavlyshenko, B.M. Financial News Analytics Using Fine-Tuned Llama 2 GPT Model. arXiv 2023, arXiv:2308.13032. [Google Scholar] [CrossRef]
  6. Kossen, J.; Rainforth, T.; Gal, Y. In-Context Learning in Large Language Models Learns Label Relationships but Is Not Conventional Learning. arXiv 2023, arXiv:2307.12375. [Google Scholar] [CrossRef]
  7. Dong, Q.; Li, L.; Dai, D.; Zheng, C.; Wu, Z.; Chang, B.; Sun, X.; Xu, J.; Li, L.; Sui, Z. A Survey on In-context Learning. arXiv 2022, arXiv:2301.00234. [Google Scholar] [CrossRef]
  8. Li, C.; Farkhoor, H.; Liu, R.; Yosinski, J. Measuring the Intrinsic Dimension of Objective Landscapes. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  9. Aghajanyan, A.; Gupta, S.; Zettlemoyer, L. Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tuning. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, 1–6 August 2021; pp. 7319–7328. [Google Scholar] [CrossRef]
  10. Ding, N.; Qin, Y.; Yang, G.; Wei, F.; Yang, Z.; Su, Y.; Hu, S.; Chen, Y.; Chan, C.; Chen, W.; et al. Parameter-efficient fine-tuning of large-scale pre-trained language models. Nat. Mac. Intell. 2023, 5, 220–235. [Google Scholar] [CrossRef]
  11. Pfeiffer, J.; Kamath, A.; Rücklé, A.; Cho, K.; Gurevych, I. AdapterFusion: Non-Destructive Task Composition for Transfer Learning. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Online, 19–23 April 2021; pp. 487–503. [Google Scholar] [CrossRef]
  12. Li, X.L.; Liang, P. Prefix-Tuning: Optimizing Continuous Prompts for Generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, 1–6 August 2021; pp. 4582–4597. [Google Scholar] [CrossRef]
  13. Lester, B.; Al-Rfou, R.; Constant, N. The Power of Scale for Parameter-Efficient Prompt Tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online and Punta Cana, Dominican Republic, 7–11 November 2021; pp. 3045–3059. [Google Scholar] [CrossRef]
  14. Ben Zaken, E.; Goldberg, Y.; Ravfogel, S. BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Dublin, Ireland, 22–27 May 2022; pp. 1–9. [Google Scholar] [CrossRef]
  15. Hu, E.J.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, S.; Wang, L.; Chen, W. LoRA: Low-Rank Adaptation of Large Language Models. In Proceedings of the International Conference on Learning Representations, Virtual Event, 25–29 April 2022. [Google Scholar]
  16. Zhang, Q.; Chen, M.; Bukharin, A.; He, P.; Cheng, Y.; Chen, W.; Zhao, T. Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning. In Proceedings of the the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  17. Louizos, C.; Welling, M.; Kingma, D.P. Learning Sparse Neural Networks through L_0 Regularization. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  18. Wang, Z.; Wohlwend, J.; Lei, T. Structured Pruning of Large Language Models. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 8–12 November 2020; pp. 6151–6162. [Google Scholar] [CrossRef]
  19. Gallego-Posada, J.; Ramirez, J.; Erraqabi, A.; Bengio, Y.; Lacoste-Julien, S. Controlled Sparsity via Constrained Optimization or: How I Learned to Stop Tuning Penalties and Love Constraints. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 29 November–1 December 2022; Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2022; Volume 35, pp. 1253–1266. [Google Scholar]
  20. Wang, A.; Singh, A.; Michael, J.; Hill, F.; Levy, O.; Bowman, S.R. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  21. Taori, R.; Gulrajani, I.; Zhang, T.; Dubois, Y.; Li, X.; Guestrin, C.; Liang, P.; Hashimoto, T.B. Stanford Alpaca: An Instruction-Following LLaMA Model. 2023. Available online: https://github.com/tatsu-lab/stanford_alpaca (accessed on 14 March 2023).
  22. Li, J.; Jia, R.; He, H.; Liang, P. Delete, Retrieve, Generate: A Simple Approach to Sentiment and Style Transfer. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), New Orleans, LA, USA, 1–6 June 2018; pp. 1865–1874. [Google Scholar] [CrossRef]
  23. Rao, S.; Tetreault, J. Dear Sir or Madam, May I Introduce the GYAFC Dataset: Corpus, Benchmarks and Metrics for Formality Style Transfer. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), New Orleans, LA, USA, 1–6 June 2018; pp. 129–140. [Google Scholar] [CrossRef]
  24. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.u.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  25. Houlsby, N.; Giurgiu, A.; Jastrzebski, S.; Morrone, B.; De Laroussilhe, Q.; Gesmundo, A.; Attariyan, M.; Gelly, S. Parameter-Efficient Transfer Learning for NLP. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 2790–2799. [Google Scholar]
  26. Mo, Y.; Yoo, J.; Kang, S. Parameter-Efficient Fine-Tuning Method for Task-Oriented Dialogue Systems. Mathematics 2023, 11, 3048. [Google Scholar] [CrossRef]
  27. Lee, J.; Tang, R.; Lin, J. What Would Elsa Do? Freezing Layers during Transformer Fine-Tuning. arXiv 2019, arXiv:1911.03090. [Google Scholar] [CrossRef]
  28. Guo, D.; Rush, A.; Kim, Y. Parameter-Efficient Transfer Learning with Diff Pruning. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, 1–6 August 2021; pp. 4884–4896. [Google Scholar] [CrossRef]
  29. Valipour, M.; Rezagholizadeh, M.; Kobyzev, I.; Ghodsi, A. DyLoRA: Parameter-Efficient Tuning of Pre-trained Models using Dynamic Search-Free Low-Rank Adaptation. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, Dubrovnik, Croatia, 2–6 May 2023; pp. 3274–3287. [Google Scholar]
  30. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  31. Rezende, D.J.; Mohamed, S.; Wierstra, D. Stochastic Backpropagation and Approximate Inference in Deep Generative Models. In Proceedings of the 1st International Conference on Machine Learning, Bejing, China, 22–24 June 2014; Xing, E.P., Jebara, T., Eds.; PMLR: Bejing, China, 2014; Volume 32, pp. 1278–1286. [Google Scholar]
  32. Brock, A.; Lim, T.; Ritchie, J.; Weston, N. Neural Photo Editing with Introspective Adversarial Networks. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  33. Lin, T.; Jin, C.; Jordan, M. On Gradient Descent Ascent for Nonconvex-Concave Minimax Problems. In Proceedings of the 37th International Conference on Machine Learning, Vienna, Austria, 13–18 July 2020; Daumé, H., III, Singh, A., Eds.; PMLR: Vienna, Austria, 2020; Volume 119, pp. 6083–6093. [Google Scholar]
  34. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  35. He, J.; Zhou, C.; Ma, X.; Berg-Kirkpatrick, T.; Neubig, G. Towards a Unified View of Parameter-Efficient Transfer Learning. In Proceedings of the International Conference on Learning Representations, Virtual Event, 25–29 April 2022. [Google Scholar]
  36. Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; et al. Transformers: State-of-the-Art Natural Language Processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Online, 16–20 November 2020; pp. 38–45. [Google Scholar] [CrossRef]
  37. Mangrulkar, S.; Gugger, S.; Debut, L.; Belkada, Y.; Paul, S. PEFT: State-of-the-Art Parameter-Efficient Fine-Tuning Methods. 2022. Available online: https://github.com/huggingface/peft (accessed on 6 July 2023).
  38. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  39. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 3–5 June 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
  40. Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.J. Bleu: A Method for Automatic Evaluation of Machine Translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Philadelphia, PA, USA, 6–12 July 2002; pp. 311–318. [Google Scholar] [CrossRef]
  41. Heafield, K. KenLM: Faster and Smaller Language Model Queries. In Proceedings of the Sixth Workshop on Statistical Machine Translation, Edinburgh, UK, 30–31 July 2011; pp. 187–197. [Google Scholar]
Figure 1. Fine-tuning performance of LoRA across different modules and layers with varying ranks on MRPC.
Figure 1. Fine-tuning performance of LoRA across different modules and layers with varying ranks on MRPC.
Mathematics 11 04317 g001
Figure 2. Structure-aware low-rank adaptation.
Figure 2. Structure-aware low-rank adaptation.
Mathematics 11 04317 g002
Figure 3. Hard-Concrete distribution with different parameters.
Figure 3. Hard-Concrete distribution with different parameters.
Mathematics 11 04317 g003
Figure 4. Fine-tuning performance under different ranks.
Figure 4. Fine-tuning performance under different ranks.
Mathematics 11 04317 g004
Figure 5. Fine-tuning performance under different sparsity levels.
Figure 5. Fine-tuning performance under different sparsity levels.
Mathematics 11 04317 g005
Figure 6. Visualization of expected sparsity b ^ and the Lagrangian multiplier λ under Lagrangian relaxation, and | | A T A I | | F 2 and | | B T B I | | F 2 under orthogonal regularization: (a) expected sparsity b ^ ; (b) Lagrangian multiplier λ ; (c) A of W O at the first layer; and (d) B of W O at the first layer.
Figure 6. Visualization of expected sparsity b ^ and the Lagrangian multiplier λ under Lagrangian relaxation, and | | A T A I | | F 2 and | | B T B I | | F 2 under orthogonal regularization: (a) expected sparsity b ^ ; (b) Lagrangian multiplier λ ; (c) A of W O at the first layer; and (d) B of W O at the first layer.
Mathematics 11 04317 g006
Figure 7. The resulting rank of each incremental matrix obtained from fine-tuning RoBERTa-base on MRPC with SaLoRA. The initial rank is set at 8, and the target sparsity is 0.5. The x-axis is the layer index and the y-axis represents different types of modules.
Figure 7. The resulting rank of each incremental matrix obtained from fine-tuning RoBERTa-base on MRPC with SaLoRA. The initial rank is set at 8, and the target sparsity is 0.5. The x-axis is the layer index and the y-axis represents different types of modules.
Mathematics 11 04317 g007
Table 1. Results with RoBERTa-base and RoBERTa-large on GLUE development set. We report the Pearson correlation for STS-B, Matthew’s correlation for CoLA, and accuracy for other tasks. We report the mean and maximum deviation of 5 runs using different random seeds. The best results are shown in bold. † indicates numbers published in prior works.
Table 1. Results with RoBERTa-base and RoBERTa-large on GLUE development set. We report the Pearson correlation for STS-B, Matthew’s correlation for CoLA, and accuracy for other tasks. We report the mean and maximum deviation of 5 runs using different random seeds. The best results are shown in bold. † indicates numbers published in prior works.
Model and Method# TrainableMNLISST-2CoLAQQPQNLIRTEMRPCSTS-BALL
ParametersACCACCMathewACCACCACCACCPearsonAvg
RoB base ( FT ) 125.00 M87.694.863.691.992.878.790.291.286.4
RoB base ( Prefix ) 1.33 M 82 . 58 ± 0.24 91 . 65 ± 0.44 47 . 45 ± 4.47 85 . 98 ± 0.11 85 . 91 ± 0.44 60 . 47 ± 0.54 88 . 09 ± 1.37 87 . 49 ± 0.63 78.70
RoB base ( Prompt ) 0.62 M 79 . 14 ± 0.99 88 . 33 ± 1.44 42 . 35 ± 5.12 73 . 30 ± 5.84 80 . 51 ± 1.64 55 . 81 ± 2.67 69 . 75 ± 0.59 76 . 94 ± 3.59 70.77
RoB base ( LoRA ) 1.33 M 87 . 49 ± 0.44 94 . 77 ± 0.53 61 . 22 ± 2.09 91 . 39 ± 0.18 92 . 85 ± 0.28 79 . 24 ± . 21 89 . 46 ± 1.23 90 . 89 ± 0.18 85.91
RoB base ( AdaLoRA ) 1.33 M 87 . 93 ± 0 . 20 94 . 59 ± 0.21 59 . 29 ± 1.06 90 . 94 ± 0.13 92 . 61 ± 0.10 76 . 39 ± 1.30 87 . 35 ± 0.39 90 . 87 ± 0.15 85.00
RoB base ( SaLoRA ) 1.33 M 87 . 83 ± 0.04 95 . 14 ± 0 . 73 63 . 39 ± 1.79 91 . 46 ± 0.09 92 . 99 ± 0 . 21 81 . 01 ± 0 . 87 90 . 20 ± 0 . 74 91 . 13 ± 0.17 86.64
RoB large ( FT ) 356.05 M90.296.468.092.294.786.690.992.488.9
RoB large ( Adapt P ) 4.05 M 90 . 2 ± 0.3 96 . 1 ± 0.3 68 . 3 ± 1.0 91 . 9 ± 0.1 94 . 8 ± 0.2 83 . 8 ± 2.9 90 . 2 ± 0.7 92 . 1 ± 0.7 88.4
RoB large ( Adapt H ) 7.05 M 89 . 5 ± 0.5 96 . 2 ± 0.3 66 . 5 ± 4.4 92 . 1 ± 0.1 94 . 7 ± . 2 83 . 4 ± 1.1 88 . 7 ± 2.9 91 . 0 ± 1.7 87.8
RoB large ( Prefix ) 3.02 M 88 . 61 ± 0.12 94 . 70 ± 0.44 60 . 06 ± 1.44 87 . 57 ± 0.25 89 . 60 ± 0.37 77 . 33 ± 1.37 89 . 85 ± . 88 89 . 97 ± 3.37 84.71
RoB large ( Prompt ) 1.09 M 85 . 65 ± 2.54 93 . 95 ± 0.83 58 . 34 ± 2.63 83 . 98 ± 1.45 84 . 92 ± 3.28 58 . 70 ± 4.83 74 . 22 ± 2.65 80 . 47 ± 0.71 77.53
RoB large ( LoRA ) 3.41 M 89 . 96 ± 0.12 96 . 10 ± 0.11 68 . 76 ± 1 . 75 88 . 67 ± 0.88 94 . 86 ± 0.07 85 . 49 ± 1.51 90 . 93 ± 0.74 92 . 25 ± 0.17 88.38
RoB large ( AdaLoRA ) 3.54 M 90 . 84 ± 0 . 03 96 . 29 ± 0.19 67 . 61 ± 0.12 91 . 12 ± 0.26 94 . 82 ± 0.11 86 . 28 ± 0.36 89 . 89 ± 0.31 92 . 27 ± 0.16 88.64
RoB large ( SaLoRA ) 3.54 M 90 . 67 ± 0.07 96 . 63 ± 0 . 28 68 . 37 ± 0.34 91 . 95 ± 0.08 94 . 98 ± 0 . 09 87 . 80 ± 1 . 29 91 . 81 ± 1 . 57 92 . 43 ± 0 . 18 89.33
Table 2. Automatic evaluation results on Yelp and GYAFC datasets. ↑ indicates that higher values mean better performance, and vice versa.
Table 2. Automatic evaluation results on Yelp and GYAFC datasets. ↑ indicates that higher values mean better performance, and vice versa.
Model and MethodYelpGYAFC
Negative to PositivePositive to NegativeInformal to FormalFormal to Informal
ACCBLEUPPLACCBLEUPPLACCBLEUPPLACCBLEUPPL
Reference64.60100102.6293.8010077.5388.4410066.8687.63100105.28
StyTrans88.4025.85173.3594.2024.90141.8832.8154.91144.1580.8627.69201.78
StyIns92.4025.98116.0189.6026.08105.7954.7360.8796.5380.5730.25132.54
TSST91.2028.95112.8694.4028.83101.9265.6261.8387.0485.8733.54128.78
LaA 7 B 2.2033.58208.690.8031.12156.1412.0160.18189.787.7534.61145.43
LaA 7 B ( LoRA ) 71.0025.9682.2092.8031.8383.0389.3461.0668.5234.4541.5982.96
LaA 7 B ( SaLoRA ) 73.2024.7676.4994.6031.9687.4189.6361.7667.5339.0440.9179.54
Table 3. Ablation studies on Lagrangian relaxation and orthogonal regularization.
Table 3. Ablation studies on Lagrangian relaxation and orthogonal regularization.
MethodMRPCSTS-BCoLA
ACCSPSACCSPSACCSPS
SaLoRA 90 . 20 ± 0 . 74 0 . 51 ± 0.00 91 . 13 ± 0.17 0 . 52 ± 0.00 63 . 39 ± 1.79 0 . 52 ± 0.00
SaLoRA λ = 0 89 . 95 ± 0.49 0 . 37 ± 0.02 91 . 20 ± 0 . 12 0 . 42 ± 0.03 63 . 65 ± 3 . 39 0 . 43 ± 0.02
SaLoRA β = 0 90 . 00 ± 0.94 0 . 51 ± 0.00 90 . 78 ± 0.27 0 . 52 ± 0.00 62 . 89 ± 2.16 0 . 52 ± 0.00
Table 4. Comparison of training efficiency between LoRA and SaLoRA on the MRPC dataset.
Table 4. Comparison of training efficiency between LoRA and SaLoRA on the MRPC dataset.
ModelBSMethodGPU MemTime
RoB base 16LoRA3.54 GB15 min
SaLoRA3.54 GB20 min
32LoRA5.34 GB14 min
SaLoRA5.35 GB15 min
64LoRA9.00 GB13 min
SaLoRA9.00 GB14 min
RoB large 16LoRA7.44 GB44 min
SaLoRA7.46 GB53 min
32LoRA12.16 GB40 min
SaLoRA12.18 GB44 min
64LoRA21.80 GB38 min
SaLoRA21.82 GB41 min
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hu, Y.; Xie, Y.; Wang, T.; Chen, M.; Pan, Z. Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning. Mathematics 2023, 11, 4317. https://doi.org/10.3390/math11204317

AMA Style

Hu Y, Xie Y, Wang T, Chen M, Pan Z. Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning. Mathematics. 2023; 11(20):4317. https://doi.org/10.3390/math11204317

Chicago/Turabian Style

Hu, Yahao, Yifei Xie, Tianfeng Wang, Man Chen, and Zhisong Pan. 2023. "Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning" Mathematics 11, no. 20: 4317. https://doi.org/10.3390/math11204317

APA Style

Hu, Y., Xie, Y., Wang, T., Chen, M., & Pan, Z. (2023). Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning. Mathematics, 11(20), 4317. https://doi.org/10.3390/math11204317

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop