LoRA Retains More
LoRA Retains More
Atharv Mittal
Vision and Language Group, Indian Institute of Technology Roorkee, Roorkee, Uttarakhand, India - 247667
atharv m@mfs.iitr.ac.in
arXiv:2411.11907v1 [cs.LG] 16 Nov 2024
Figure 1: Results of ResNet-50 and ViT When Tested Using Various Unlearning Approaches (in percent accuracy)
dataset. Our unlearning task focused on removing the influ- Future Scope
ence of the forget class while maintaining performance on There is significant potential for further research and ex-
the remaining classes. To establish an exact unlearning base- perimentation to strengthen and validate our hypothesis. A
line, we retrain both the models on the remaining dataset promising avenue for future research is the application of
for 200 and 90 epochs respectively. We used L2 Pruning to this method to Large Language Models (LLMs) and Vision-
prune 50% of the specific layers in each model, Convolu- Language Models (VLMs). These models, with their vast
tional layers were pruned in ResNet50 and linear and atten- parameter spaces, emphasize the need for efficient unlearn-
tion layers were pruned in ViT. After final finetuning on the ing techniques. Although computational constraints limited
remaining dataset for 5/10 epochs, we evaluate the models our ability to explore this direction, scaling our approach
based on the following metrics: to these larger models could help develop adaptable and
• Unlearning accuracy (UA): 1-Acc(Df), where Acc(Df) is privacy-preserving AI systems.
the accuracy of the unlearned model on the forget dataset.
• Membership inference attack (MIA-Efficacy): Applying Conclusion
the confidence-based MIA predictor to the unlearned This study addresses the challenge of machine unlearning in
model on the forgetting dataset (Df). A higher MIA- light of growing privacy regulations and the need for adapt-
Efficacy implies less information about Df in the model. able AI systems. We present a novel approach, PruneLoRA
• Remaining accuracy (RA): This refers to the accuracy of to LoRA to fine-tune sparse models. Our findings highlight
the unlearned model on the retain dataset. the efficacy of LoRA, especially when combined with prun-
• Testing accuracy (TA): This refers to the accuracy of the ing, in achieving high unlearning performance with minimal
unlearned model on the testing dataset of the remaining computational cost and memory requirements while main-
classes. taining general accuracy on remaining classes. These results
advances the research in exploring parameter efficient ma-
• Run-time efficiency (RTE): This measures the computa- chine approximate unlearning techniques, thus laying the
tion efficiency of the MU method (run time cost). groundwork for applying these methods to complex mod-
• GPU Memory (GPU): This measures the memory re- els such as Large Language Models and Vision-Language
quirements of the MU method for a model. Models.
Results References
Table 1 presents the accuracy metrics for both model un- Biderman, D.; Portes, J.; Ortiz, J. J. G.; Paul, M.; Greengard,
der the given five paradigms: It is observed that all methods P.; Jennings, C.; King, D.; Havens, S.; Chiley, V.; Frankle,
achieved perfect or near-perfect Unlearning Accuracy (UA) J.; Blakeney, C.; and Cunningham, J. P. 2024. LoRA Learns
and Membership Inference Attack (MIA) efficacy, indicat- Less and Forgets Less. Transactions on Machine Learning
ing successful removal of the target class information. For Research. Featured Certification.
ResNet-50, PruneLoRA outperformed all methods, achiev- Hu, E. J.; yelong shen; Wallis, P.; Allen-Zhu, Z.; Li, Y.;
ing the highest Remaining (RA) and Testing accuracy (TA), Wang, S.; Wang, L.; and Chen, W. 2022. LoRA: Low-Rank
while experiencing near-perfect UA. For the ViT model, Adaptation of Large Language Models. In International
PruneLoRA significantly outperformed other methods (ex- Conference on Learning Representations.
cept LoRA) in terms of RA and TA. Moreover, while LoRA
Liu, J.; Ram, P.; Yao, Y.; Liu, G.; Liu, Y.; SHARMA, P.;
demonstrated a drastically low UA, PruneLoRA achieved
Liu, S.; et al. 2024. Model sparsity can simplify machine
perfect UA. These results suggest that PruneLoRA offers a
unlearning. Advances in Neural Information Processing Sys-
balance between effective unlearning, retained model per-
tems, 36.
formance, and computational efficiency.
Appendix
Experiment Details
We trained ResNet-50 and Vision Transformer (ViT) on
CIFAR10, using custom implementations. The ResNet-50
model was trained for 200 epochs, and the ViT model was
trained for 90 epochs, both on a P100 GPU. They achieved
a test accuracy of 95.56% and 83.77% respectively. (a) Fine-tuned R (b) PruneFT R (c) LoRA R
All further experiments were conducted on a T4 GPU.
To allow for meaningful comparison between the various
fine-tuning techniques employed, we consistently used the
Adam optimizer with a learning rate of 10−3 , along with
cross-entropy loss.
Future Scope
Due to lack of computational resources and funding, we
were only able to perform a limited number of experiments,