Despite large neural networks demonstrating remarkable abilities to complete different tasks, they require excessive memory usage to store the optimization states for training. To alleviate this, the low rank adaptation (LoRA) is proposed to reduce the optimization states by training fewer parameters. However, LoRA restricts overall weight update matrices to be low-rank, limiting the model performance. In this work, we investigate the dynamics of LoRA and identify that it can be approximated by a random projection. Based on this observation, we propose Flora, which is able to achieve high-rank updates by resampling the projection matrices while enjoying the sublinear space complexity of optimization states. We conduct experiments across different tasks and model architectures to verify the effectiveness of our approach.
Bibtex
@misc{hao2024flora,
title={Flora: Low-Rank Adapters Are Secretly Gradient Compressors},
author={Yongchang Hao and Yanshuai Cao and Lili Mou},
year={2024},
eprint={2402.03293},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
Related Research
-
Designing Scalable Multi-Tenant Data Pipelines with Dagster’s Declarative Orchestration
Designing Scalable Multi-Tenant Data Pipelines with Dagster’s Declarative Orchestration
B. Zhang.
Research
-
Training foundation models up to 10x more efficiently with Memory-Mapped Datasets
Training foundation models up to 10x more efficiently with Memory-Mapped Datasets
T. Badamdorj, and M. Anand.
Research
-
DeepRRTime: Robust Time-series Forecasting with a Regularized INR Basis
DeepRRTime: Robust Time-series Forecasting with a Regularized INR Basis
C.S. Sastry, M. Gilany, K. Y. C. Lui, M. Magill, and A. Pashevich. Transactions on Machine Learning Research (TMLR)
Publications
Careers
Artificial Intelligence is reshaping finance. Every day, our teams uncover new opportunities that advance the field of AI, building products that impact millions of people across Canada and beyond. Explore open roles!
Explore opportunities