LoRA: LOW-RANK ADAPTATION OF LARGE LANGUAGE MODELS (2021)
LoRA is motivated by the findings of Li et al. (2018) and Aghajanyan et al. (2020), which show that overparameterized models tend to converge to solutions that lie within a low-dimensional intrinsic subspace.
The intrinsic dimension refers to the minimum number of trainable parameters needed to reach satisfactory performance on a given task.
LoRA introduces low-rank adaptation by decomposing the weight matrices in dense layers. Instead of fine-tuning all parameters of a pre-trained model, LoRA freezes the original weights and learns two low-rank matrices during training.