Zihao Fu,1 Haoran Yang,2 Anthony Man-Cho So,2 Wai Lam,2 Lidong Bing,3 Nigel Collier1
1Language Technology Lab, University of Cambridge
2The Chinese University of Hong Kong
3DAMO Academy, Alibaba Group
2The Chinese University of Hong Kong
3DAMO Academy, Alibaba Group
[Paper (Full+Appendix)] [Slides] [Video]
- This paper gives a comprehensive explanation of why parameter-efficient models (such as Adapters, LoRA, Bitfit, etc.) achieve promising results.
- This paper unveils how the sparsity itself improves the model stability and generalization capability theoretically and empirically.
- This paper proposes a provable approximately best method to choose the tunable parameters for parameter-efficient models.
This code is the SAM model proposed in the paper. We suggest to create a new conda env to install the dependencies.
git clone https://github.com/fuzihaofzh/AnalyzeParameterEfficientFinetune.git
cd AnalyzeParameterEfficientFinetune
./scripts/install.sh
Run the following code to train our SAM model on the CoLA dataset.
./scripts/train.sh