You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Project Supervisor: Xuezhe Ma (xuezhema@isi.edu) (University of Southern California)
Reference
If you use this code as part of any published research, please acknowledge the following paper (it encourages researchers who publish their code!):
@inproceedings{yang-2022-stability,
author = "Yang, Chenghao and Ma, Xuezhe",
title = "Improving Stability of Fine-Tuning Pretrained Language Models via Component-Wise Gradient Norm Clipping",
booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",
year = "2022"
}
Dependency Installation
conda create -p ./env python=3.7
conda activate ./env # the environment position is optional, you can choose whatever places you like to save dependencies. Here I choose ./env for example.
pip install -r requirements.txt
cd ./CustomToolBox
pip install -e .# you also need to download the dataset (e.g., anli) and model checkpoints, see below
We will upload the experiment results as well as baseline replication results soon.
Currently doing some works to move the data from the institutional server. Stay tuned!
Special Acknowledgement
Our work is based on the finding of Marius Mosbach's ICLR 2021 paper on fine-tuning stability for pretrained language models.
We would like to thank Marius Mosbach for his help on baseline replication.
Also check out their wonderful docker-supported codebase!
About
Code and data of the EMNLP 2022 paper "Improving Stability of Fine-Tuning Pretrained Language Models via Component-Wise Gradient Norm Clipping""