Parameter-Efficient Transfer Learning for NLP

Compact adapter layers in each transformer block — freeze pre-trained weights, train only 3–8% of parameters for downstream tasks.
Efficient Adaptation
Author

Imad Dabbura

Published

August 7, 2020

Parameter-Efficient Transfer Learning for NLP

#nlp #llm #fine-tuning

Back to top