LoRA: Low-Rank Adaptation of Large Language Models

Low-rank decomposition matrices for transformer attention weights — minimal trainable parameters, no inference latency overhead, efficient task switching.
Efficient Adaptation
Author

Imad Dabbura

Published

June 11, 2022

LoRA: Low-Rank Adaptation of Large Language Models

#nlp #llm #fine-tuning

Back to top