Transformer-Adaptation-Playbook
PublicAn empirical study of Transformer adaptation techniques. Pre-training from scratch (MLM), classic fine-tuning, and from-scratch implementations of PEFT methods (LoRA, Adapters). Tuning both encoder (BERT) and decoder (OPT) models.