Adapters in Transformers. A New Paradigm for Transfer Learning…?

Speaker:
Jonas Pfeiffer (Technical University of Darmstadt)
Abstract:
Adapters have recently been introduced as an alternative transfer learning strategy. Instead of fine-tuning all weights of a pre-trained transformer-based model, small neural network components are introduced at every layer. While the pre-trained parameters are frozen, only the newly introduced adapter weights are fine-tuned, achieving an encapsulation of the down-stream task information in designated parts of the model. In this talk we will provide an introduction to adapter-training in natural language processing. We will go into detail on how the encapsulated knowledge can be leveraged for compositional transfer learning, as well as cross-lingual transfer. We will further briefly touch the efficiency of adapters in terms of trainable parameters as well as (wall-clock) training time. Finally, we will provide an outlook to recent alternative adapter approaches and training strategies.
Length:
01:11:59
Date:
29/11/2021
views: 668

Images:
Attachments: (video, slides, etc.)
199.0 MB
669 downloads