Dave's Blog

Prefix Tuning: Lightweight Adaptation of Large Language Models for Customized Natural Language Generation

Source: https://arxiv.org/abs/2101.00190 Introduction Natural language generation (NLG) models like the GPT series have enabled remarkable progress on conditional text generation tasks such as summarization and table-to-text.  However, fine-tuning these large pretrained models on downstream NLG tasks requires updating all model parameters, which is computationally expensive. Fortunately, researchers have identified lightweight fine-tuning methods that reduce the number of parameters that must be updated when fine-tuning a model for a specific task. One such method is prefix tuning, proposed by Li and Liang (2021). Prefix tuning keeps the parameters of the pretrained model fixed, and only trains a small continuous “prefix” that

Read More »

Share

Most Popular

Categories