Fig. 3: The power of scale of delta-tuning methods. | Nature Machine Intelligence

Fig. 3: The power of scale of delta-tuning methods.

From: Parameter-efficient fine-tuning of large-scale pre-trained language models

Fig. 3

ao, We perform all delta-tuning methods on different scales of T5: T5SMALL(), T5BASE() and T5XXL(). We report the performance of Adapter in (a–c), LoRA in (df), Prefix-tuning in (gi), Last-layer tuning in (jl), and Selective-module tuning in (mo). From this figure, we can observe that with the scale of T5 increasing, all delta-tuning methods could converge faster and achieve better performance on MNLI, QNLI and SST-2.

Back to article page