Publication | Open Access
HyperPELT: Unified Parameter-Efficient Language Model Tuning for Both Language and Vision-and-Language Tasks
12
Citations
16
References
2023
Year
Unknown Venue
Shared HypernetworkLlm Fine-tuningEngineeringMachine LearningMultilingual PretrainingLarge Language ModelNatural Language ProcessingMultimodal LlmSyntaxVision-and-language TasksComputational LinguisticsMulti-task LearningGrammarLanguage StudiesMachine TranslationLarge Ai ModelComputer ScienceGlue BenchmarkDeep LearningBoth LanguageVisual ModalityLinguistics
With the scale and capacity of pretrained models growing rapidly, parameter-efficient language model tuning has emerged as a popular paradigm for solving various NLP and Vision-and-Language (V&L) tasks. In this paper, we design a unified parameter-efficient multitask learning framework that works effectively on both NLP and V&L tasks. In particular, we use a shared hypernetwork that takes trainable hyper-embeddings and visual modality as input, and outputs weights for different modules in a pretrained language model, such as the parameters inserted into multi-head attention blocks (i.e., prefix-tuning) and feed-forward blocks (i.e., adapter-tuning.). Our proposed framework adds fewer trainable parameters in multi-task learning while achieving superior performances and transfer ability compared to state-of-the-art methods. Empirical results on the GLUE benchmark and multiple V&L tasks confirm the effectiveness of our framework.
| Year | Citations | |
|---|---|---|
Page 1
Page 1