Parameter-Efficient Multi-Task Learning via Optimized Continuous Prompts
A novel approach known as Parameter-Efficient Multi-Task Learning (PEML) has been introduced for the simultaneous fine-tuning of large language models (LLMs) across various tasks. Current PEFT techniques, such as LoRA and Prefix Tuning, focus on individual tasks; LoRA modifies model weights but overlooks prompt tuning in multi-task scenarios, whereas Prefix Tuning's straightforward design restricts its adaptability to multiple tasks. PEML leverages optimized continuous prompts to facilitate effective multi-task fine-tuning, minimizing resource usage by merging tasks within a single model. This method meets the increasing need for multi-task LLMs, which benefit from reduced data requirements due to shared characteristics among tasks. The research can be found on arXiv with the ID 2605.14055.
Key facts
- PEML stands for Parameter-Efficient Multi-Task Learning with Optimized Continuous Prompts.
- The paper is published on arXiv with ID 2605.14055.
- PEFT methods like LoRA and Prefix Tuning are designed for single-task adaptation.
- LoRA aligns model weights but overlooks prompt tuning in multi-task learning.
- Prefix Tuning uses a simple architecture that limits multi-task adaptation.
- PEML introduces optimized continuous prompts for multi-task fine-tuning.
- Multi-task fine-tuning reduces overall data requirements due to shared features.
- Deploying a single model for multiple tasks consumes significantly fewer resources than individual models.
Entities
Institutions
- arXiv