Solving Composite Problems in One Inference: The MeteoRA Architecture for Scalable Integration of Knowledge Modules in Large Language Models Based on MoE
In the field of large language models, the pre-training + fine-tuning paradigm has become an important foundation for deploying various downstream applications. Within this framework, the use of low-rank adaptation (LoRA) methods for efficient fine-tuning of large model parameters (PEFT) has resulted in a large number of reusable LoRA adapters tailored for specific tasks. However, … Read more