DR-LoRA: Dynamic Rank LoRA for Fine-Tuning Mixture-of-Experts Models

📰 ArXiv cs.AI

arXiv:2601.04823v4 Announce Type: replace Abstract: Mixture-of-Experts (MoE) has become a prominent paradigm for scaling Large Language Models (LLMs). Parameter-efficient fine-tuning methods, such as LoRA, are widely adopted to adapt pretrained MoE LLMs to downstream tasks. However, existing approaches typically assign identical LoRA ranks to all expert modules, ignoring the heterogeneous specialization of pretrained experts. This uniform allocation leads to a resource mismatch: task-relevant ex

Published 2 Apr 2026
Read full paper → ← Back to News