Adaptive Capacity Allocation for Vision Language Action Fine-tuning
This paper introduces LoRA-SP, a rank-adaptive fine-tuning method that dynamically allocates parameter capacity using a router and energy-based selection to overcome the limitations of fixed-rank LoRA in Vision-Language-Action models, thereby achieving superior multi-task generalization and efficiency on real robots.