菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-03
📄 Abstract - Flexi-LoRA with Input-Adaptive Ranks: Efficient Finetuning for Speech and Reasoning Tasks

Parameter-efficient fine-tuning methods like Low-Rank Adaptation (LoRA) have become essential for deploying large language models, yet their static parameter allocation remains suboptimal for inputs of varying complexity. We present Flexi-LoRA, a novel framework that dynamically adjusts LoRA ranks based on input complexity during both training and inference. Through empirical analysis across question answering, mathematical reasoning, and speech tasks, we demonstrate that maintaining consistency between training and inference dynamics is important for effective adaptation, particularly for sequential reasoning tasks. Our findings reveal that input-dependent parameter allocation achieves higher performance with fewer parameters by optimally matching rank configurations to question complexity. Furthermore, task-specific dependency on rank dynamics varies, with mathematical reasoning tasks exhibiting higher dependency than QA tasks. Successful adaptation manifests not only in correctness but also in reasoning quality and instruction adherence. Flexi-LoRA consistently outperforms static LoRA while using fewer parameters, with performance gains more pronounced on tasks requiring strict reasoning chains. Our approach realizes key benefits of mixture-of-experts frameworks through a more streamlined implementation, reducing parameter redundancy while improving model capabilities. We provide comprehensive empirical studies across diverse tasks, establishing a basis for future work in input-adaptive and efficient fine-tuning approaches.

顶级标签: llm model training audio
详细标签: parameter-efficient finetuning low-rank adaptation input-adaptive ranks reasoning tasks speech tasks 或 搜索:

Flexi-LoRA:输入自适应秩的高效微调方法,适用于语音与推理任务 / Flexi-LoRA with Input-Adaptive Ranks: Efficient Finetuning for Speech and Reasoning Tasks


1️⃣ 一句话总结

本文提出了一种名为Flexi-LoRA的新方法,它能根据输入问题的复杂程度动态调整模型微调时使用的参数数量,从而在更少参数的情况下,比传统固定参数的LoRA方法在问答、数学推理和语音任务上表现更优,尤其擅长处理需要严格推理链的任务。

源自 arXiv: 2605.01959