马斯特里赫特大学在AMIYA项目中的研究:通过微调和最小贝叶斯风险解码使大语言模型适应阿拉伯语方言 / Maastricht University at AMIYA: Adapting LLMs for Dialectal Arabic using Fine-tuning and MBR Decoding
1️⃣ 一句话总结
这项研究提出了一种结合微调、适配器合并和方言感知解码的方法,有效提升了大型语言模型在多种阿拉伯语方言上的生成准确性和地道性,同时保持了语义的准确性。
Large Language Models (LLMs) are becoming increasingly multilingual, supporting hundreds of languages, especially high resource ones. Unfortunately, Dialect variations are still underrepresented due to limited data and linguistic variation. In this work, we adapt a pre-trained LLM to improve dialectal performance. Specifically, we use Low Rank Adaptation (LoRA) fine-tuning on monolingual and English Dialect parallel data, adapter merging and dialect-aware MBR decoding to improve dialectal fidelity generation and translation. Experiments on Syrian, Moroccan, and Saudi Arabic show that merging and MBR improve dialectal fidelity while preserving semantic accuracy. This combination provides a compact and effective framework for robust dialectal Arabic generation.
马斯特里赫特大学在AMIYA项目中的研究:通过微调和最小贝叶斯风险解码使大语言模型适应阿拉伯语方言 / Maastricht University at AMIYA: Adapting LLMs for Dialectal Arabic using Fine-tuning and MBR Decoding
这项研究提出了一种结合微调、适配器合并和方言感知解码的方法,有效提升了大型语言模型在多种阿拉伯语方言上的生成准确性和地道性,同时保持了语义的准确性。
源自 arXiv: 2602.09703