菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-09
📄 Abstract - Afri-MCQA: Multimodal Cultural Question Answering for African Languages

Africa is home to over one-third of the world's languages, yet remains underrepresented in AI research. We introduce Afri-MCQA, the first Multilingual Cultural Question-Answering benchmark covering 7.5k Q&A pairs across 15 African languages from 12 countries. The benchmark offers parallel English-African language Q&A pairs across text and speech modalities and was entirely created by native speakers. Benchmarking large language models (LLMs) on Afri-MCQA shows that open-weight models perform poorly across evaluated cultures, with near-zero accuracy on open-ended VQA when queried in native language or speech. To evaluate linguistic competence, we include control experiments meant to assess this specific aspect separate from cultural knowledge, and we observe significant performance gaps between native languages and English for both text and speech. These findings underscore the need for speech-first approaches, culturally grounded pretraining, and cross-lingual cultural transfer. To support more inclusive multimodal AI development in African languages, we release our Afri-MCQA under academic license or CC BY-NC 4.0 on HuggingFace (this https URL)

顶级标签: natural language processing benchmark multi-modal
详细标签: multilingual qa cultural knowledge speech evaluation african languages llm benchmarking 或 搜索:

Afri-MCQA:面向非洲语言的多模态文化问答基准 / Afri-MCQA: Multimodal Cultural Question Answering for African Languages


1️⃣ 一句话总结

该论文创建了首个覆盖15种非洲语言的多模态文化问答基准Afri-MCQA,并发现当前大语言模型在处理这些语言的文本和语音时表现极差,凸显了开发面向非洲语言的、文化敏感的多模态AI系统的紧迫性。

源自 arXiv: 2601.05699