菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-14
📄 Abstract - What Makes Words Hard? Sakura at BEA 2026 Shared Task on Vocabulary Difficulty Prediction

We describe two types of models for vocabulary difficulty prediction: a high-accuracy black-box model, which achieved the top shared task result in the open track, and an explainable model, which outperforms a fine-tuned encoder baseline. As the black-box model, we fine-tuned an LLM using a soft-target loss function for effective application to the rating task, achieving r > 0.91. The explainable model provides insights into what impacts the difficulty of each item while maintaining a strong correlation (r > 0.77). We further analyze the results, demonstrating that the difficulty of items in the British Council's Knowledge-based Vocabulary Lists (KVL) is often affected by spelling difficulty or the construction of the test items, in addition to the genuine production difficulty of the words. We make our code available online at this https URL .

顶级标签: llm natural language processing model evaluation
详细标签: vocabulary difficulty explainability fine-tuning soft-target loss rating prediction 或 搜索:

什么让单词变难?——BEA 2026词汇难度预测共享任务中的Sakura系统 / What Makes Words Hard? Sakura at BEA 2026 Shared Task on Vocabulary Difficulty Prediction


1️⃣ 一句话总结

本文为词汇难度预测任务设计了两种模型:一个高精度的黑盒模型(通过软目标损失微调大语言模型,准确率排名第一),和一个可解释模型(揭示单词难度受拼写和题目设计影响,而不仅仅是单词本身的难度),并开源了代码。

源自 arXiv: 2605.14257