菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-25
📄 Abstract - Small Wins Big: Comparing Large Language Models and Domain Fine-Tuned Models for Sarcasm Detection in Code-Mixed Hinglish Text

Sarcasm detection in multilingual and code-mixed environments remains a challenging task for natural language processing models due to structural variations, informal expressions, and low-resource linguistic availability. This study compares four large language models, Llama 3.1, Mistral, Gemma 3, and Phi-4, with a fine-tuned DistilBERT model for sarcasm detection in code-mixed Hinglish text. The results indicate that the smaller, sequentially fine-tuned DistilBERT model achieved the highest overall accuracy of 84%, outperforming all of the LLMs in zero and few-shot set ups, using minimal LLM generated code-mixed data used for fine-tuning. These findings indicate that domain-adaptive fine-tuning of smaller transformer based models may significantly improve sarcasm detection over general LLM inference, in low-resource and data scarce settings.

顶级标签: natural language processing llm model evaluation
详细标签: sarcasm detection code-mixed text hinglish fine-tuning low-resource nlp 或 搜索:

小胜大:比较大语言模型与领域微调模型在混合印地-英语文本中的讽刺检测能力 / Small Wins Big: Comparing Large Language Models and Domain Fine-Tuned Models for Sarcasm Detection in Code-Mixed Hinglish Text


1️⃣ 一句话总结

这项研究发现,在资源有限的混合印地-英语文本讽刺检测任务中,经过针对性微调的小型模型(DistilBERT)表现优于多种通用大语言模型,表明特定领域的精细调整比模型规模本身更为关键。

源自 arXiv: 2602.21933