菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-05
📄 Abstract - Feature-Augmented Transformers for Robust AI-Text Detection Across Domains and Generators

AI-generated text is nowadays produced at scale across domains and heterogeneous generation pipelines, making robustness to distribution shift a central requirement for supervised binary detectors. We train transformer-based detectors on HC3 PLUS and calibrate a single decision threshold by maximising balanced accuracy on held-out validation; this threshold is then kept fixed for all downstream test distributions, revealing domain- and generator-dependent error asymmetries under shift. We evaluate in-domain on HC3 PLUS, under cross-dataset transfer to the multi-domain, multi-generator M4 benchmark, and on the external AI-Text-Detection-Pile. Although base models achieve near-ceiling in-domain performance (up to 99.5% balanced accuracy), performance under shift is brittle and strongly model-dependent. Feature augmentation via attention-based linguistic feature fusion improves transfer, with our best model (DeBERTa-v3-base+FeatAttn) achieving 85.9% balanced accuracy on M4. Multi-seed experiments confirm high stability. Under the same fixed-threshold protocol, our model outperforms strong zero-shot baselines by up to +7.22 points. Category-level ablations further show that readability and vocabulary features contribute most to robustness under shift. Overall, these results demonstrate that feature augmentation and a modern DeBERTa backbone significantly outperform earlier BERT/RoBERTa models, while the fixed-threshold protocol provides a more realistic and informative assessment of practical detector robustness.

顶级标签: machine learning natural language processing
详细标签: ai-generated text detection transformer feature augmentation distribution shift benchmark 或 搜索:

基于特征增强的Transformer模型:实现跨领域与跨生成器的鲁棒性AI文本检测 / Feature-Augmented Transformers for Robust AI-Text Detection Across Domains and Generators


1️⃣ 一句话总结

本文提出一种通过注意力机制融合语言特征(如可读性和词汇特征)的Transformer检测器(DeBERTa-v3-base+FeatAttn),在固定阈值下实现跨领域、跨生成器的稳健AI文本检测,在M4基准上达到85.9%的平衡准确率,显著优于传统BERT/RoBERTa模型及零样本方法。

源自 arXiv: 2605.03969