📄
Abstract - Feature-Augmented Transformers for Robust AI-Text Detection Across Domains and Generators
AI-generated text is nowadays produced at scale across domains and heterogeneous generation pipelines, making robustness to distribution shift a central requirement for supervised binary detectors. We train transformer-based detectors on HC3 PLUS and calibrate a single decision threshold by maximising balanced accuracy on held-out validation; this threshold is then kept fixed for all downstream test distributions, revealing domain- and generator-dependent error asymmetries under shift. We evaluate in-domain on HC3 PLUS, under cross-dataset transfer to the multi-domain, multi-generator M4 benchmark, and on the external AI-Text-Detection-Pile. Although base models achieve near-ceiling in-domain performance (up to 99.5% balanced accuracy), performance under shift is brittle and strongly model-dependent. Feature augmentation via attention-based linguistic feature fusion improves transfer, with our best model (DeBERTa-v3-base+FeatAttn) achieving 85.9% balanced accuracy on M4. Multi-seed experiments confirm high stability. Under the same fixed-threshold protocol, our model outperforms strong zero-shot baselines by up to +7.22 points. Category-level ablations further show that readability and vocabulary features contribute most to robustness under shift. Overall, these results demonstrate that feature augmentation and a modern DeBERTa backbone significantly outperform earlier BERT/RoBERTa models, while the fixed-threshold protocol provides a more realistic and informative assessment of practical detector robustness.
基于特征增强的Transformer模型:实现跨领域与跨生成器的鲁棒性AI文本检测 /
Feature-Augmented Transformers for Robust AI-Text Detection Across Domains and Generators
1️⃣ 一句话总结
本文提出一种通过注意力机制融合语言特征(如可读性和词汇特征)的Transformer检测器(DeBERTa-v3-base+FeatAttn),在固定阈值下实现跨领域、跨生成器的稳健AI文本检测,在M4基准上达到85.9%的平衡准确率,显著优于传统BERT/RoBERTa模型及零样本方法。