ArXiv TLDR

Feature-Augmented Transformers for Robust AI-Text Detection Across Domains and Generators

🐦 Tweet
2605.03969

Mohamed Mady, Johannes Reschke, Björn Schuller

cs.CLcs.AI

TLDR

A new feature-augmented transformer model significantly improves robust AI-text detection across diverse domains and generators, outperforming prior methods.

Key contributions

  • Developed feature-augmented transformers for robust AI-text detection across domains.
  • Achieved 85.9% balanced accuracy on M4, surpassing zero-shot baselines by 7.22 points.
  • Identified readability and vocabulary features as key for cross-domain robustness.
  • Showed DeBERTa-v3-base with feature attention outperforms older BERT/RoBERTa models.

Why it matters

This paper significantly improves robust AI-text detection across diverse domains using feature-augmented transformers. This is crucial for reliably identifying AI content under distribution shift, combating misinformation, and maintaining trust.

Original Abstract

AI-generated text is nowadays produced at scale across domains and heterogeneous generation pipelines, making robustness to distribution shift a central requirement for supervised binary detectors. We train transformer-based detectors on HC3 PLUS and calibrate a single decision threshold by maximising balanced accuracy on held-out validation; this threshold is then kept fixed for all downstream test distributions, revealing domain- and generator-dependent error asymmetries under shift. We evaluate in-domain on HC3 PLUS, under cross-dataset transfer to the multi-domain, multi-generator M4 benchmark, and on the external AI-Text-Detection-Pile. Although base models achieve near-ceiling in-domain performance (up to 99.5% balanced accuracy), performance under shift is brittle and strongly model-dependent. Feature augmentation via attention-based linguistic feature fusion improves transfer, with our best model (DeBERTa-v3-base+FeatAttn) achieving 85.9% balanced accuracy on M4. Multi-seed experiments confirm high stability. Under the same fixed-threshold protocol, our model outperforms strong zero-shot baselines by up to +7.22 points. Category-level ablations further show that readability and vocabulary features contribute most to robustness under shift. Overall, these results demonstrate that feature augmentation and a modern DeBERTa backbone significantly outperform earlier BERT/RoBERTa models, while the fixed-threshold protocol provides a more realistic and informative assessment of practical detector robustness.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.