ArXiv TLDR

Interpretable facial dynamics as behavioral and perceptual traces of deepfakes

🐦 Tweet
2604.21760

Timothy Joseph Murphy, Jennifer Cook, Hélio Clemente José Cuve

cs.CVcs.HCcs.LG

TLDR

This paper uses interpretable facial dynamics to detect deepfakes, finding emotional expressions are key and model/human detection strategies differ.

Key contributions

  • Developed interpretable bio-behavioral facial dynamics features to detect deepfakes.
  • Found deepfakes exhibit higher-order temporal irregularities, especially during emotional expressions.
  • Showed emotive signals are systematically degraded in deepfakes, improving detection accuracy.
  • Revealed model and human detection strategies differ, offering complementary routes to deepfake identification.

Why it matters

This paper introduces an interpretable deepfake detection method. It reveals emotional expressions are a key fingerprint, showing degraded emotive signals. This work offers insights for robust, explainable, and human-aligned deepfake detection.

Original Abstract

Deepfake detection research has largely converged on deep learning approaches that, despite strong benchmark performance, offer limited insight into what distinguishes real from manipulated facial behavior. This study presents an interpretable alternative grounded in bio-behavioral features of facial dynamics and evaluates how computational detection strategies relate to human perceptual judgments. We identify core low-dimensional patterns of facial movement, from which temporal features characterizing spatiotemporal structure were derived. Traditional machine learning classifiers trained on these features achieved modest but significant above-chance deepfake classification, driven by higher-order temporal irregularities that were more pronounced in manipulated than real facial dynamics. Notably, detection was substantially more accurate for videos containing emotive expressions than those without. An emotional valence classification analysis further indicated that emotive signals are systematically degraded in deepfakes, explaining the differential impact of emotive dynamics on detection. Furthermore, we provide an additional and often overlooked dimension of explainability by assessing the relationship between model decisions and human perceptual detection. Model and human judgments converged for emotive but diverged for non-emotive videos, and even where outputs aligned, underlying detection strategies differed. These findings demonstrate that face-swapped deepfakes carry a measurable behavioral fingerprint, most salient during emotional expression. Additionally, model-human comparisons suggest that interpretable computational features and human perception may offer complementary rather than redundant routes to detection.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.