StyleID: A Perception-Aware Dataset and Metric for Stylization-Agnostic Facial Identity Recognition
Kwan Yun, Changmin Lee, Ayeong Jeong, Youngseo Kim, Seungmi Lee + 1 more
TLDR
StyleID introduces a human perception-aware dataset and framework to improve facial identity recognition robustness across diverse artistic stylizations.
Key contributions
- Introduces StyleID, a human perception-aware dataset for facial identity under stylization.
- StyleBench-H captures human judgments on same-different verification across styles.
- StyleBench-S provides supervision from psychometric recognition-strength curves.
- Fine-tuned models show higher correlation with human judgments and improved robustness.
Why it matters
Existing identity encoders struggle with stylized faces, limiting their use in creative applications. StyleID addresses this by providing a human perception-aware framework, leading to models that accurately recognize identity across various artistic styles. This significantly enhances robustness for real-world stylized portraits.
Original Abstract
Creative face stylization aims to render portraits in diverse visual idioms such as cartoons, sketches, and paintings while retaining recognizable identity. However, current identity encoders, which are typically trained and calibrated on natural photographs, exhibit severe brittleness under stylization. They often mistake changes in texture or color palette for identity drift or fail to detect geometric exaggerations. This reveals the lack of a style-agnostic framework to evaluate and supervise identity consistency across varying styles and strengths. To address this gap, we introduce StyleID, a human perception-aware dataset and evaluation framework for facial identity under stylization. StyleID comprises two datasets: (i) StyleBench-H, a benchmark that captures human same-different verification judgments across diffusion- and flow-matching-based stylization at multiple style strengths, and (ii) StyleBench-S, a supervision set derived from psychometric recognition-strength curves obtained through controlled two-alternative forced-choice (2AFC) experiments. Leveraging StyleBench-S, we fine-tune existing semantic encoders to align their similarity orderings with human perception across styles and strengths. Experiments demonstrate that our calibrated models yield significantly higher correlation with human judgments and enhanced robustness for out-of-domain, artist drawn portraits. All of our datasets, code, and pretrained models are publicly available at https://kwanyun.github.io/StyleID_page/
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.