ArXiv TLDR

Generative Texture Diversification of 3D Pedestrians for Robust Autonomous Driving Perception

🐦 Tweet
2605.13755

Arka Bhowmick, Enes Ozeren, Ahmed Abdullah, Oliver Wasenmuller

cs.CV

TLDR

Generates diverse 3D pedestrian textures using StyleGAN2 for synthetic data, enhancing autonomous driving perception robustness.

Key contributions

  • Proposes StyleGAN2-based method for scalable 3D pedestrian texture diversification.
  • Synthesizes diverse facial textures and appearances from a single 3D base asset.
  • Improves 2D object detection robustness using controlled synthetic data mixing.
  • Highlights 3D perception models' sensitivity to geometric domain gaps.

Why it matters

Autonomous driving requires vast, diverse data, which is expensive to collect. This paper provides a scalable generative AI solution for synthetic pedestrian data, improving 2D perception robustness. It also highlights challenges for 3D perception with synthetic data.

Original Abstract

In recent years, autonomous driving has significantly in creased the demand for high-quality data to train 2D and 3D perception models for safety-critical scenarios. Real world datasets struggle to meet this demand as require ments continuously evolve and large-scale annotated data collection remains costly and time-consuming making syn thetic data a scalable, practical and controllable alterna tive. Pedestrian detection is among the most safety-critical tasks in autonomous driving. In this paper, we propose a simple yet effective method for scaling variability in 3D pedestrian assets for synthetic scene generation. Starting from a single 3D base asset, we generate multiple distinct pedestrian instances by synthesizing diverse facial textures and identity-level appearance variations using StyleGAN2 and automatically mapping them onto 3D meshes. This ap proach enables scalable appearance-level asset diversifica tion without requiring the design of new geometries for each instance. Using the assets, we construct synthetic datasets and study the impact of mixing real and synthetic data for RGB-based object detection. Through complementary ex periments, we analyze geometry-driven distribution shifts in point cloud perception for 3D object detection. Our findings demonstrate that controlled synthetic diversifica tion improves robustness in 2D detection while revealing the sensitivity of 3D perception models to geometric domain gaps. Overall, this work highlights how generative AI en ables scalable, simulation-ready pedestrian diversification through controlled facial texture synthesis, along with the benefits and limitations of cross-domain training strategies in autonomous driving pipelines.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.