ArXiv TLDR

FIT: A Large-Scale Dataset for Fit-Aware Virtual Try-On

🐦 Tweet
2604.08526

Johanna Karras, Yuanhao Wang, Yingwei Li, Ira Kemelmacher-Shlizerman

cs.CVcs.GR

TLDR

Introduces FIT, a 1.13M image dataset with precise body/garment measurements, enabling fit-aware virtual try-on models for realistic clothing depiction.

Key contributions

  • Introduces FIT, a 1.13M image dataset for virtual try-on with precise body and garment measurements.
  • Overcomes data scarcity for "ill-fit" scenarios by programmatically generating 3D garments with physics simulation.
  • Uses a novel re-texturing framework to transform synthetic renderings into photorealistic images.
  • Enables training of fit-aware VTO models, setting a new state-of-the-art and robust benchmark.

Why it matters

Current virtual try-on methods ignore garment fit, leading to unrealistic results. This paper addresses this by providing the first large-scale dataset, FIT, specifically designed for fit-aware VTO. It enables the development of more accurate and realistic virtual try-on experiences, crucial for e-commerce.

Original Abstract

Given a person and a garment image, virtual try-on (VTO) aims to synthesize a realistic image of the person wearing the garment, while preserving their original pose and identity. Although recent VTO methods excel at visualizing garment appearance, they largely overlook a crucial aspect of the try-on experience: the accuracy of garment fit -- for example, depicting how an extra-large shirt looks on an extra-small person. A key obstacle is the absence of datasets that provide precise garment and body size information, particularly for "ill-fit" cases, where garments are significantly too large or too small. Consequently, current VTO methods default to generating well-fitted results regardless of the garment or person size. In this paper, we take the first steps towards solving this open problem. We introduce FIT (Fit-Inclusive Try-on), a large-scale VTO dataset comprising over 1.13M try-on image triplets accompanied by precise body and garment measurements. We overcome the challenges of data collection via a scalable synthetic strategy: (1) We programmatically generate 3D garments using GarmentCode and drape them via physics simulation to capture realistic garment fit. (2) We employ a novel re-texturing framework to transform synthetic renderings into photorealistic images while strictly preserving geometry. (3) We introduce person identity preservation into our re-texturing model to generate paired person images (same person, different garments) for supervised training. Finally, we leverage our FIT dataset to train a baseline fit-aware virtual try-on model. Our data and results set the new state-of-the-art for fit-aware virtual try-on, as well as offer a robust benchmark for future research. We will make all data and code publicly available on our project page: https://johannakarras.github.io/FIT.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.