ChatPaper.aiChatPaper

Garments2Look:一个支持服装配饰高保真全身试穿的多参考数据集

Garments2Look: A Multi-Reference Dataset for High-Fidelity Outfit-Level Virtual Try-On with Clothing and Accessories

March 14, 2026
作者: Junyao Hu, Zhongwei Cheng, Waikeung Wong, Xingxing Zou
cs.AI

摘要

虚拟试衣技术虽已实现单件服装的可视化,但现实时尚场景以包含多件服装、配饰、细粒度品类、叠穿搭配及多样风格的整体造型为核心,这仍是当前虚拟试衣系统的盲区。现有数据集存在品类局限且缺乏造型多样性。我们推出首个面向整体造型虚拟试衣的大规模多模态数据集Garments2Look,包含8万组多服装至单造型配对,涵盖40个主要品类与300余个细分子类别。每组数据包含含3-12张参考服装图像(平均4.48张)的完整造型、模特上身效果图,以及详细的单品描述与试穿文本标注。为平衡真实性与多样性,我们提出合成流水线:先启发式构建造型清单再生成试穿效果,全程经过严格自动过滤与人工校验以确保数据质量。通过适配现有顶尖虚拟试衣方法与通用图像编辑模型建立基准测试,发现当前方法难以无缝试穿完整造型,也无法准确推断叠穿逻辑与搭配风格,导致错位与伪影问题。
English
Virtual try-on (VTON) has advanced single-garment visualization, yet real-world fashion centers on full outfits with multiple garments, accessories, fine-grained categories, layering, and diverse styling, remaining beyond current VTON systems. Existing datasets are category-limited and lack outfit diversity. We introduce Garments2Look, the first large-scale multimodal dataset for outfit-level VTON, comprising 80K many-garments-to-one-look pairs across 40 major categories and 300+ fine-grained subcategories. Each pair includes an outfit with 3-12 reference garment images (Average 4.48), a model image wearing the outfit, and detailed item and try-on textual annotations. To balance authenticity and diversity, we propose a synthesis pipeline. It involves heuristically constructing outfit lists before generating try-on results, with the entire process subjected to strict automated filtering and human validation to ensure data quality. To probe task difficulty, we adapt SOTA VTON methods and general-purpose image editing models to establish baselines. Results show current methods struggle to try on complete outfits seamlessly and to infer correct layering and styling, leading to misalignment and artifacts.
PDF23March 18, 2026