A generic method of wearable items virtual try-on

Pengpeng Hu, Nastaran Nourbakhsh Kaashki, Jing Tian, Stephan Sturges, Vasile Dadarlat, Adrian Munteanu

Research output: Contribution to journalArticle

2 Citations (Scopus)

Abstract

Virtual try-on synthesizes garments for the target bodies in 2D/3D domains. Even though existing virtual try-on methods
focus on redressing garments, the virtual try-on hair, shoes and wearable accessories are still under-reached. In this
paper, we present the first general method for virtual try-ons that is fully automatic and suitable for many items including
garments, hair, shoes, watches, necklaces, hats, and so on. Starting with the pre-defined wearable items on a reference
human body model, an automatic method is proposed to deform the reference body mesh to fit a target body for
obtaining dense triangle correspondences. Then, an improved fit metric is used to represent the interaction between
wearable items and the body. For the next step, with the help of triangle correspondences and the fit metric, the
wearable items can be fast and efficiently inferred by the shape and posture of the targeted body. Extensive experimental
results show that, besides automation and efficiency, the proposed method can be easily extended to implement the
dynamic try-on by applying rigging and importing motion capture data, being able to handle both tight and loose
garments, and even multi-layer clothing.
Original languageEnglish
Pages (from-to)2161-2174
Number of pages14
JournalTextile research journal
Volume90
Issue number19-20
DOIs
Publication statusPublished - Mar 2020

Fingerprint

Dive into the research topics of 'A generic method of wearable items virtual try-on'. Together they form a unique fingerprint.

Cite this