3D garment digitisation for virtual wardrobe using a commodity depth sensor

Dongjoe Shin, Yu Chen

Research output: Chapter in Book/Report/Conference proceedingConference contribution

261 Downloads (Pure)


A practical garment digitisation should be efficient and robust to minimise the cost of processing a large volume of garments manufactured in every season. In addition, the quality of a texture map needs to be high to deliver a better user experience of VR/AR applications using garment models such as digital wardrobe or virtual fitting room. To address this, we propose a novel pipeline for fast, low-cost, and robust 3D garment digitisation with minimal human involvement. The proposed system is simply configured with a commodity RGB-D sensor (e.g. Kinect) and a rotating platform where a mannequin is placed to put on a target garment. Since a conventional reconstruction pipeline such as Kinect Fusion (KF) tends to fail to track the correct camera pose under fast rotation, we modelled the camera motion and fed this as a guidance of the ICP process in KF. The proposed method is also designed to produce a high-quality texture map by stitching the best views from a single rotation, and a modified shape from silhouettes algorithm has been developed to extract a garment model from a mannequin.
Original languageEnglish
Title of host publication2017 IEEE International Conference on Computer Vision Workshops (ICCVW)
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)978-1538610343
ISBN (Print)978-1538610350
Publication statusPublished - 23 Jan 2018
EventInternational Conference on Computer Vision: ICCV 2017 - Venice, Italy
Duration: 22 Oct 201729 Oct 2017

Publication series

NameIEEE ICCVW Proceedings Series
ISSN (Electronic)2473-9944


ConferenceInternational Conference on Computer Vision
Internet address


  • virtual try-on
  • Kinect fusion
  • image stitching
  • image reconstruction


Dive into the research topics of '3D garment digitisation for virtual wardrobe using a commodity depth sensor'. Together they form a unique fingerprint.

Cite this