3D virtual try-on with human parsing

In recent years, the world saw an increase in online shopping traffic, especially during the Covid-19 pandemic, where people are stuck indoors. This led to virtual try-ons becoming increasingly popular in ecommerce platforms, because they provide a virtual fitting room space to allow shoppers to try...

Full description

Saved in:
Bibliographic Details
Main Author: Tay, Yu Xuan
Other Authors: Lin Guosheng
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2023
Subjects:
Online Access:https://hdl.handle.net/10356/166088
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-166088
record_format dspace
spelling sg-ntu-dr.10356-1660882023-04-21T15:37:14Z 3D virtual try-on with human parsing Tay, Yu Xuan Lin Guosheng School of Computer Science and Engineering gslin@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision In recent years, the world saw an increase in online shopping traffic, especially during the Covid-19 pandemic, where people are stuck indoors. This led to virtual try-ons becoming increasingly popular in ecommerce platforms, because they provide a virtual fitting room space to allow shoppers to try-on new clothing. With a set up as simple as a camera and internet connection, they can see how they appear in different styles of clothing, all because of virtual try-ons. This project aims to utilise a computer vision technique known as human parsing, which is a pixel level classification task, to perform a virtual try-on. After conducting some research on already implemented virtual try-ons, this project’s objective is to implement a virtual try-on that is computationally cheap and easy to use, while being able to give the user a good idea of how they appear in different clothing. It consists of 3 stages of the try-on pipeline, the image mask generation, the thin-plate spline warp, and the 2D to 3D reconstruction. The first stage takes in the user's image and clothing image for human parsing, and prepares the necessary semantic part of both images for the second stage. The clothing mask is then warped into the shape of the user’s body and pasted onto the user, completing the 2D fit. The last step is to convert the 2D image into a 3D model for better visualisation of the fit. The 3D reconstruction uses monocular depth estimation, a depth estimation technique with a single image, followed by a point cloud generation to create a 3D space, then a poisson reconstruction of the point clouds, to produce a 3D mesh of the try-on. The virtual try-on results show decent results at the 2D stage, but the 3D reconstruction needs sharpening. Bachelor of Engineering (Computer Engineering) 2023-04-21T05:18:20Z 2023-04-21T05:18:20Z 2023 Final Year Project (FYP) Tay, Y. X. (2023). 3D virtual try-on with human parsing. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/166088 https://hdl.handle.net/10356/166088 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
spellingShingle Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
Tay, Yu Xuan
3D virtual try-on with human parsing
description In recent years, the world saw an increase in online shopping traffic, especially during the Covid-19 pandemic, where people are stuck indoors. This led to virtual try-ons becoming increasingly popular in ecommerce platforms, because they provide a virtual fitting room space to allow shoppers to try-on new clothing. With a set up as simple as a camera and internet connection, they can see how they appear in different styles of clothing, all because of virtual try-ons. This project aims to utilise a computer vision technique known as human parsing, which is a pixel level classification task, to perform a virtual try-on. After conducting some research on already implemented virtual try-ons, this project’s objective is to implement a virtual try-on that is computationally cheap and easy to use, while being able to give the user a good idea of how they appear in different clothing. It consists of 3 stages of the try-on pipeline, the image mask generation, the thin-plate spline warp, and the 2D to 3D reconstruction. The first stage takes in the user's image and clothing image for human parsing, and prepares the necessary semantic part of both images for the second stage. The clothing mask is then warped into the shape of the user’s body and pasted onto the user, completing the 2D fit. The last step is to convert the 2D image into a 3D model for better visualisation of the fit. The 3D reconstruction uses monocular depth estimation, a depth estimation technique with a single image, followed by a point cloud generation to create a 3D space, then a poisson reconstruction of the point clouds, to produce a 3D mesh of the try-on. The virtual try-on results show decent results at the 2D stage, but the 3D reconstruction needs sharpening.
author2 Lin Guosheng
author_facet Lin Guosheng
Tay, Yu Xuan
format Final Year Project
author Tay, Yu Xuan
author_sort Tay, Yu Xuan
title 3D virtual try-on with human parsing
title_short 3D virtual try-on with human parsing
title_full 3D virtual try-on with human parsing
title_fullStr 3D virtual try-on with human parsing
title_full_unstemmed 3D virtual try-on with human parsing
title_sort 3d virtual try-on with human parsing
publisher Nanyang Technological University
publishDate 2023
url https://hdl.handle.net/10356/166088
_version_ 1764208055688364032