3D virtual try-on with human parsing
In recent years, the world saw an increase in online shopping traffic, especially during the Covid-19 pandemic, where people are stuck indoors. This led to virtual try-ons becoming increasingly popular in ecommerce platforms, because they provide a virtual fitting room space to allow shoppers to try...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/166088 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-166088 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1660882023-04-21T15:37:14Z 3D virtual try-on with human parsing Tay, Yu Xuan Lin Guosheng School of Computer Science and Engineering gslin@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision In recent years, the world saw an increase in online shopping traffic, especially during the Covid-19 pandemic, where people are stuck indoors. This led to virtual try-ons becoming increasingly popular in ecommerce platforms, because they provide a virtual fitting room space to allow shoppers to try-on new clothing. With a set up as simple as a camera and internet connection, they can see how they appear in different styles of clothing, all because of virtual try-ons. This project aims to utilise a computer vision technique known as human parsing, which is a pixel level classification task, to perform a virtual try-on. After conducting some research on already implemented virtual try-ons, this project’s objective is to implement a virtual try-on that is computationally cheap and easy to use, while being able to give the user a good idea of how they appear in different clothing. It consists of 3 stages of the try-on pipeline, the image mask generation, the thin-plate spline warp, and the 2D to 3D reconstruction. The first stage takes in the user's image and clothing image for human parsing, and prepares the necessary semantic part of both images for the second stage. The clothing mask is then warped into the shape of the user’s body and pasted onto the user, completing the 2D fit. The last step is to convert the 2D image into a 3D model for better visualisation of the fit. The 3D reconstruction uses monocular depth estimation, a depth estimation technique with a single image, followed by a point cloud generation to create a 3D space, then a poisson reconstruction of the point clouds, to produce a 3D mesh of the try-on. The virtual try-on results show decent results at the 2D stage, but the 3D reconstruction needs sharpening. Bachelor of Engineering (Computer Engineering) 2023-04-21T05:18:20Z 2023-04-21T05:18:20Z 2023 Final Year Project (FYP) Tay, Y. X. (2023). 3D virtual try-on with human parsing. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/166088 https://hdl.handle.net/10356/166088 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision |
spellingShingle |
Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Tay, Yu Xuan 3D virtual try-on with human parsing |
description |
In recent years, the world saw an increase in online shopping traffic, especially during the Covid-19 pandemic, where people are stuck indoors. This led to virtual try-ons becoming increasingly popular in ecommerce platforms, because they provide a virtual fitting room space to allow shoppers to try-on new clothing. With a set up as simple as a camera and internet connection, they can see how they appear in different styles of clothing, all because of virtual try-ons. This project aims to utilise a computer vision technique known as human parsing, which is a pixel level classification task, to perform a virtual try-on. After conducting some research on already implemented virtual try-ons, this project’s objective is to implement a virtual try-on that is computationally cheap and easy to use, while being able to give the user a good idea of how they appear in different clothing. It consists of 3 stages of the try-on pipeline, the image mask generation, the thin-plate spline warp, and the 2D to 3D reconstruction. The first stage takes in the user's image and clothing image for human parsing, and prepares the necessary semantic part of both images for the second stage. The clothing mask is then warped into the shape of the user’s body and pasted onto the user, completing the 2D fit. The last step is to convert the 2D image into a 3D model for better visualisation of the fit. The 3D reconstruction uses monocular depth estimation, a depth estimation technique with a single image, followed by a point cloud generation to create a 3D space, then a poisson reconstruction of the point clouds, to produce a 3D mesh of the try-on. The virtual try-on results show decent results at the 2D stage, but the 3D reconstruction needs sharpening. |
author2 |
Lin Guosheng |
author_facet |
Lin Guosheng Tay, Yu Xuan |
format |
Final Year Project |
author |
Tay, Yu Xuan |
author_sort |
Tay, Yu Xuan |
title |
3D virtual try-on with human parsing |
title_short |
3D virtual try-on with human parsing |
title_full |
3D virtual try-on with human parsing |
title_fullStr |
3D virtual try-on with human parsing |
title_full_unstemmed |
3D virtual try-on with human parsing |
title_sort |
3d virtual try-on with human parsing |
publisher |
Nanyang Technological University |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/166088 |
_version_ |
1764208055688364032 |