Mesh R-CNN++ for 3D Mesh generation: from single to multiple views
Inferring the 3-dimensional structure and geometry of scenes and objects from one or multiple 2-dimensional images has been one of the primary goals of image-based 3D reconstruction. In recent years, with the improved progress of deep learning techniques, and the increasing availability of large 3D...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/156477 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-156477 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1564772022-04-17T12:01:38Z Mesh R-CNN++ for 3D Mesh generation: from single to multiple views Zhang, HengKai Qian Kemao School of Computer Science and Engineering MKMQian@ntu.edu.sg Engineering::Computer science and engineering Inferring the 3-dimensional structure and geometry of scenes and objects from one or multiple 2-dimensional images has been one of the primary goals of image-based 3D reconstruction. In recent years, with the improved progress of deep learning techniques, and the increasing availability of large 3D training datasets, led to significant advances in 3D shape understanding using deep learning. Inspired by traditional multiple view geometry methods, this project proposed, Mesh R-CNN++, a multi-view deep learning shape predictor. Extensive experiments against current state-of-the-art single and multi-view deep learning shape predictors showed that Mesh R-CNN++ produces 3D models with accurate thin structures and surface details using multiple images. Bachelor of Engineering (Computer Engineering) 2022-04-17T12:01:38Z 2022-04-17T12:01:38Z 2022 Final Year Project (FYP) Zhang, H. (2022). Mesh R-CNN++ for 3D Mesh generation: from single to multiple views. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/156477 https://hdl.handle.net/10356/156477 en SCSE21-4080 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering |
spellingShingle |
Engineering::Computer science and engineering Zhang, HengKai Mesh R-CNN++ for 3D Mesh generation: from single to multiple views |
description |
Inferring the 3-dimensional structure and geometry of scenes and objects from one or multiple 2-dimensional images has been one of the primary goals of image-based 3D reconstruction. In recent years, with the improved progress of deep learning techniques, and the increasing availability of large 3D training datasets, led to significant advances in 3D shape understanding using deep learning. Inspired by traditional multiple view geometry methods, this project proposed, Mesh R-CNN++, a multi-view deep learning shape predictor. Extensive experiments against current state-of-the-art single and multi-view deep learning shape predictors showed that Mesh R-CNN++ produces 3D models with accurate thin structures and surface details using multiple images. |
author2 |
Qian Kemao |
author_facet |
Qian Kemao Zhang, HengKai |
format |
Final Year Project |
author |
Zhang, HengKai |
author_sort |
Zhang, HengKai |
title |
Mesh R-CNN++ for 3D Mesh generation: from single to multiple views |
title_short |
Mesh R-CNN++ for 3D Mesh generation: from single to multiple views |
title_full |
Mesh R-CNN++ for 3D Mesh generation: from single to multiple views |
title_fullStr |
Mesh R-CNN++ for 3D Mesh generation: from single to multiple views |
title_full_unstemmed |
Mesh R-CNN++ for 3D Mesh generation: from single to multiple views |
title_sort |
mesh r-cnn++ for 3d mesh generation: from single to multiple views |
publisher |
Nanyang Technological University |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/156477 |
_version_ |
1731235766314467328 |