Vision-based 3D information modeling and applications

We can infer the 3D structure of our surroundings simply by looking. It is long hoped that imaging devices can mirror such an ability, which is crucial to many computer vision tasks. This thesis is about our work on developing algorithms, as well as utilizing novel optical devices, in particular lig...

Full description

Saved in:
Bibliographic Details
Main Author: Ni, Yun
Other Authors: Lap-Pui Chau
Format: Thesis-Doctor of Philosophy
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/145895
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-145895
record_format dspace
spelling sg-ntu-dr.10356-1458952023-07-04T17:40:17Z Vision-based 3D information modeling and applications Ni, Yun Lap-Pui Chau School of Electrical and Electronic Engineering elpchau@ntu.edu.sg Engineering::Electrical and electronic engineering We can infer the 3D structure of our surroundings simply by looking. It is long hoped that imaging devices can mirror such an ability, which is crucial to many computer vision tasks. This thesis is about our work on developing algorithms, as well as utilizing novel optical devices, in particular light-field cameras, to infer 3D information without active illuminations, and how such information can be used in various practical applications: • We develop an algorithm to synthesize novel views. When we shift to a different viewpoint, certain scene points not captured in the input image will be revealed. We first infer the colors of these points, based on 3D plane notations, and then use the expanded scene points to generate the target image. • We propose using depths calculated from light field (LF) data to remove reflections. The depths are used to roughly identify background and reflection scene points. We then reconstruct the background and the reflection layers using scene points identified. • Finally, we utilize depth information to help identify different kinds of materials. Given images captured using a multi-camera cell phone, we estimate a depth probability map. The estimated depth probability map, together with one of the color images, are then inputted into a trained neural network to determine the material type. The first method utilizes 3D plane notations, while the rest use depths recovered from LF data or stereo images. 3D information is crucial in accomplishing these tasks. Doctor of Philosophy 2021-01-14T01:18:06Z 2021-01-14T01:18:06Z 2021 Thesis-Doctor of Philosophy Ni, Y. (2021). Vision-based 3D information modeling and applications. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/145895 https://hdl.handle.net/10356/145895 10.32657/10356/145895 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
spellingShingle Engineering::Electrical and electronic engineering
Ni, Yun
Vision-based 3D information modeling and applications
description We can infer the 3D structure of our surroundings simply by looking. It is long hoped that imaging devices can mirror such an ability, which is crucial to many computer vision tasks. This thesis is about our work on developing algorithms, as well as utilizing novel optical devices, in particular light-field cameras, to infer 3D information without active illuminations, and how such information can be used in various practical applications: • We develop an algorithm to synthesize novel views. When we shift to a different viewpoint, certain scene points not captured in the input image will be revealed. We first infer the colors of these points, based on 3D plane notations, and then use the expanded scene points to generate the target image. • We propose using depths calculated from light field (LF) data to remove reflections. The depths are used to roughly identify background and reflection scene points. We then reconstruct the background and the reflection layers using scene points identified. • Finally, we utilize depth information to help identify different kinds of materials. Given images captured using a multi-camera cell phone, we estimate a depth probability map. The estimated depth probability map, together with one of the color images, are then inputted into a trained neural network to determine the material type. The first method utilizes 3D plane notations, while the rest use depths recovered from LF data or stereo images. 3D information is crucial in accomplishing these tasks.
author2 Lap-Pui Chau
author_facet Lap-Pui Chau
Ni, Yun
format Thesis-Doctor of Philosophy
author Ni, Yun
author_sort Ni, Yun
title Vision-based 3D information modeling and applications
title_short Vision-based 3D information modeling and applications
title_full Vision-based 3D information modeling and applications
title_fullStr Vision-based 3D information modeling and applications
title_full_unstemmed Vision-based 3D information modeling and applications
title_sort vision-based 3d information modeling and applications
publisher Nanyang Technological University
publishDate 2021
url https://hdl.handle.net/10356/145895
_version_ 1772826532957388800