Vision-based 3D information modeling and applications
We can infer the 3D structure of our surroundings simply by looking. It is long hoped that imaging devices can mirror such an ability, which is crucial to many computer vision tasks. This thesis is about our work on developing algorithms, as well as utilizing novel optical devices, in particular lig...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Doctor of Philosophy |
Language: | English |
Published: |
Nanyang Technological University
2021
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/145895 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-145895 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1458952023-07-04T17:40:17Z Vision-based 3D information modeling and applications Ni, Yun Lap-Pui Chau School of Electrical and Electronic Engineering elpchau@ntu.edu.sg Engineering::Electrical and electronic engineering We can infer the 3D structure of our surroundings simply by looking. It is long hoped that imaging devices can mirror such an ability, which is crucial to many computer vision tasks. This thesis is about our work on developing algorithms, as well as utilizing novel optical devices, in particular light-field cameras, to infer 3D information without active illuminations, and how such information can be used in various practical applications: • We develop an algorithm to synthesize novel views. When we shift to a different viewpoint, certain scene points not captured in the input image will be revealed. We first infer the colors of these points, based on 3D plane notations, and then use the expanded scene points to generate the target image. • We propose using depths calculated from light field (LF) data to remove reflections. The depths are used to roughly identify background and reflection scene points. We then reconstruct the background and the reflection layers using scene points identified. • Finally, we utilize depth information to help identify different kinds of materials. Given images captured using a multi-camera cell phone, we estimate a depth probability map. The estimated depth probability map, together with one of the color images, are then inputted into a trained neural network to determine the material type. The first method utilizes 3D plane notations, while the rest use depths recovered from LF data or stereo images. 3D information is crucial in accomplishing these tasks. Doctor of Philosophy 2021-01-14T01:18:06Z 2021-01-14T01:18:06Z 2021 Thesis-Doctor of Philosophy Ni, Y. (2021). Vision-based 3D information modeling and applications. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/145895 https://hdl.handle.net/10356/145895 10.32657/10356/145895 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering |
spellingShingle |
Engineering::Electrical and electronic engineering Ni, Yun Vision-based 3D information modeling and applications |
description |
We can infer the 3D structure of our surroundings simply by looking. It is long hoped that imaging devices can mirror such an ability, which is crucial to many computer vision tasks. This thesis is about our work on developing algorithms, as well as utilizing novel optical devices, in particular light-field cameras, to infer 3D information without active illuminations, and how such information can be used in various practical applications:
• We develop an algorithm to synthesize novel views. When we shift to a different viewpoint, certain scene points not captured in the input image will be revealed. We first infer the colors of these points, based on 3D plane notations, and then use the expanded scene points to generate the target image.
• We propose using depths calculated from light field (LF) data to remove reflections. The depths are used to roughly identify background and reflection scene points. We then reconstruct the background and the reflection layers using scene points identified.
• Finally, we utilize depth information to help identify different kinds of materials. Given images captured using a multi-camera cell phone, we estimate a depth probability map. The estimated depth probability map, together with one of the color images, are then inputted into a trained neural network to determine the material type.
The first method utilizes 3D plane notations, while the rest use depths recovered from LF data or stereo images. 3D information is crucial in accomplishing these tasks. |
author2 |
Lap-Pui Chau |
author_facet |
Lap-Pui Chau Ni, Yun |
format |
Thesis-Doctor of Philosophy |
author |
Ni, Yun |
author_sort |
Ni, Yun |
title |
Vision-based 3D information modeling and applications |
title_short |
Vision-based 3D information modeling and applications |
title_full |
Vision-based 3D information modeling and applications |
title_fullStr |
Vision-based 3D information modeling and applications |
title_full_unstemmed |
Vision-based 3D information modeling and applications |
title_sort |
vision-based 3d information modeling and applications |
publisher |
Nanyang Technological University |
publishDate |
2021 |
url |
https://hdl.handle.net/10356/145895 |
_version_ |
1772826532957388800 |