DANI-Net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering
Uncalibrated photometric stereo (UPS) is challenging due to the inherent ambiguity brought by the unknown light. Although the ambiguity is alleviated on non-Lambertian objects, the problem is still difficult to solve for more general objects with complex shapes introducing irregular shadows and gen...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/165854 https://cvpr2023.thecvf.com/Conferences/2023 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-165854 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1658542023-08-29T00:51:06Z DANI-Net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering Li, Zongrui Zheng, Qian Shi, Boxin Pan, Gang Jiang, Xudong School of Electrical and Electronic Engineering IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2023) Rapid-Rich Object Search (ROSE) Lab Engineering::Computer science and engineering Uncalibrated Photometric Stereo Shadow Handling Uncalibrated photometric stereo (UPS) is challenging due to the inherent ambiguity brought by the unknown light. Although the ambiguity is alleviated on non-Lambertian objects, the problem is still difficult to solve for more general objects with complex shapes introducing irregular shadows and general materials with complex reflectance like anisotropic reflectance. To exploit cues from shadow and reflectance to solve UPS and improve performance on general materials, we propose DANI-Net, an inverse rendering framework with differentiable shadow handling and anisotropic reflectance modeling. Unlike most previous methods that use non-differentiable shadow maps and assume isotropic material, our network benefits from cues of shadow and anisotropic reflectance through two differentiable paths. Experiments on multiple real-world datasets demonstrate our superior and robust performance. Submitted/Accepted version This work is supported by National Natural Science Foundation of China under Grant No. 61925603, 62136001, 62088102. 2023-08-22T08:35:33Z 2023-08-22T08:35:33Z 2023 Conference Paper Li, Z., Zheng, Q., Shi, B., Pan, G. & Jiang, X. (2023). DANI-Net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2023). https://dx.doi.org/10.1109/CVPR52729.2023.00810 https://hdl.handle.net/10356/165854 10.1109/CVPR52729.2023.00810 arXiv:2303.15101 https://cvpr2023.thecvf.com/Conferences/2023 en © 2023 The Author(s). Published by Computer Vision Foundation. This is an open-access article distributed under the terms of the Creative Commons Attribution License. The final published version of the proceedings is available on IEEE Xplore. application/pdf application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Uncalibrated Photometric Stereo Shadow Handling |
spellingShingle |
Engineering::Computer science and engineering Uncalibrated Photometric Stereo Shadow Handling Li, Zongrui Zheng, Qian Shi, Boxin Pan, Gang Jiang, Xudong DANI-Net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering |
description |
Uncalibrated photometric stereo (UPS) is challenging due to the inherent ambiguity brought by the unknown light. Although the ambiguity is alleviated on non-Lambertian objects, the problem is still difficult to solve for more
general objects with complex shapes introducing irregular shadows and general materials with complex reflectance like anisotropic reflectance. To exploit cues from shadow and reflectance to solve UPS and improve performance on
general materials, we propose DANI-Net, an inverse rendering framework with differentiable shadow handling and anisotropic reflectance modeling. Unlike most previous methods that use non-differentiable shadow maps and assume isotropic material, our network benefits from cues of shadow and anisotropic reflectance through two differentiable paths. Experiments on multiple real-world datasets demonstrate our superior and robust performance. |
author2 |
School of Electrical and Electronic Engineering |
author_facet |
School of Electrical and Electronic Engineering Li, Zongrui Zheng, Qian Shi, Boxin Pan, Gang Jiang, Xudong |
format |
Conference or Workshop Item |
author |
Li, Zongrui Zheng, Qian Shi, Boxin Pan, Gang Jiang, Xudong |
author_sort |
Li, Zongrui |
title |
DANI-Net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering |
title_short |
DANI-Net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering |
title_full |
DANI-Net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering |
title_fullStr |
DANI-Net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering |
title_full_unstemmed |
DANI-Net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering |
title_sort |
dani-net: uncalibrated photometric stereo by differentiable shadow handling, anisotropic reflectance modeling, and neural inverse rendering |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/165854 https://cvpr2023.thecvf.com/Conferences/2023 |
_version_ |
1779156282347880448 |