Diffuse3D: Wide-angle 3D photography via bilateral diffusion

This paper aims to resolve the challenging problem of wide-angle novel view synthesis from a single image, a.k.a. wide-angle 3D photography. Existing approaches rely on local context and treat them equally to inpaint occluded RGB and depth regions, which fail to deal with large-region occlusion (i.e...

Full description

Saved in:
Bibliographic Details
Main Authors: JIANG, Yutao, ZHOU, Yang, LIANG, Yuan, LIU, Wenxi, JIAO, Jianbo, QUAN, Yuhui, HE, Shengfeng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8558
https://ink.library.smu.edu.sg/context/sis_research/article/9561/viewcontent/Diffuse3D_Wide_Angle_3D_Photography_via_Bilateral_Diffusion_ICCV_2023_oa.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9561
record_format dspace
spelling sg-smu-ink.sis_research-95612024-04-15T05:43:49Z Diffuse3D: Wide-angle 3D photography via bilateral diffusion JIANG, Yutao ZHOU, Yang LIANG, Yuan LIU, Wenxi JIAO, Jianbo QUAN, Yuhui HE, Shengfeng This paper aims to resolve the challenging problem of wide-angle novel view synthesis from a single image, a.k.a. wide-angle 3D photography. Existing approaches rely on local context and treat them equally to inpaint occluded RGB and depth regions, which fail to deal with large-region occlusion (i.e., observing from an extreme angle) and foreground layers might blend into background inpainting. To address the above issues, we propose Diffuse3D which employs a pre-trained diffusion model for global synthesis, while amending the model to activate depth-aware inference. Our key insight is to alter the convolution mechanism in the denoising process. We inject depth information into the denoising convolution operation with bilateral kernels, i.e., a depth kernel and a spatial kernel, to consider layered correlations among pixels. In this way, foreground regions are overlooked in background inpainting and only pixels close in depth are leveraged. On the other hand, we propose a global-local balancing approach to maximize both contextual understandings. Extensive experiments demonstrate that our approach outperforms state-of-the-art methods in novel view synthesis, especially in wide-angle scenarios. More importantly, our method does not require any training and is a plug-and-play module that can be integrated with any diffusion model. Our code can be found at https://github.com/yutaojiang1/Diffuse3D. 2023-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8558 info:doi/10.1109/ICCV51070.2023.00826 https://ink.library.smu.edu.sg/context/sis_research/article/9561/viewcontent/Diffuse3D_Wide_Angle_3D_Photography_via_Bilateral_Diffusion_ICCV_2023_oa.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Diffusion model wide-angle 3D photography Diffuse3D Computer Sciences Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Diffusion model
wide-angle 3D photography
Diffuse3D
Computer Sciences
Graphics and Human Computer Interfaces
spellingShingle Diffusion model
wide-angle 3D photography
Diffuse3D
Computer Sciences
Graphics and Human Computer Interfaces
JIANG, Yutao
ZHOU, Yang
LIANG, Yuan
LIU, Wenxi
JIAO, Jianbo
QUAN, Yuhui
HE, Shengfeng
Diffuse3D: Wide-angle 3D photography via bilateral diffusion
description This paper aims to resolve the challenging problem of wide-angle novel view synthesis from a single image, a.k.a. wide-angle 3D photography. Existing approaches rely on local context and treat them equally to inpaint occluded RGB and depth regions, which fail to deal with large-region occlusion (i.e., observing from an extreme angle) and foreground layers might blend into background inpainting. To address the above issues, we propose Diffuse3D which employs a pre-trained diffusion model for global synthesis, while amending the model to activate depth-aware inference. Our key insight is to alter the convolution mechanism in the denoising process. We inject depth information into the denoising convolution operation with bilateral kernels, i.e., a depth kernel and a spatial kernel, to consider layered correlations among pixels. In this way, foreground regions are overlooked in background inpainting and only pixels close in depth are leveraged. On the other hand, we propose a global-local balancing approach to maximize both contextual understandings. Extensive experiments demonstrate that our approach outperforms state-of-the-art methods in novel view synthesis, especially in wide-angle scenarios. More importantly, our method does not require any training and is a plug-and-play module that can be integrated with any diffusion model. Our code can be found at https://github.com/yutaojiang1/Diffuse3D.
format text
author JIANG, Yutao
ZHOU, Yang
LIANG, Yuan
LIU, Wenxi
JIAO, Jianbo
QUAN, Yuhui
HE, Shengfeng
author_facet JIANG, Yutao
ZHOU, Yang
LIANG, Yuan
LIU, Wenxi
JIAO, Jianbo
QUAN, Yuhui
HE, Shengfeng
author_sort JIANG, Yutao
title Diffuse3D: Wide-angle 3D photography via bilateral diffusion
title_short Diffuse3D: Wide-angle 3D photography via bilateral diffusion
title_full Diffuse3D: Wide-angle 3D photography via bilateral diffusion
title_fullStr Diffuse3D: Wide-angle 3D photography via bilateral diffusion
title_full_unstemmed Diffuse3D: Wide-angle 3D photography via bilateral diffusion
title_sort diffuse3d: wide-angle 3d photography via bilateral diffusion
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8558
https://ink.library.smu.edu.sg/context/sis_research/article/9561/viewcontent/Diffuse3D_Wide_Angle_3D_Photography_via_Bilateral_Diffusion_ICCV_2023_oa.pdf
_version_ 1814047469704577024