Delving into multi-illumination monocular depth estimation: A new dataset and method

Monocular depth prediction has received significant attention in recent years. However, the impact of illumination variations, which can shift scenes to unseen domains, has often been overlooked. To address this, we introduce the first indoor scene dataset featuring RGB-D images captured under multi...

Full description

Saved in:
Bibliographic Details
Main Authors: LIANG, Yuan, ZHANG, Zitian, XIAN, Chuhua, HE, Shengfeng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2024
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8658
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9661
record_format dspace
spelling sg-smu-ink.sis_research-96612024-02-22T03:00:04Z Delving into multi-illumination monocular depth estimation: A new dataset and method LIANG, Yuan ZHANG, Zitian XIAN, Chuhua HE, Shengfeng Monocular depth prediction has received significant attention in recent years. However, the impact of illumination variations, which can shift scenes to unseen domains, has often been overlooked. To address this, we introduce the first indoor scene dataset featuring RGB-D images captured under multiple illumination conditions, allowing for a comprehensive exploration of indoor depth prediction. Additionally, we propose a novel method, MI-Transformer, which leverages global illumination understanding through large receptive fields to capture depth-attention contexts. This enables our network to overcome local window limitations and effectively mitigate the influence of changing illumination conditions. To evaluate the performance and robustness, we conduct extensive qualitative and quantitative analyses on both the proposed dataset and existing benchmarks, comparing our method with state-of-the-art approaches. The experimental results demonstrate the superiority of our method across various metrics, making it the first solution to achieve robust monocular depth estimation under diverse illumination conditions. We provide the codes, pre-trained models, and dataset openly accessible at https://github.com/ViktorLiang/midepth. 2024-01-01T08:00:00Z text https://ink.library.smu.edu.sg/sis_research/8658 info:doi/10.1109/TMM.2024.3353544 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Decoding Depth Estimation Estimation Lighting Multi-illuminations Dataset Robustness Three-dimensional displays Training Transformer-enhanced Network Transformers Databases and Information Systems OS and Networks
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Decoding
Depth Estimation
Estimation
Lighting
Multi-illuminations Dataset
Robustness
Three-dimensional displays
Training
Transformer-enhanced Network
Transformers
Databases and Information Systems
OS and Networks
spellingShingle Decoding
Depth Estimation
Estimation
Lighting
Multi-illuminations Dataset
Robustness
Three-dimensional displays
Training
Transformer-enhanced Network
Transformers
Databases and Information Systems
OS and Networks
LIANG, Yuan
ZHANG, Zitian
XIAN, Chuhua
HE, Shengfeng
Delving into multi-illumination monocular depth estimation: A new dataset and method
description Monocular depth prediction has received significant attention in recent years. However, the impact of illumination variations, which can shift scenes to unseen domains, has often been overlooked. To address this, we introduce the first indoor scene dataset featuring RGB-D images captured under multiple illumination conditions, allowing for a comprehensive exploration of indoor depth prediction. Additionally, we propose a novel method, MI-Transformer, which leverages global illumination understanding through large receptive fields to capture depth-attention contexts. This enables our network to overcome local window limitations and effectively mitigate the influence of changing illumination conditions. To evaluate the performance and robustness, we conduct extensive qualitative and quantitative analyses on both the proposed dataset and existing benchmarks, comparing our method with state-of-the-art approaches. The experimental results demonstrate the superiority of our method across various metrics, making it the first solution to achieve robust monocular depth estimation under diverse illumination conditions. We provide the codes, pre-trained models, and dataset openly accessible at https://github.com/ViktorLiang/midepth.
format text
author LIANG, Yuan
ZHANG, Zitian
XIAN, Chuhua
HE, Shengfeng
author_facet LIANG, Yuan
ZHANG, Zitian
XIAN, Chuhua
HE, Shengfeng
author_sort LIANG, Yuan
title Delving into multi-illumination monocular depth estimation: A new dataset and method
title_short Delving into multi-illumination monocular depth estimation: A new dataset and method
title_full Delving into multi-illumination monocular depth estimation: A new dataset and method
title_fullStr Delving into multi-illumination monocular depth estimation: A new dataset and method
title_full_unstemmed Delving into multi-illumination monocular depth estimation: A new dataset and method
title_sort delving into multi-illumination monocular depth estimation: a new dataset and method
publisher Institutional Knowledge at Singapore Management University
publishDate 2024
url https://ink.library.smu.edu.sg/sis_research/8658
_version_ 1794549706624335872