Delving into multi-illumination monocular depth estimation: A new dataset and method
Monocular depth prediction has received significant attention in recent years. However, the impact of illumination variations, which can shift scenes to unseen domains, has often been overlooked. To address this, we introduce the first indoor scene dataset featuring RGB-D images captured under multi...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8658 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9661 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-96612024-02-22T03:00:04Z Delving into multi-illumination monocular depth estimation: A new dataset and method LIANG, Yuan ZHANG, Zitian XIAN, Chuhua HE, Shengfeng Monocular depth prediction has received significant attention in recent years. However, the impact of illumination variations, which can shift scenes to unseen domains, has often been overlooked. To address this, we introduce the first indoor scene dataset featuring RGB-D images captured under multiple illumination conditions, allowing for a comprehensive exploration of indoor depth prediction. Additionally, we propose a novel method, MI-Transformer, which leverages global illumination understanding through large receptive fields to capture depth-attention contexts. This enables our network to overcome local window limitations and effectively mitigate the influence of changing illumination conditions. To evaluate the performance and robustness, we conduct extensive qualitative and quantitative analyses on both the proposed dataset and existing benchmarks, comparing our method with state-of-the-art approaches. The experimental results demonstrate the superiority of our method across various metrics, making it the first solution to achieve robust monocular depth estimation under diverse illumination conditions. We provide the codes, pre-trained models, and dataset openly accessible at https://github.com/ViktorLiang/midepth. 2024-01-01T08:00:00Z text https://ink.library.smu.edu.sg/sis_research/8658 info:doi/10.1109/TMM.2024.3353544 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Decoding Depth Estimation Estimation Lighting Multi-illuminations Dataset Robustness Three-dimensional displays Training Transformer-enhanced Network Transformers Databases and Information Systems OS and Networks |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Decoding Depth Estimation Estimation Lighting Multi-illuminations Dataset Robustness Three-dimensional displays Training Transformer-enhanced Network Transformers Databases and Information Systems OS and Networks |
spellingShingle |
Decoding Depth Estimation Estimation Lighting Multi-illuminations Dataset Robustness Three-dimensional displays Training Transformer-enhanced Network Transformers Databases and Information Systems OS and Networks LIANG, Yuan ZHANG, Zitian XIAN, Chuhua HE, Shengfeng Delving into multi-illumination monocular depth estimation: A new dataset and method |
description |
Monocular depth prediction has received significant attention in recent years. However, the impact of illumination variations, which can shift scenes to unseen domains, has often been overlooked. To address this, we introduce the first indoor scene dataset featuring RGB-D images captured under multiple illumination conditions, allowing for a comprehensive exploration of indoor depth prediction. Additionally, we propose a novel method, MI-Transformer, which leverages global illumination understanding through large receptive fields to capture depth-attention contexts. This enables our network to overcome local window limitations and effectively mitigate the influence of changing illumination conditions. To evaluate the performance and robustness, we conduct extensive qualitative and quantitative analyses on both the proposed dataset and existing benchmarks, comparing our method with state-of-the-art approaches. The experimental results demonstrate the superiority of our method across various metrics, making it the first solution to achieve robust monocular depth estimation under diverse illumination conditions. We provide the codes, pre-trained models, and dataset openly accessible at https://github.com/ViktorLiang/midepth. |
format |
text |
author |
LIANG, Yuan ZHANG, Zitian XIAN, Chuhua HE, Shengfeng |
author_facet |
LIANG, Yuan ZHANG, Zitian XIAN, Chuhua HE, Shengfeng |
author_sort |
LIANG, Yuan |
title |
Delving into multi-illumination monocular depth estimation: A new dataset and method |
title_short |
Delving into multi-illumination monocular depth estimation: A new dataset and method |
title_full |
Delving into multi-illumination monocular depth estimation: A new dataset and method |
title_fullStr |
Delving into multi-illumination monocular depth estimation: A new dataset and method |
title_full_unstemmed |
Delving into multi-illumination monocular depth estimation: A new dataset and method |
title_sort |
delving into multi-illumination monocular depth estimation: a new dataset and method |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2024 |
url |
https://ink.library.smu.edu.sg/sis_research/8658 |
_version_ |
1794549706624335872 |