Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks

Representing the world in 3D space provides vivid texture and depth information. However, 3D datasets currently do not match the scale of 2D datasets. There is a growing trend in representing 3D data as multi-view 2D images and using large-scale 2D models, to solve 3D tasks. In this work, we present...

Full description

Saved in:
Bibliographic Details
Main Authors: Yang, Xiaofeng, Liu, Fayao, Lin, Guosheng
Other Authors: School of Computer Science and Engineering
Format: Article
Language:English
Published: 2024
Subjects:
Online Access:https://hdl.handle.net/10356/179095
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-179095
record_format dspace
spelling sg-ntu-dr.10356-1790952024-07-17T07:05:10Z Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks Yang, Xiaofeng Liu, Fayao Lin, Guosheng School of Computer Science and Engineering Computer and Information Science CLIP NeRF Representing the world in 3D space provides vivid texture and depth information. However, 3D datasets currently do not match the scale of 2D datasets. There is a growing trend in representing 3D data as multi-view 2D images and using large-scale 2D models, to solve 3D tasks. In this work, we present the Neural Radiance Selector, a method that automatically selects the optimal 2D representations of 3D data. Instead of indiscriminately sampling multi-view 2D images, we define the optimal 2D views as those capable of reconstructing the entire 3D scene with a conditional neural radiance field. We propose two distinct methods for 3D point cloud data and 3D implicit models to achieve faster inference. We demonstrate the efficacy of our methods in various 3D tasks, including zero-shot 3D point cloud classification, 3D implicit model classification, and language-guided NeRF editing. Agency for Science, Technology and Research (A*STAR) Ministry of Education (MOE) This research is supported by the MoE AcRF Tier 2 grant (MOE-T2EP 20220-0007) and the MoE AcRF Tier 1 grant (RG14/22). This research is also supported by the Agency for Science, Technology and Research (A*STAR), Singapore under its MTCYoung Individual Research Grant (Grant No. M21K3c0130). 2024-07-17T07:05:10Z 2024-07-17T07:05:10Z 2024 Journal Article Yang, X., Liu, F. & Lin, G. (2024). Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks. Knowledge-Based Systems, 299, 112002-. https://dx.doi.org/10.1016/j.knosys.2024.112002 0950-7051 https://hdl.handle.net/10356/179095 10.1016/j.knosys.2024.112002 2-s2.0-85195220282 299 112002 en MOE-T2EP 20220-0007 RG14/22 M21K3c0130 Knowledge-Based Systems © 2024 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
CLIP
NeRF
spellingShingle Computer and Information Science
CLIP
NeRF
Yang, Xiaofeng
Liu, Fayao
Lin, Guosheng
Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks
description Representing the world in 3D space provides vivid texture and depth information. However, 3D datasets currently do not match the scale of 2D datasets. There is a growing trend in representing 3D data as multi-view 2D images and using large-scale 2D models, to solve 3D tasks. In this work, we present the Neural Radiance Selector, a method that automatically selects the optimal 2D representations of 3D data. Instead of indiscriminately sampling multi-view 2D images, we define the optimal 2D views as those capable of reconstructing the entire 3D scene with a conditional neural radiance field. We propose two distinct methods for 3D point cloud data and 3D implicit models to achieve faster inference. We demonstrate the efficacy of our methods in various 3D tasks, including zero-shot 3D point cloud classification, 3D implicit model classification, and language-guided NeRF editing.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Yang, Xiaofeng
Liu, Fayao
Lin, Guosheng
format Article
author Yang, Xiaofeng
Liu, Fayao
Lin, Guosheng
author_sort Yang, Xiaofeng
title Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks
title_short Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks
title_full Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks
title_fullStr Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks
title_full_unstemmed Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks
title_sort neural radiance selector: find the best 2d representations of 3d data for clip based 3d tasks
publishDate 2024
url https://hdl.handle.net/10356/179095
_version_ 1814047232180092928