Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks
Representing the world in 3D space provides vivid texture and depth information. However, 3D datasets currently do not match the scale of 2D datasets. There is a growing trend in representing 3D data as multi-view 2D images and using large-scale 2D models, to solve 3D tasks. In this work, we present...
Saved in:
Main Authors: | , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/179095 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-179095 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1790952024-07-17T07:05:10Z Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks Yang, Xiaofeng Liu, Fayao Lin, Guosheng School of Computer Science and Engineering Computer and Information Science CLIP NeRF Representing the world in 3D space provides vivid texture and depth information. However, 3D datasets currently do not match the scale of 2D datasets. There is a growing trend in representing 3D data as multi-view 2D images and using large-scale 2D models, to solve 3D tasks. In this work, we present the Neural Radiance Selector, a method that automatically selects the optimal 2D representations of 3D data. Instead of indiscriminately sampling multi-view 2D images, we define the optimal 2D views as those capable of reconstructing the entire 3D scene with a conditional neural radiance field. We propose two distinct methods for 3D point cloud data and 3D implicit models to achieve faster inference. We demonstrate the efficacy of our methods in various 3D tasks, including zero-shot 3D point cloud classification, 3D implicit model classification, and language-guided NeRF editing. Agency for Science, Technology and Research (A*STAR) Ministry of Education (MOE) This research is supported by the MoE AcRF Tier 2 grant (MOE-T2EP 20220-0007) and the MoE AcRF Tier 1 grant (RG14/22). This research is also supported by the Agency for Science, Technology and Research (A*STAR), Singapore under its MTCYoung Individual Research Grant (Grant No. M21K3c0130). 2024-07-17T07:05:10Z 2024-07-17T07:05:10Z 2024 Journal Article Yang, X., Liu, F. & Lin, G. (2024). Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks. Knowledge-Based Systems, 299, 112002-. https://dx.doi.org/10.1016/j.knosys.2024.112002 0950-7051 https://hdl.handle.net/10356/179095 10.1016/j.knosys.2024.112002 2-s2.0-85195220282 299 112002 en MOE-T2EP 20220-0007 RG14/22 M21K3c0130 Knowledge-Based Systems © 2024 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science CLIP NeRF |
spellingShingle |
Computer and Information Science CLIP NeRF Yang, Xiaofeng Liu, Fayao Lin, Guosheng Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks |
description |
Representing the world in 3D space provides vivid texture and depth information. However, 3D datasets currently do not match the scale of 2D datasets. There is a growing trend in representing 3D data as multi-view 2D images and using large-scale 2D models, to solve 3D tasks. In this work, we present the Neural Radiance Selector, a method that automatically selects the optimal 2D representations of 3D data. Instead of indiscriminately sampling multi-view 2D images, we define the optimal 2D views as those capable of reconstructing the entire 3D scene with a conditional neural radiance field. We propose two distinct methods for 3D point cloud data and 3D implicit models to achieve faster inference. We demonstrate the efficacy of our methods in various 3D tasks, including zero-shot 3D point cloud classification, 3D implicit model classification, and language-guided NeRF editing. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Yang, Xiaofeng Liu, Fayao Lin, Guosheng |
format |
Article |
author |
Yang, Xiaofeng Liu, Fayao Lin, Guosheng |
author_sort |
Yang, Xiaofeng |
title |
Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks |
title_short |
Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks |
title_full |
Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks |
title_fullStr |
Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks |
title_full_unstemmed |
Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks |
title_sort |
neural radiance selector: find the best 2d representations of 3d data for clip based 3d tasks |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/179095 |
_version_ |
1814047232180092928 |