SpSequenceNet : semantic segmentation network on 4D point clouds

Point clouds are useful in many applications like autonomous driving and robotics as they provide natural 3D information of the surrounding environments. While there are extensive research on 3D point clouds, scene understanding on 4D point clouds, a series of consecutive 3D point clouds frames...

Full description

Saved in:
Bibliographic Details
Main Authors: Shi, Hanyu, Lin, Guosheng, Wang, Hao, Hung, Tzu-Yi, Wang, Zhenhua
Other Authors: School of Computer Science and Engineering
Format: Conference or Workshop Item
Language:English
Published: 2020
Subjects:
Online Access:https://hdl.handle.net/10356/143545
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-143545
record_format dspace
spelling sg-ntu-dr.10356-1435452020-10-26T05:08:13Z SpSequenceNet : semantic segmentation network on 4D point clouds Shi, Hanyu Lin, Guosheng Wang, Hao Hung, Tzu-Yi Wang, Zhenhua School of Computer Science and Engineering IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2020 Zhejiang University of Technology Engineering::Computer science and engineering Segmentation Computer Vision Point clouds are useful in many applications like autonomous driving and robotics as they provide natural 3D information of the surrounding environments. While there are extensive research on 3D point clouds, scene understanding on 4D point clouds, a series of consecutive 3D point clouds frames, is an emerging topic and yet underinvestigated. With 4D point clouds (3D point cloud videos), robotic systems could enhance their robustness by leveraging the temporal information from previous frames. However, the existing semantic segmentation methods on 4D point clouds suffer from low precision due to the spatial and temporal information loss in their network structures. In this paper, we propose SpSequenceNet to address this problem. The network is designed based on 3D sparse convolution, and it includes two novel modules, a cross-frame global attention module and a cross-frame local interpolation module, to capture spatial and temporal information in 4D point clouds. We conduct extensive experiments on SemanticKITTI, and achieve the state-of-the-art result of 43.1% on mIoU, which is 1.5% higher than the previous best approach. Ministry of Education (MOE) National Research Foundation (NRF) Published version This work is supported by the Delta-NTU Corporate Lab with funding support from Delta Electronics Inc. and the National Research Foundation (NRF) Singapore. This work is also partly supported by the National Research Foundation Singapore under its AI Singapore Programme (Award Number: AISG-RP-2018-003), the MOE Tier-1 research grant: RG22/19 (S), and the National Natural Science Foundation of China (61802348). 2020-09-08T06:55:28Z 2020-09-08T06:55:28Z 2020 Conference Paper Shi, H., Lin, G., Wang, H., Hung, T.-Y., & Wang, Z. (2020). SpSequenceNet : semantic segmentation network on 4D point clouds. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2020. doi:10.1109/CVPR42600.2020.00463 https://hdl.handle.net/10356/143545 10.1109/CVPR42600.2020.00463 en Delta-NTU Corporate Lab AISG-RP-2018-003 RG22/19 (S) © 2020 The Author(s) (published by IEEE). This is an open-access article distributed under the terms of the Creative Commons Attribution License. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Segmentation
Computer Vision
spellingShingle Engineering::Computer science and engineering
Segmentation
Computer Vision
Shi, Hanyu
Lin, Guosheng
Wang, Hao
Hung, Tzu-Yi
Wang, Zhenhua
SpSequenceNet : semantic segmentation network on 4D point clouds
description Point clouds are useful in many applications like autonomous driving and robotics as they provide natural 3D information of the surrounding environments. While there are extensive research on 3D point clouds, scene understanding on 4D point clouds, a series of consecutive 3D point clouds frames, is an emerging topic and yet underinvestigated. With 4D point clouds (3D point cloud videos), robotic systems could enhance their robustness by leveraging the temporal information from previous frames. However, the existing semantic segmentation methods on 4D point clouds suffer from low precision due to the spatial and temporal information loss in their network structures. In this paper, we propose SpSequenceNet to address this problem. The network is designed based on 3D sparse convolution, and it includes two novel modules, a cross-frame global attention module and a cross-frame local interpolation module, to capture spatial and temporal information in 4D point clouds. We conduct extensive experiments on SemanticKITTI, and achieve the state-of-the-art result of 43.1% on mIoU, which is 1.5% higher than the previous best approach.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Shi, Hanyu
Lin, Guosheng
Wang, Hao
Hung, Tzu-Yi
Wang, Zhenhua
format Conference or Workshop Item
author Shi, Hanyu
Lin, Guosheng
Wang, Hao
Hung, Tzu-Yi
Wang, Zhenhua
author_sort Shi, Hanyu
title SpSequenceNet : semantic segmentation network on 4D point clouds
title_short SpSequenceNet : semantic segmentation network on 4D point clouds
title_full SpSequenceNet : semantic segmentation network on 4D point clouds
title_fullStr SpSequenceNet : semantic segmentation network on 4D point clouds
title_full_unstemmed SpSequenceNet : semantic segmentation network on 4D point clouds
title_sort spsequencenet : semantic segmentation network on 4d point clouds
publishDate 2020
url https://hdl.handle.net/10356/143545
_version_ 1683493878556000256