Residential College | false |
Status | 已發表Published |
FPS-Net: A convolutional fusion network for large-scale LiDAR point cloud segmentation | |
Xiao, Aoran1; Yang, Xiaofei2; Lu, Shijian1; Guan, Dayan1; Huang, Jiaxing1 | |
2021-06-01 | |
Source Publication | ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING |
ISSN | 0924-2716 |
Volume | 176Pages:237-249 |
Abstract | Scene understanding based on LiDAR point cloud is an essential task for autonomous cars to drive safely, which often employs spherical projection to map 3D point cloud into multi-channel 2D images for semantic segmentation. Most existing methods simply stack different point attributes/modalities (e.g. coordinates, intensity, depth, etc.) as image channels to increase information capacity, but ignore distinct characteristics of point attributes in different image channels. We design FPS-Net, a convolutional fusion network that exploits the uniqueness and discrepancy among the projected image channels for optimal point cloud segmentation. FPS-Net adopts an encoder-decoder structure. Instead of simply stacking multiple channel images as a single input, we group them into different modalities to first learn modality-specific features separately and then map the learnt features into a common high-dimensional feature space for pixel-level fusion and learning. Specifically, we design a residual dense block with multiple receptive fields as a building block in encoder which preserves detailed information in each modality and learns hierarchical modality-specific and fused features effectively. In the FPS-Net decoder, we use a recurrent convolution block likewise to hierarchically decode fused features into output space for pixel-level classification. Extensive experiments conducted on two widely adopted point cloud datasets show that FPS-Net achieves superior semantic segmentation as compared with state-of-the-art projection-based methods. Specifically, FPS-Net outperforms the state-of-the-art in both accuracy (4.9% higher than RangeNet++ and 2.8% higher than PolarNet in mIoU) and computation speed (15.0 FPS faster than SqueezeSegV3) for SemanticKITTI benchmark. For KITTI benchmark, FPS-Net achieves significant accuracy improvement (12.6% higher than RangeNet++ in mIoU) with comparable computation speed. In addition, the proposed modality fusion idea is compatible with typical projection-based methods and can be incorporated into them with consistent performance improvement. |
Keyword | Autonomous Driving Lidar Point Cloud Scene Understanding Semantic Segmentation Spherical Projection |
DOI | 10.1016/j.isprsjprs.2021.04.011 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Physical Geography ; Geology ; Remote Sensing ; Imaging Science & Photographic Technology |
WOS Subject | Geography, Physical ; Geosciences, Multidisciplinary ; Remote Sensing ; Imaging Science & Photographic Technology |
WOS ID | WOS:000655474600018 |
Publisher | ELSEVIER, RADARWEG 29, 1043 NX AMSTERDAM, NETHERLANDS |
Scopus ID | 2-s2.0-85105590610 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | University of Macau |
Corresponding Author | Lu, Shijian |
Affiliation | 1.Singtel Cognitive and Artificial Intelligence Lab for Enterprises, Nanyang Technological University, Singapore, 50 Nanyang Avenue, 639798, Singapore 2.University of Macau, Avenida da Universidade Taipa, 999078, Macao |
Recommended Citation GB/T 7714 | Xiao, Aoran,Yang, Xiaofei,Lu, Shijian,et al. FPS-Net: A convolutional fusion network for large-scale LiDAR point cloud segmentation[J]. ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2021, 176, 237-249. |
APA | Xiao, Aoran., Yang, Xiaofei., Lu, Shijian., Guan, Dayan., & Huang, Jiaxing (2021). FPS-Net: A convolutional fusion network for large-scale LiDAR point cloud segmentation. ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 176, 237-249. |
MLA | Xiao, Aoran,et al."FPS-Net: A convolutional fusion network for large-scale LiDAR point cloud segmentation".ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING 176(2021):237-249. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment