Residential College | false |
Status | 已發表Published |
HPPLO-Net: Unsupervised LiDAR Odometry Using a Hierarchical Point-to-Plane Solver | |
Zhou,Beibei1; Tu,Yiming1; Jin,Zhong1; Xu,Chengzhong2; Kong,Hui2 | |
2023 | |
Source Publication | IEEE Transactions on Intelligent Vehicles |
ISSN | 2379-8858 |
Volume | 9Issue:1Pages:2727-2739 |
Abstract | High-precision LiDAR odometry (LO) plays an essential role in autonomous driving. Generally, due to inaccurate data associations and the existence of outliers, it is a challenging task to estimate ego-motions reliably and efficiently given sparse and complexly distributed point clouds in outdoor environments. In this paper, we propose an unsupervised method for LiDAR odometry, named HPPLO-Net, to predict the relative pose of a LiDAR sensor in a hierarchical way. Specifically, we achieve accurate 6-DoF (Degree of Freedom) pose estimation between the source and target point clouds using a differentiable Point-to-Plane solver with the assistance of scene flow. The novel Point-to-Plane solver consists of a multi-scale aggregation (MSA) normal estimation layer and a differentiable weighted Point-to-Plane SVD module. The MSA layer is introduced to find reliable normal vectors of the pseudo target point cloud by aggregating multi-scale contextual information. The differentiable weighted Point-to-Plane SVD is embedded in the network to solve the pose matrix and alleviate the problem of lacking accurate data association in two LiDAR scans, which exists in the point-to-point alternatives. To reduce the impact of noises and outliers, our method can learn and update the inlier mask and the (residual) flow uncertainty at each layer of the hierarchy. We demonstrate the effectiveness of our method on the KITTI Odometry Dataset, Ford Campus Vision and Lidar DataSet, and the Apollo-SouthBay Dataset. Our method has achieved superior performance than recent unsupervised learning-based methods and some traditional geometry-based methods and has also achieved promising performance close to A-LOAM with mapping optimization. The source code of our method is available at https://github.com/IMRL/HPPLO-Net. |
Keyword | Costs Feature Extraction Hierarchical Framework Laser Radar Lidar Odometry Msa Layer Odometry Optimization Point Cloud Compression Scene Flow Three-dimensional Displays Weighted Point-to-plane Svd |
DOI | 10.1109/TIV.2023.3288943 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Computer Science ; Engineering ; Transportation |
WOS Subject | Computer Science, Artificial Intelligence ; Engineering, Electrical & Electronic ; Transportation Science & Technology |
WOS ID | WOS:001173317800227 |
Publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 445 HOES LANE, PISCATAWAY, NJ 08855-4141 |
Scopus ID | 2-s2.0-85163425582 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | Faculty of Science and Technology THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU) DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Affiliation | 1.Nanjing University of Science and Technology, Nanjing, Jiangsu, China 2.State Key Laboratory of Internet of Things for Smart City (SKL-IOTSC), Department of Electromechanical Engineering (EME), University of Macau, Macau, China |
Recommended Citation GB/T 7714 | Zhou,Beibei,Tu,Yiming,Jin,Zhong,et al. HPPLO-Net: Unsupervised LiDAR Odometry Using a Hierarchical Point-to-Plane Solver[J]. IEEE Transactions on Intelligent Vehicles, 2023, 9(1), 2727-2739. |
APA | Zhou,Beibei., Tu,Yiming., Jin,Zhong., Xu,Chengzhong., & Kong,Hui (2023). HPPLO-Net: Unsupervised LiDAR Odometry Using a Hierarchical Point-to-Plane Solver. IEEE Transactions on Intelligent Vehicles, 9(1), 2727-2739. |
MLA | Zhou,Beibei,et al."HPPLO-Net: Unsupervised LiDAR Odometry Using a Hierarchical Point-to-Plane Solver".IEEE Transactions on Intelligent Vehicles 9.1(2023):2727-2739. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment