Residential Collegefalse
Status已發表Published
Geometry-Aware Network for Unsupervised Learning of Monocular Camera's Ego-Motion
Zhou, Beibei1; Xie, Jin1; Jin, Zhong1; Kong, Hui2
2023-12-01
Source PublicationIEEE Transactions on Intelligent Transportation Systems
ISSN1524-9050
Volume24Issue:12Pages:14226-14236
Abstract

Deep neural networks have been shown to be effective for unsupervised monocular visual odometry that can predict the camera's ego-motion based on an input of monocular video sequence. However, most existing unsupervised monocular methods haven't fully exploited the extracted information from both local geometric structure and visual appearance of the scenes, resulting in degraded performance. In this paper, a novel geometry-aware network is proposed to predict the camera's ego-motion by learning representations in both 2D and 3D space. First, to extract geometry-aware features, we design an RGB-PointCloud feature fusion module to capture information from both geometric structure and the visual appearance of the scenes by fusing local geometric features from depth-map-derived point clouds and visual features from RGB images. Furthermore, the fusion module can adaptively allocate different weights to the two types of features to emphasize important regions. Then, we devise a relevant feature filtering module to build consistency between the two views and preserve informative features with high relevance. It can capture the correlation of frame pairs in the feature-embedding space by attention mechanisms. Finally, the obtained features are fed into the pose estimator to recover the 6-DoF poses of the camera. Extensive experiments show that our method achieves promising results among the unsupervised monocular deep learning methods on the KITTI odometry and TUM-RGBD datasets.

Keyword6-dof Poses Geometry-aware Monocular Visual Odometry Point Clouds Visual Appearance
DOI10.1109/TITS.2023.3298715
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaEngineering ; Transportation
WOS SubjectEngineering, Civil ; Engineering, Electrical & Electronic ; Transportation Science & Technology
WOS IDWOS:001047543000001
PublisherInstitute of Electrical and Electronics Engineers Inc.
Scopus ID2-s2.0-85166779935
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionTHE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
Corresponding AuthorKong, Hui
Affiliation1.Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, PCA Lab, Key Lab Intelligent Percept & Syst High Dimens Inf, Nanjing 210094, Jiangsu, Peoples R China
2.University of Macau, State Key Laboratory of Internet of Things for Smart City (SKL-IOTSC), Department of Electromechanical Engineering (EME), Macao
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Zhou, Beibei,Xie, Jin,Jin, Zhong,et al. Geometry-Aware Network for Unsupervised Learning of Monocular Camera's Ego-Motion[J]. IEEE Transactions on Intelligent Transportation Systems, 2023, 24(12), 14226-14236.
APA Zhou, Beibei., Xie, Jin., Jin, Zhong., & Kong, Hui (2023). Geometry-Aware Network for Unsupervised Learning of Monocular Camera's Ego-Motion. IEEE Transactions on Intelligent Transportation Systems, 24(12), 14226-14236.
MLA Zhou, Beibei,et al."Geometry-Aware Network for Unsupervised Learning of Monocular Camera's Ego-Motion".IEEE Transactions on Intelligent Transportation Systems 24.12(2023):14226-14236.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Zhou, Beibei]'s Articles
[Xie, Jin]'s Articles
[Jin, Zhong]'s Articles
Baidu academic
Similar articles in Baidu academic
[Zhou, Beibei]'s Articles
[Xie, Jin]'s Articles
[Jin, Zhong]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Zhou, Beibei]'s Articles
[Xie, Jin]'s Articles
[Jin, Zhong]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.