Residential Collegefalse
Status已發表Published
Lane change strategies for autonomous vehicles: a deep reinforcement learning approach based on transformer
Li, Guofa1; Qiu, Yifan2; Yang, Yifan2; Li, Zhenning3; Li, Shen4; Chu, Wenbo1,5; Green, Paul6; Li, Shengbo Eben7
2022-12-09
Source PublicationIEEE Transactions on Intelligent Vehicles
ISSN2379-8858
Volume8Issue:3Pages:2197 - 2211
Abstract

End-to-end approaches are one of the most promising solutions for autonomous vehicles (AVs) decision-making. However, the deployment of these technologies is usually constrained by the high computational burden. To alleviate this problem, we proposed a lightweight transformer-based end-to-end model with risk awareness ability for AV decision-making. Specifically, a lightweight network with depth-wise separable convolution and transformer modules was firstly proposed for image semantic extraction from time sequences of trajectory data. Then, we assessed driving risk by a probabilistic model with position uncertainty. This model was integrated into deep reinforcement learning (DRL) to find strategies with minimum expected risk. Finally, the proposed method was evaluated in three lane change scenarios to validate its superiority.

KeywordAutonomous Vehicles Decision-making Reinforcement Learning Lane Change Transformer
DOI10.1109/TIV.2022.3227921
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaComputer Science ; Engineering ; Transportation
WOS SubjectComputer Science, Artificial Intelligence ; Engineering, Electrical & Electronic ; Transportation Science & Technology
WOS IDWOS:000981348100018
PublisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 445 HOES LANE, PISCATAWAY, NJ 08855-4141
Scopus ID2-s2.0-85144753699
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionDEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
Corresponding AuthorLi, Shen
Affiliation1.College of Mechanical and Vehicle Engineering, Chongqing University, Chongqing, China
2.College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen, Guangdong, China
3.State Key Laboratory of Internet of Things for Smart City and the Department of Computer and Information Science, University of Macau, Macau, China
4.School of Civil Engineering, Tsinghua University, Beijing, China
5.Western China Science City Innovation Center of Intelligent and Connected Vehicles (Chongqing) Co, Ltd., Chongqing, China
6.University of Michigan Transportation Research Institute (UMTRI) & Department of Industrial and Operations Engineering, University of Michigan, Ann Arbor, MI, USA
7.State Key Lab of Automotive Safety and Energy, School of Vehicle and Mobility, Tsinghua University, Beijing, China
Recommended Citation
GB/T 7714
Li, Guofa,Qiu, Yifan,Yang, Yifan,et al. Lane change strategies for autonomous vehicles: a deep reinforcement learning approach based on transformer[J]. IEEE Transactions on Intelligent Vehicles, 2022, 8(3), 2197 - 2211.
APA Li, Guofa., Qiu, Yifan., Yang, Yifan., Li, Zhenning., Li, Shen., Chu, Wenbo., Green, Paul., & Li, Shengbo Eben (2022). Lane change strategies for autonomous vehicles: a deep reinforcement learning approach based on transformer. IEEE Transactions on Intelligent Vehicles, 8(3), 2197 - 2211.
MLA Li, Guofa,et al."Lane change strategies for autonomous vehicles: a deep reinforcement learning approach based on transformer".IEEE Transactions on Intelligent Vehicles 8.3(2022):2197 - 2211.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Li, Guofa]'s Articles
[Qiu, Yifan]'s Articles
[Yang, Yifan]'s Articles
Baidu academic
Similar articles in Baidu academic
[Li, Guofa]'s Articles
[Qiu, Yifan]'s Articles
[Yang, Yifan]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Li, Guofa]'s Articles
[Qiu, Yifan]'s Articles
[Yang, Yifan]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.