arXiv Analytics

Sign in

arXiv:2304.10891 [cs.LG]AbstractReferencesReviewsResources

Transformer-based models and hardware acceleration analysis in autonomous driving: A survey

Juan Zhong, Zheng Liu, Xi Chen

Published 2023-04-21Version 1

Transformer architectures have exhibited promising performance in various autonomous driving applications in recent years. On the other hand, its dedicated hardware acceleration on portable computational platforms has become the next critical step for practical deployment in real autonomous vehicles. This survey paper provides a comprehensive overview, benchmark, and analysis of Transformer-based models specifically tailored for autonomous driving tasks such as lane detection, segmentation, tracking, planning, and decision-making. We review different architectures for organizing Transformer inputs and outputs, such as encoder-decoder and encoder-only structures, and explore their respective advantages and disadvantages. Furthermore, we discuss Transformer-related operators and their hardware acceleration schemes in depth, taking into account key factors such as quantization and runtime. We specifically illustrate the operator level comparison between layers from convolutional neural network, Swin-Transformer, and Transformer with 4D encoder. The paper also highlights the challenges, trends, and current insights in Transformer-based models, addressing their hardware deployment and acceleration issues within the context of long-term autonomous driving applications.

Related articles: Most relevant | Search more
arXiv:1809.01564 [cs.LG] (Published 2018-09-05)
Traffic Density Estimation using a Convolutional Neural Network
arXiv:2105.04232 [cs.LG] (Published 2021-05-10)
De-homogenization using Convolutional Neural Networks
arXiv:1603.03657 [cs.LG] (Published 2016-03-11)
Efficient forward propagation of time-sequences in convolutional neural networks using Deep Shifting