arXiv Analytics

Sign in

arXiv:1607.00971 [cs.CV]AbstractReferencesReviewsResources

Can we unify monocular detectors for autonomous driving by using the pixel-wise semantic segmentation of CNNs?

Eduardo Romera, Luis M. Bergasa, Roberto Arroyo

Published 2016-07-04Version 1

Autonomous driving is a challenging topic that requires complex solutions in perception tasks such as recognition of road, lanes, traffic signs or lights, vehicles and pedestrians. Through years of research, computer vision has grown capable of tackling these tasks with monocular detectors that can provide remarkable detection rates with relatively low processing times. However, the recent appearance of Convolutional Neural Networks (CNNs) has revolutionized the computer vision field and has made possible approaches to perform full pixel-wise semantic segmentation in times close to real time (even on hardware that can be carried on a vehicle). In this paper, we propose to use full image segmentation as an approach to simplify and unify most of the detection tasks required in the perception module of an autonomous vehicle, analyzing major concerns such as computation time and detection performance.

Comments: Extended abstract presented in IV16-WS Deepdriving (http://iv2016.berkeleyvision.org/)
Categories: cs.CV
Related articles: Most relevant | Search more
arXiv:1505.00256 [cs.CV] (Published 2015-05-01)
DeepDriving: Learning Affordance for Direct Perception in Autonomous Driving
arXiv:1803.06184 [cs.CV] (Published 2018-03-16)
The ApolloScape Dataset for Autonomous Driving
Xinyu Huang et al.
arXiv:1901.09270 [cs.CV] (Published 2019-01-26)
Challenges in Designing Datasets and Validation for Autonomous Driving