TY - GEN
T1 - NeurAll
T2 - 2019 IEEE Intelligent Transportation Systems Conference, ITSC 2019
AU - Sistu, Ganesh
AU - Leang, Isabelle
AU - Chennupati, Sumanth
AU - Yogamani, Senthil
AU - Hughes, Ciaran
AU - Milz, Stefan
AU - Rawashdeh, Samir
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/10
Y1 - 2019/10
N2 - Convolutional Neural Networks (CNNs) are successfully used for the important automotive visual perception tasks including object recognition, motion and depth estimation, visual SLAM, etc. However, these tasks are typically independently explored and modeled. In this paper, we propose a joint multi-task network design for learning several tasks simultaneously. Our main motivation is the computational efficiency achieved by sharing the expensive initial convolutional layers between all tasks. Indeed, the main bottleneck in automated driving systems is the limited processing power available on deployment hardware. There is also some evidence for other benefits in improving accuracy for some tasks and easing development effort. It also offers scalability to add more tasks leveraging existing features and achieving better generalization. We survey various CNN based solutions for visual perception tasks in automated driving. Then we propose a unified CNN model for the important tasks and discuss several advanced optimization and architecture design techniques to improve the baseline model. The paper is partly review and partly positional with demonstration of several preliminary results promising for future research. We first demonstrate results of multi-stream learning and auxiliary learning which are important ingredients to scale to a large multi-task model. Finally, we implement a two-stream three-task network which performs better in many cases compared to their corresponding single-task models, while maintaining network size.
AB - Convolutional Neural Networks (CNNs) are successfully used for the important automotive visual perception tasks including object recognition, motion and depth estimation, visual SLAM, etc. However, these tasks are typically independently explored and modeled. In this paper, we propose a joint multi-task network design for learning several tasks simultaneously. Our main motivation is the computational efficiency achieved by sharing the expensive initial convolutional layers between all tasks. Indeed, the main bottleneck in automated driving systems is the limited processing power available on deployment hardware. There is also some evidence for other benefits in improving accuracy for some tasks and easing development effort. It also offers scalability to add more tasks leveraging existing features and achieving better generalization. We survey various CNN based solutions for visual perception tasks in automated driving. Then we propose a unified CNN model for the important tasks and discuss several advanced optimization and architecture design techniques to improve the baseline model. The paper is partly review and partly positional with demonstration of several preliminary results promising for future research. We first demonstrate results of multi-stream learning and auxiliary learning which are important ingredients to scale to a large multi-task model. Finally, we implement a two-stream three-task network which performs better in many cases compared to their corresponding single-task models, while maintaining network size.
UR - http://www.scopus.com/inward/record.url?scp=85076804665&partnerID=8YFLogxK
U2 - 10.1109/ITSC.2019.8917043
DO - 10.1109/ITSC.2019.8917043
M3 - Conference contribution
AN - SCOPUS:85076804665
T3 - 2019 IEEE Intelligent Transportation Systems Conference, ITSC 2019
SP - 796
EP - 803
BT - 2019 IEEE Intelligent Transportation Systems Conference, ITSC 2019
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 27 October 2019 through 30 October 2019
ER -