Dense-ACSSD for End-To-End Traffic Scenes Recognition

Zhiwei Cheng1, Zhenyang Wang1, Hongcheng Huang1, Yanbo Liu2

  • 1Shanghai Jiao Tong University
  • 2Shang Hai Jiao Tong University

Details

11:00 - 12:30 | Mon 10 Jun | Room 5 | MoAM_P1.9

Session: Poster 1: AV + Vision

Abstract

Traffic scenes recognition is the cornerstone of autonomous driving. However, most of the current algorithms are individually trained for tasks such as object detection and road segmentation. In addition, the training data used are mainly concentrated in small datasets such as KITTI, and the trained models are highly susceptible to weather, lighting and other factors. In order to solve the above problems, we propose an end-to-end CNN model for driveable area segmentation and multiple object detection. The feature extraction part of the network is powerful DenseNet. The atrous convolution and spatial pyramid pooling are used for road segmentation, and single shot detection is used for multiple object detection. According to its characteristics, we named the network Dense-ACSSD. Dense-ACSSD is trained on the current largest autonomous driving dataset, called BDD100K. The final training results show that the mIOU of the driveable area segmentation part is as high as 84.15%, and the mAP of the multiple object detection part reaches 30.82%. In addition, the inference time of Dense-ACSSD can meet real-time requirements.