DSNet: An efficient CNN for road scene segmentation

Ping Rong Chen, Hsueh Ming Hang*, Sheng Wei Chan, Jing Jhih Lin

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Road scene understanding is a critical component in an autonomous driving system. Although the deep learning-based road scene segmentation can achieve very high accuracy, its complexity is also very high for developing real-time applications. It is challenging to design a neural net with high accuracy and low computational complexity. To address this issue, we investigate the advantages and disadvantages of several popular convolutional neural network (CNN) architectures in terms of speed, storage, and segmentation accuracy. We start from the fully convolutional network with VGG, and then we study ResNet and DenseNet. Through detailed experiments, we pick up the favorable components from the existing architectures and at the end, we construct a light-weight network architecture based on the DenseNet. Our proposed network, called DSNet, demonstrates a real-time testing (inferencing) ability (on the popular GPU platform) and it maintains an accuracy comparable with most previous systems. We test our system on several datasets including the challenging Cityscapes dataset (resolution of 1024 × 512) with an Mean Intersection over Union (mIoU) of about 69.1% and runtime of 0.0147 s/image on a single GTX 1080Ti. We also design a more accurate model but at the price of a slower speed, which has an mIoU of about 72.6% on the CamVid dataset.

Original languageEnglish
Article numbere27
JournalAPSIPA Transactions on Signal and Information Processing
DOIs
StateAccepted/In press - 2020

Keywords

  • CNN architecture
  • Real-time CNN segmentation
  • Road scene segmentation
  • Semantic segmentation

Fingerprint Dive into the research topics of 'DSNet: An efficient CNN for road scene segmentation'. Together they form a unique fingerprint.

Cite this