DOI QR코드

DOI QR Code

Concurrent Detection for Vehicles and Lanes Using Light-Weight Model of Multi-Task CNN

멀티 테스크 CNN의 경량화 모델을 이용한 차량 및 차선의 동시 검출

  • Shin, Hyeon-Sik (College of Electrical & Computer Engineering, ChungBuk University) ;
  • Kim, Hyung-Won (College of Electrical & Computer Engineering, ChungBuk University) ;
  • Hong, Sang-Wook (College of Electrical & Computer Engineering, ChungBuk University)
  • Received : 2021.12.23
  • Accepted : 2021.12.29
  • Published : 2022.03.31

Abstract

As deep learning-based autonomous driving technology develops, artificial intelligence models for various purposes have been studied. Based on these studies, several models were used simultaneously to develop autonomous driving systems. It can occur by increasing hardware resource consumption. We propose a multi-tasks model using a shared backbone to solve this problem. This can solve the increase in the number of backbones for using AI models. As a result, in the proposed lightweight model, the model parameters could be reduced by more than 50% compared to the existing model, and the speed could be improved. In addition, each lane can be classified through lane detection using the instance segmentation method. However, further research is needed on the decrease in accuracy compared to the existing model.

딥러닝 기반 자율 주행 기술이 발전함에 따라 다양한 목적의 인공지능 모델이 연구되었다. 연구된 여러 모델들을 동시에 구동하여 자율주행 시스템을 개발한다. 그러나 동시에 인공지능 모델을 사용하면서 많은 하드웨어 자원 소비가 증가한다. 이를 해결하기 위해 본 논문은 백본 모델을 공유하며 다중 태스크를 고속으로 수행할 수 있는 Multi-Task CNN 모델을 제안한다. 이를 통해 AI모델을 사용하기 위한 백본 수의 증가를 해결할 수 있었습니다. 제안하는 CNN 모델은 기존 모델 대비 50% 이상 웨이트 파라미터 수를 감소시키며, 3배 이상의 FPS 속도를 향상시켰다. 또한, 차선인식은 Instance segmentation 기반으로 차선검출 및 차선별 Labeling을 모두 출력한다. 그러나 기존 모델에 비해 정확도가 감소하는 부분에 대해서는 추가적인 연구가 필요하다.

Keywords

Acknowledgement

This work was partly supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No.2020-0-01304, Development of Self-learnable Mobile Recursive Neural Network Processor Technology) and also supported by the MSIT(Ministry of Science and ICT), Korea, under the Grand Information Technology Research Center support program(IITP-2022-2020-0-01462) and supervised by the IITP(Institute for Information & communications Technology Planning & Evaluation)". and also supported by the AURI(Korea Association of University, Research institute and Industry) grant funded by the Korea Government(MSS : Ministry of SMEs and Startups). (No. S2929950, HRD program for 2020)

References

  1. R. Girshick, J. Donahue, T. Darrell, and J. Malik, "Rich feature hierarchies for accurate object detection and semantic segmentation," in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 580-587, Jun. 2014.
  2. R. Girshick, "Fast R-CNN," in Proceedings of the IEEE international conference on computer vision, pp. 1440-1448, Dec. 2015.
  3. J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, "You only look once: Unified, real-time object detection," in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779-788, Dec. 2016.
  4. X. Li, J. Li, X. Hu, and J. Yang, "Line-CNN: End-to-end traffic line detection with line proposal unit," IEEE Transactions on Intelligent Transportation Systems, vol. 21, iss. 1, pp. 248-258, Jan. 2019. https://doi.org/10.1109/tits.2019.2890870
  5. L. Tabelini, R. Berriel, T. M. Paixao, C. Badue, A. F. De Souza, and T. Oliveira-Santos, "Keep your eyes on the lane: Real-time attention-guided lane detection," in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 294-302, Jun. 2021.
  6. J. Su, C. Chen, K. Zhang, J. Luo, X. Wei, and X. Wei, "Structure Guided Lane Detection," in Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, pp. 997-1003, 2021.
  7. D. Wu, M. Liao, W. Zhang, and X. Wang, "YOLOP: You Only Look Once for Panoptic Driving Perception," arXiv preprint arXiv:2108.11250, Feb. 2021.
  8. Z. Qin, H. Wang, and X. Li, "Ultra fast structure-aware deep lane detection," in Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, pp. 276-291, Aug. 2020.
  9. A. Bochkovskiy, C. Y. Wang, and H. Y. M. Liao, "Yolov4: Optimal speed and accuracy of object detection," arXiv preprint arXiv:2004.10934, Apr. 2020.
  10. M. Teichmann, M. Weber, M. Zollner, R. Cipolla, and R. Urtasun, "Multinet: Real-time joint semantic reasoning for autonomous driving," in 2018 IEEE Intelligent Vehicles Symposium (IV), pp. 1013-1020, Jun. 2018.
  11. Y. Qian, J. M. Dolan, and M. Yang, "DLT-Net: Joint detection of drivable areas, lane lines, and traffic objects," IEEE Transactions on Intelligent Transportation Systems, vol. 21, no. 11, pp. 4670-4679, Dec. 2019. https://doi.org/10.1109/tits.2019.2943777
  12. S. Ren, K. He, R. Girshick, and J. Sun, "Faster R-CNN: Towards real-time object detection with region proposal networks," Advances in neural information processing systems, vol. 28, pp. 91-99, 2015.
  13. yolov5 [Internet]. Available: https://github.com/ultralytics/yolov5.
  14. L. C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, "Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFS," IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 4, pp. 834-848, Apr. 2017. https://doi.org/10.1109/TPAMI.2017.2699184
  15. X. Pan, J. Shi, P. Luo, X. Wang, and X. Tang , "Spatial as deep: Spatial CNN for traffic scene understanding," in Thirty- Second AAAI Conference on Artificial Intelligence, vol. 32, no. 1, Apr. 2018.