Citation: | Zhang C, Huang Y P, Guo Z Y, et al. Real-time lane detection method based on semantic segmentation[J]. Opto-Electron Eng, 2022, 49(5): 210378. doi: 10.12086/oee.2022.210378 |
[1] | Wang Y, Shen D G, Teoh E K. Lane detection using spline model[J]. Pattern Recognit Lett, 2000, 21(8): 677−689. doi: 10.1016/S0167-8655(00)00021-0 |
[2] | Jiang R Y, Reinhard K, Tobi V, et al. Lane detection and tracking using a new lane model and distance transform[J]. Mach Vis Appl, 2011, 22(4): 721−737. doi: 10.1007/s00138-010-0307-7 |
[3] | 朱鸿宇, 杨帆, 高晓倩, 等. 基于级联霍夫变换的车道线快速检测算法[J]. 计算机技术与发展, 2021, 31(1): 88−93. doi: 10.3969/j.issn.1673-629X.2021.01.016 Zhu H Y, Yang F, Gao X Q, et al. A fast lane detection algorithm based on cascade Hough transform[J]. Comput Technol Dev, 2021, 31(1): 88−93. doi: 10.3969/j.issn.1673-629X.2021.01.016 |
[4] | Tian J, Liu S W, Zhong X Y, et al. LSD-based adaptive lane detection and tracking for ADAS in structured road environment[J]. Soft Comput, 2021, 25(7): 5709−5722. doi: 10.1007/s00500-020-05566-4 |
[5] | Qin Z Q, Wang H Y, Li X. Ultra fast structure-aware deep lane detection[C]//16th European Conference on Computer Vision, 2020: 276–291. |
[6] | Chen Z P, Liu Q F, Lian C F. PointLaneNet: efficient end-to-end CNNs for accurate real-time lane detection[C]//2019 IEEE Intelligent Vehicles Symposium (IV), 2019: 2563–2568. |
[7] | Tabelini L, Berriel R, Paixão T M, et al. PolyLaneNet: lane estimation via deep polynomial regression[C]//2020 25th International Conference on Pattern Recognition (ICPR), 2021: 6150–6156. |
[8] | Ji G Q, Zheng Y C. Lane line detection system based on improved Yolo V3 algorithm[Z]. Research Square: 2021. https://doi.org/10.21203/rs.3.rs-961172/v1. |
[9] | Neven D, De Brabandere B, Georgoulis S, et al. Towards end-to-end lane detection: an instance segmentation approach[C]//2018 IEEE Intelligent Vehicles Symposium (IV), 2018: 286–291. |
[10] | 刘彬, 刘宏哲. 基于改进Enet网络的车道线检测算法[J]. 计算机科学, 2020, 47(4): 142−149. doi: 10.11896/jsjkx.190500021 Liu B, Liu H Z. Lane detection algorithm based on improved Enet network[J]. Comput Sci, 2020, 47(4): 142−149. doi: 10.11896/jsjkx.190500021 |
[11] | 田锦, 袁家政, 刘宏哲. 基于实例分割的车道线检测及自适应拟合算法[J]. 计算机应用, 2020, 40(7): 1932−1937. doi: 10.11772/j.issn.1001-9081.2019112030 Tian J, Yuan J Z, Liu H Z. Instance segmentation based lane line detection and adaptive fitting algorithm[J]. J Comput Appl, 2020, 40(7): 1932−1937. doi: 10.11772/j.issn.1001-9081.2019112030 |
[12] | Pan X G, Shi J P, Luo P, et al. Spatial as deep: spatial CNN for traffic scene understanding[C]//Thirty-Second AAAI Conference on Artificial Intelligence, 2018: 7276–7283. |
[13] | Woo S, Park J, Lee J Y, et al. CBAM: convolutional block attention module[C]//Proceedings of the 15th European Conference on Computer Vision (ECCV), 2018: 3–19. |
[14] | Howard A G, Zhu M L, Chen B, et al. MobileNets: efficient convolutional neural networks for mobile vision applications[Z]. arXiv:1704.04861, 2017. http://www.arxiv.org/abs/1704.04861. |
[15] | Yu F, Koltun V. Multi-scale context aggregation by dilated convolutions[C]//4th International Conference on Learning Representations, 2016. |
[16] | Chollet F. Xception: deep learning with depthwise separable convolutions[C]//2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017: 1800–1807. |
[17] | Szegedy C, Vanhoucke V, Ioffe S, et al. Rethinking the inception architecture for computer vision[C]//2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016: 2818–2826. |
[18] | Wu B C, Wan A, Yue X Y, et al. SqueezeSeg: convolutional neural nets with recurrent CRF for real-time road-object segmentation from 3D LiDAR point cloud[C]//2018 IEEE International Conference on Robotics and Automation (ICRA), 2018: 1887–1893. |
[19] | Ren S Q, He K M, Girshick R, et al. Faster R-CNN: towards real-time object detection with region proposal networks[J]. IEEE Trans Pattern Anal Mach Intell, 2017, 39(6): 1137−1149. doi: 10.1109/TPAMI.2016.2577031 |
In recent years, the rapid development of neural network has greatly improved the efficiency of lane detection. However, convolutional neural network has become a new problem restricting the development of lane detection because of its large amount of calculation and high hardware requirements. Lane detection methods based on deep learning can be divided into two categories: detection based methods and segmentation based methods. The method based on detection has the advantages of high speed and strong ability to deal with straight lane. However, when the environment is complex and there are many curves, the detection effect is obviously not as good as the segmentation based method. This paper adopts the segmentation based method, and considers that the performance of lane detection can be improved by establishing global context correlation and enhancing the effective expression of important Lane feature channels. Attention mechanism is a model that can significantly improve network performance. It imitates the human visual processing mechanism, strengthens the attention to important information, so as to reasonably allocate network resources and improve the detection efficiency and accuracy of the network. Therefore, this paper uses the CBAM model. In this model, channel attention and spatial attention are serial to obtain better feature representation ability. Spatial attention learns the positional relationship between lane line pixels, and channel attention learns the importance of different channel features. In addition, in order to solve the problem of complex convolution calculation and slow running speed based on segmentation model, a more efficient convolution structure is proposed to improve the computational efficiency. A new fast down sampling module laneconv and a new fast up sampling module laneconv are introduced, and the depth separable convolution is introduced to further reduce the amount of calculation. They are located in the coding part of the network. The decoding part outputs the binary segmentation result. Then, the results are clustered by DBSCAN to obtain the lane line. After clustering, compared with the complex post-processing in other literature, this paper only uses simple cubic fitting to fit the lane line, which further improves the speed. Therefore, the running speed of the model proposed in this paper is better than most segmentation based methods. Finally, a large number of experiments are carried out on tusimple Lane database. The results show that the method has good robustness under various road conditions, especially in the case of occlusion. Compared with the existing models, it has comprehensive advantages in detection accuracy and speed.
Framework of the method
The parameters of the network and illustration of LaneConv and LaneDeconv
Depth separable convolution.
(a) Laneconv structure; (b) Lanedeconv structure
(a) Channel attention; (b) Spatial attention
DBSCAN cluster
The output in different stages.
Comparison between visualization results of baseline and our method on TuSimple.
Comparison of effects before and after adding CBAM.
Visual results generated by our method on some of typical scenarios
Framework of the method