• Laser & Optoelectronics Progress
  • Vol. 61, Issue 4, 0412004 (2024)
Yang Wang1, Dudu Guo2、*, Qingqing Wang1, Fei Zhou1, and Ying Qin1
Author Affiliations
  • 1School of Intelligent Manufacturing Modern Industry, Xinjiang University, Urumqi 830017, Xinjiang , China
  • 2School of Traffic and Transportation Engineering, Xinjiang University, Urumqi 830017, Xinjiang , China
  • show less
    DOI: 10.3788/LOP231270 Cite this Article Set citation alerts
    Yang Wang, Dudu Guo, Qingqing Wang, Fei Zhou, Ying Qin. UAV Highway Guardrail Inspection Based on Improved DeepLabV3+[J]. Laser & Optoelectronics Progress, 2024, 61(4): 0412004 Copy Citation Text show less

    Abstract

    To address the problems of slow prediction speed and low segmentation accuracy of existing semantic segmentation methods for highway guardrail detection, an UAV highway guardrail detection method based on improved DeepLabV3+ is proposed. First, the MobileNetv2 network was used to replace the backbone of the original model and outputs the middle layer's features to reduce number of parameters while recovering the spatial information lost in the downsampling process; then an atrous spatial pyramid pooling was improved by the densely connected expansive convolution to reduce the phenomenon of missed segmentation; finally, the spatial group-wise enhance (SGE) attention mechanism was introduced in the encoder part to reduce the phenomenon of wrong segmentation. The experimental results show that the average intersection over union, average pixel accuracy, and frames per second transmission of the improved model can reach 79.20%, 87.89%, and 52.59, which are 2.59%, 2.93%, and 56.70% higher than the base model, respectively, and number of parameters is reduced by 78.85%. This method can thus improve the segmentation accuracy for the guardrail while guaranteeing the model's prediction speed.
    Yang Wang, Dudu Guo, Qingqing Wang, Fei Zhou, Ying Qin. UAV Highway Guardrail Inspection Based on Improved DeepLabV3+[J]. Laser & Optoelectronics Progress, 2024, 61(4): 0412004
    Download Citation