Please wait a minute...
Journal of ZheJiang University (Engineering Science)  2026, Vol. 60 Issue (4): 738-750    DOI: 10.3785/j.issn.1008-973X.2026.04.006
    
Improved algorithm for identifying occluded vehicles and pedestrians in foggy images
Tianhe YU(),Wenlong WANG(),Yong LIU,Zhuangzhuang YANG,Shanchong HOU
School of Measurement and Control Technology and Communication Engineering, Harbin University of Science and Technology, Harbin 150006, China
Download: HTML     PDF(5503KB) HTML
Export: BibTeX | EndNote (RIS)      

Abstract  

Two key technical improvements were proposed to address the issue of decreased target detection accuracy caused by target occlusion and fog interference in intelligent autonomous driving scenarios. Firstly, an improved detection method was proposed for the problem of target occlusion. A lightweight MobileNetV3_small with an integrated enhanced attention mechanism was used as the SSD backbone feature extraction network. The detection accuracy in occlusion scenarios was enhanced by combining a multi-scale feature fusion mechanism and an adaptive hyperparameter Soft-NMS algorithm. The confidence loss function was reconstructed by improving the adaptive Focal Loss to alleviate the imbalance between positive and negative samples and the sensitivity to noisy labels. Secondly, an improved lightweight AOD-Net defogging method was proposed to address the problem of target blurriness in foggy images. The image contrast was effectively enhanced and texture details were improved, by constructing a multi-scale feature extraction network based on depthwise separable convolution, optimizing the cross-layer connection structure, and introducing a boundary enhancement module. Finally, an end-to-end collaborative optimization of the defogging network and the detection network was achieved through a joint loss function, providing a more reliable optimization path for the detection task of occluded targets in foggy images. Experimental results showed that the joint optimization model improved the target detection performance in foggy occlusion scenarios, achieving efficient detection with an accuracy of 93.85% and a detection speed of 47.61 frames per second, and demonstrating excellent model robustness.



Key wordsoccluded target      foggy image      feature fusion      target detection      feature extraction     
Received: 26 July 2025      Published: 19 March 2026
CLC:  TP 391.41  
  U 463.6  
Cite this article:

Tianhe YU,Wenlong WANG,Yong LIU,Zhuangzhuang YANG,Shanchong HOU. Improved algorithm for identifying occluded vehicles and pedestrians in foggy images. Journal of ZheJiang University (Engineering Science), 2026, 60(4): 738-750.

URL:

https://www.zjujournals.com/eng/10.3785/j.issn.1008-973X.2026.04.006     OR     https://www.zjujournals.com/eng/Y2026/V60/I4/738


改进的有雾图像中被遮挡车辆及行人识别算法

针对智能自动驾驶场景中目标遮挡与雾天干扰导致的目标检测精度下降问题,提出2项关键技术改进. 针对目标遮挡问题提出改进检测方法,以集成增强注意力机制的轻量化MobileNetV3_small作为SSD骨干特征提取网络,结合多尺度特征融合机制与自适应超参数Soft-NMS算法提升遮挡场景下的检测精度,通过改进自适应Focal Loss重构置信度损失函数,缓解正负样本不平衡及噪声标签敏感性问题. 针对雾天图像中目标模糊的问题提出改进轻量化AOD-Net去雾方法,通过构建基于深度可分离卷积的多尺度特征提取网络,优化跨层连接结构并引入边界增强模块,有效提升图像对比度、增强纹理细节. 通过联合损失函数对去雾网络与检测网络进行端到端协同优化,为有雾图像中的遮挡目标检测任务提供更可靠的优化路径. 实验结果表明,联合优化模型提升了雾天遮挡场景下的目标检测性能,以93.85%的准确率和47.61帧/s的检测速度实现了高效检测,并表现出优异的模型鲁棒性.


关键词: 被遮挡目标,  雾天图像,  特征融合,  目标检测,  特征提取 
Fig.1 AOD-Net dehazing algorithm process
Fig.2 ABMD-Net dehazing network model
Fig.3 Boundary enhancement module
Fig.4 SSD target detection network diagram
Fig.5 MsF-SSD-Net network structure
Fig.6 Feature extraction network architecture diagram
Fig.7 Comparison of detection results by feature extraction networks
Fig.8 Ordinary-deep separable convolutional double branch pooling splicing fusion module
不同场景$ \beta $V/km
薄雾0.0050.6≤V<1.0
中雾0.0100.3≤V<0.6
重雾0.020V<0.3
Tab.1 Defogging evaluation index
模型薄雾中雾重雾FPS/(帧·s?1)
PSNRSSIMPSNRSSIMPSNRSSIM
AOD-Net19.490.882818.150.803217.790.766438.46
MSCNN[25]19.800.783418.150.821916.920.793927.78
GCANet[25]22.430.928218.110.757516.940.76557.46
GFN[25]21.150.923917.330.760916.930.738714.70
DehazeNet[25]20.460.875319.180.836518.890.827523.81
ABMD-Net24.020.939120.320.846120.030.835186.96
Improve AOD-Net[25]21.800.922817.790.851014.140.8024113.64
Tab.2 PSNR and SSIM values on foggy images
Fig.9 Defogging effect comparison of network models
模型FPS/
(帧·s?1)
F1PrRemAP
SoPoSe
SSD15.380.81350.83050.75630.77570.53250.3883
Faster
R-CNN
8.130.84130.87280.78230.80490.69260.5264
YOLOv1055.560.87680.91020.82260.83360.75540.5727
MsF-SSD-Net74.070.85450.89010.81890.92410.81360.5812
DCT-
YOLO[26]
0.84000.91600.77500.8870
improved
YOLOv7[27]
252.000.77000.78200.76900.8220
Tab.3 Performance comparison of various object detection models based on CityPersons dataset under occlusion scenarios
Fig.10 Comparison of training evaluation metrics for different models based on CityPersons dataset
模型mAPFPS/(帧·s?1)
清晰图像薄雾中雾重雾SoPoSe
SSD0.78190.71230.65560.53450.72850.47840.345414.92
Faster R-CNN0.81470.74870.68900.57890.75310.61320.48238.04
YOLOv100.83380.76340.70120.60560.77920.65730.495652.62
MsF-SSD-Net0.88760.84010.79340.66780.80520.71490.503471.43
lightweight YOLOv8[28]0.8100166.00
Tab.4 Quantitative performance comparison of various object detection models on Foggy Cityscapes dataset with fog data
Fig.11 Visual comparison chart of image detection results after defogging by ABMD-Net
模型mAPFPS/(帧·s?1)
薄雾中雾重雾SoPoSe
ABMD-Net+SSD0.75270.69560.59670.76450.52840.37639.74
ABMD-Net+Faster R-CNN0.80750.73900.62580.79840.68420.51665.38
ABMD-Net+YOLOv100.82340.78120.67340.80260.71440.542832.79
AOD-Net+MsF-SSD-Net0.88370.83650.71490.83320.74830.555625.03
ABMD-Net+MsF-SSD-Net0.91920.85340.73780.87780.76340.563345.21
Defog YOLO[29]0.8670
YOLOv5-Transformer[30]0.8280
AO YOLO[31]0.8910
ABMD-Net-MsF-SSD-Net0.93850.86830.78240.89520.78930.574847.61
Tab.5 Quantitative comparison of detection performance of each model after ABMD-Net defogging treatment
Fig.12 Comparison of training losses of different network models on Foggy Cityscapes dataset
算法改动薄雾中雾重雾FPS/(帧·s?1)
PSNRSSIMPSNRSSIMPSNRSSIM
AOD-Net20.310.852118.150.803217.790.766438.17
AOD-Net120.820.865918.870.805317.920.798437.88
AOD-Net1(DW)20.730.855818.780.791317.860.7965117.65
AOD-Net1(DW)(Pyramid)22.430.897419.440.824618.840.8147107.53
ABMD-Net24.020.939120.320.846120.030.835186.96
Tab.6 Ablation experimental results of AOD-Net algorithm
Fig.13 Comparison of AOD-Net ablation experimental results
算法改动mAPF1Time/msFPS/(帧·s?1)PrReParams/106
SSD0.78140.813565.115.360.83050.7563138.36
MobileNetV3+SSD0.77830.79748.5117.650.82270.739425.54
MobileNetV3-EC+SSD0.85450.830213.275.760.85790.762826.73
MobileNetV3-EC+SSD10.87320.841714.867.570.86710.779328.47
MobileNetV3-EC+SSD1+Soft-NMS0.88270.854315.365.360.88280.792428.47
MsF-SSD-Net0.88750.855413.574.070.89010.808928.47
L-SSD[32]0.7380106.00
Tab.7 Ablation experiment results of SSD algorithm
[1]   TAHIR N U A, ZHANG Z, ASIM M, et al Object detection in autonomous vehicles under adverse weather: a review of traditional and deep learning approaches[J]. Algorithms, 2024, 17 (3): 103
doi: 10.3390/a17030103
[2]   ABDEL-ATY M, CAI Q, WU Y, et al Evaluation of automated emergency braking system’s avoidance of pedestrian crashes at intersections under occluded conditions within a virtual simulator[J]. Accident Analysis and Prevention, 2022, 176: 106797
doi: 10.1016/j.aap.2022.106797
[3]   SHETTY A, YU M, KURZHANSKIY A, et al Safety challenges for autonomous vehicles in the absence of connectivity[J]. Transportation Research Part C: Emerging Technologies, 2021, 128: 103133
doi: 10.1016/j.trc.2021.103133
[4]   MO Y, VIJAY R, RUFUS R, et al Enhanced perception for autonomous vehicles at obstructed intersections: an implementation of vehicle to infrastructure (V2I) collaboration[J]. Sensors, 2024, 24 (3): 936
doi: 10.3390/s24030936
[5]   ZHOU J, ZHANG D, ZOU P, et al Retinex-based Laplacian pyramid method for image defogging[J]. IEEE Access, 2019, 7: 122459- 122472
doi: 10.1109/ACCESS.2019.2934981
[6]   HE K, SUN J, TANG X Single image haze removal using dark channel prior[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010, 33 (12): 2341- 2353
doi: 10.18535/ijecs/v5i1.12
[7]   CAI B, XU X, JIA K, et al DehazeNet: an end-to-end system for single image haze removal[J]. IEEE Transactions on Image Processing, 2016, 25 (11): 5187- 5198
doi: 10.1109/TIP.2016.2598681
[8]   LIU W, YAO R, QIU G A physics based generative adversarial network for single image defogging[J]. Image and Vision Computing, 2019, 92: 103815
doi: 10.1016/j.imavis.2019.10.001
[9]   CAI D, BAMISILE O, ZHANG W, et al Anti-occlusion multi-object surveillance based on improved deep learning approach and multi-feature enhancement for unmanned smart grid safety[J]. Energy Reports, 2023, 9: 594- 603
doi: 10.1016/j.egyr.2023.01.074
[10]   TANG T, ZHOU S, DENG Z, et al Vehicle detection in aerial images based on region convolutional neural networks and hard negative example mining[J]. Sensors, 2017, 17 (2): 336
doi: 10.3390/s17020336
[11]   SU J, WANG F, ZHUANG W An improved YOLOv7 tiny algorithm for vehicle and pedestrian detection with occlusion in autonomous driving[J]. Chinese Journal of Electronics, 2025, 34 (1): 282- 294
doi: 10.23919/cje.2023.00.256
[12]   AGHAEE F, FAZL-ERSI E, NOORI H MDSSD-MobV2: an embedded deconvolutional multispectral pedestrian detection based on SSD-MobileNetV2[J]. Multimedia Tools and Applications, 2024, 83 (15): 43801- 43829
doi: 10.1007/s11042-023-17188-7
[13]   ZHANG C, WANG H, CAI Y, et al TransFusion: multi-modal robust fusion for 3D object detection in foggy weather based on spatial vision transformer[J]. IEEE Transactions on Intelligent Transportation Systems, 2024, 25 (9): 10652- 10666
doi: 10.1109/TITS.2024.3420432
[14]   LI B, PENG X, WANG Z, et al. AOD-net: all-in-one dehazing network [C]// Proceedings of the IEEE International Conference on Computer Vision. Venice: IEEE, 2017: 4780–4788.
[15]   LIU W, ANGUELOV D, ERHAN D, et al. SSD: single shot MultiBox detector [C]// Computer Vision – ECCV 2016. Cham: Springer, 2016: 21–37.
[16]   CAO Z, LI J, FANG L, et al Research on efficient classification algorithm for coal and gangue based on improved MobilenetV3-small[J]. International Journal of Coal Preparation and Utilization, 2025, 45 (2): 437- 462
doi: 10.1080/19392699.2024.2353128
[17]   HE H, LI Z, TIAN G, et al Towards accurate dense pedestrian detection via occlusion-prediction aware label assignment and hierarchical-NMS[J]. Pattern Recognition Letters, 2023, 174: 78- 84
doi: 10.1016/j.patrec.2023.08.019
[18]   LIU Y, LI D AdaXod: a new adaptive and momental bound algorithm for training deep neural networks[J]. The Journal of Supercomputing, 2023, 79 (15): 17691- 17715
doi: 10.1007/s11227-023-05338-5
[19]   ROLNICK D, VEIT A, BELONGIE S, et al. Deep learning is robust to massive label noise [EB/OL]. (2017−05−30) [2025−08−05]. https://doi.org/10.48550/arXiv.1705.10694.
[20]   ARAZO E, ORTEDO D, ALBERT P, et al. Unsupervised label noise modeling and loss cor-rection [C]// International Conference on Machine Learning. Long Beach: Curran Associates, Inc. 2019: 312–321.
[21]   ZHANG Z, SABUNCU M. Generalized cross entropy loss for training deep neural networks with noisy labels [EB/OL]. (2018−12−03) [2025−08−05]. https://proceedings.neurips.cc/paper/2018/hash/f2925f97bc13ad2852a7a551802feea0-Abstract.html.
[22]   LI B, REN W, FU D, et al. REalistic single image dehazing (RESIDE): a benchmark for single image dehazing [DS/OL]. (2017−12−01) [2025−08−05]. https://github.com/Boyiliee/RESIDE-dataset-link.
[23]   ZHANG S, BENENSON R, SCHIELE B. CityPersons: a diverse dataset for pedestrian detection [DS/OL]. (2017−08−09) [2025−08−05]. https://www.cityscapes-dataset.com/citypersonsdownload/.
[24]   CORDTS M, OMRAN M, RAMOS S. The cityscapes dataset for semantic urban scene understanding [DS/OL]. (2016−06−01) [2025−08−05]. https://www.cityscapes-dataset.com/.
[25]   都雪静, 刘世新. 基于改进AOD-Net的交通场景去雾算法 [EB/OL]. (2025−06−12). https://kns.cnki.net/kcms/detail/23.1498.N.20250611.1938.008.html.
[26]   张壮, 孟祥瑞 基于DCT-YOLO的密集遮挡行人检测算法[J]. 宁夏师范大学学报, 2025, 46 (7): 73- 85
ZHANG Zhuang, MENG Xiangrui Dense occlusion pedestrian detection algorithm based on DCT-YOLO[J]. Journal of Ningxia Normal University, 2025, 46 (7): 73- 85
doi: 10.3969/j.issn.1674-1331.2025.07.007
[27]   SU J, WANG F, ZHUANG W An improved YOLOv7 tiny algorithm for vehicle and pedestrian detection with occlusion in autonomous driving[J]. Chinese Journal of Electronics, 2025, 34 (1): 282- 294
doi: 10.23919/cje.2023.00.256
[28]   CHENG L, ZHANG D, ZHENG Y Road object detection in foggy complex scenes based on improved YOLOv8[J]. IEEE Access, 2024, 12: 107420- 107430
doi: 10.1109/ACCESS.2024.3438612
[29]   SHI X, SONG A Defog YOLO for road object detection in foggy weather[J]. The Computer Journal, 2024, 67 (11): 3115- 3127
doi: 10.1093/comjnl/bxae074
[30]   LIU Z, YAN J, ZHANG J Research on a recognition algorithm for traffic signs in foggy environments based on image defogging and transformer[J]. Sensors, 2024, 24 (13): 4370
doi: 10.3390/s24134370
[31]   SU J, MAO S, ZHUANG W AOYOLO algorithm oriented vehicle and pedestrian detection in foggy weather[J]. Chinese Journal of Electronics, 2025, 34 (2): 661- 672
doi: 10.23919/cje.2023.00.280
[1] Xiaoan BAO,Enlin CHEN,Na ZHANG,Xiaomei TU,Biao WU,Qingqi ZHANG. 3D human pose estimation based on multi-scale encoder fusion[J]. Journal of ZheJiang University (Engineering Science), 2026, 60(3): 565-573.
[2] Guoyan LI,Penghui LI,Rong LIU,Yupeng MEI,Minghui ZHANG. Remote sensing road extraction by fusing multi-scale resolution and strip feature[J]. Journal of ZheJiang University (Engineering Science), 2026, 60(3): 585-593.
[3] Weiqun LUO,Jingwei LU,Jiadi WU,Yuying LIANG,Chuanpeng SHEN,Rui ZHU. Lightweight detection model for typical environmental terrain target in Tibetan Plateau[J]. Journal of ZheJiang University (Engineering Science), 2026, 60(3): 594-603.
[4] Yuyu MENG,Chuile KONG,Jiuyuan HUO,Zeyu WU. UAV small target detection algorithm based on reconstruction of YOLOv11[J]. Journal of ZheJiang University (Engineering Science), 2026, 60(2): 303-312.
[5] Jian XIAO,Xinze HE,Hongliang CHENG,Xiaoyuan YANG,Xin HU. Aerial small target detection algorithm based on multi-scale feature enhancement[J]. Journal of ZheJiang University (Engineering Science), 2026, 60(1): 19-31.
[6] Siyao ZHOU,Nan XIA,Jiahong JIANG. Pose-guided dual-branch network for clothing-changing person re-identification[J]. Journal of ZheJiang University (Engineering Science), 2026, 60(1): 71-80.
[7] Chaoqun DONG,Zhan WANG,Ping LIAO,Shuai XIE,Yujie RONG,Jingsong ZHOU. Lightweight YOLOv5s-OCG rail sleeper crack detection algorithm[J]. Journal of ZheJiang University (Engineering Science), 2025, 59(9): 1838-1845.
[8] Wenxin CHENG,Guanghui YAN,Wenwen CHANG,Baijing WU,Yaning HUANG. Channel-weighted multimodal feature fusion for EEG-based fatigue driving detection[J]. Journal of ZheJiang University (Engineering Science), 2025, 59(9): 1775-1783.
[9] Zhuguo ZHOU,Yujun LU,Liye LV. Improved YOLOv5s-based algorithm for printed circuit board defect detection[J]. Journal of ZheJiang University (Engineering Science), 2025, 59(8): 1608-1616.
[10] Yahong ZHAI,Yaling CHEN,Longyan XU,Yu GONG. Improved YOLOv8s lightweight small target detection algorithm of UAV aerial image[J]. Journal of ZheJiang University (Engineering Science), 2025, 59(8): 1708-1717.
[11] Jiarui FU,Zhaofei LI,Hao ZHOU,Wei HUANG. Camouflaged object detection based on Convnextv2 and texture-edge guidance[J]. Journal of ZheJiang University (Engineering Science), 2025, 59(8): 1718-1726.
[12] Shenchong LI,Xinhua ZENG,Chuanqu LIN. Multi-task environment perception algorithm for autonomous driving based on axial attention[J]. Journal of ZheJiang University (Engineering Science), 2025, 59(4): 769-777.
[13] Zhenli ZHANG,Xinkai HU,Fan LI,Zhicheng FENG,Zhichao CHEN. Semantic segmentation algorithm for multiscale remote sensing images based on CNN and Efficient Transformer[J]. Journal of ZheJiang University (Engineering Science), 2025, 59(4): 778-786.
[14] Dengfeng LIU,Wenjing GUO,Shihai CHEN. Content-guided attention-based lane detection network[J]. Journal of ZheJiang University (Engineering Science), 2025, 59(3): 451-459.
[15] Liming LIANG,Pengwei LONG,Jiaxin JIN,Renjie LI,Lu ZENG. Steel surface defect detection algorithm based on improved YOLOv8s[J]. Journal of ZheJiang University (Engineering Science), 2025, 59(3): 512-522.