Please wait a minute...
浙江大学学报(工学版)  2026, Vol. 60 Issue (4): 738-750    DOI: 10.3785/j.issn.1008-973X.2026.04.006
计算机技术     
改进的有雾图像中被遮挡车辆及行人识别算法
于天河(),王文龙(),刘镛,杨壮壮,侯善冲
哈尔滨理工大学 测控技术与通信工程学院,黑龙江 哈尔滨 150006
Improved algorithm for identifying occluded vehicles and pedestrians in foggy images
Tianhe YU(),Wenlong WANG(),Yong LIU,Zhuangzhuang YANG,Shanchong HOU
School of Measurement and Control Technology and Communication Engineering, Harbin University of Science and Technology, Harbin 150006, China
 全文: PDF(5503 KB)   HTML
摘要:

针对智能自动驾驶场景中目标遮挡与雾天干扰导致的目标检测精度下降问题,提出2项关键技术改进. 针对目标遮挡问题提出改进检测方法,以集成增强注意力机制的轻量化MobileNetV3_small作为SSD骨干特征提取网络,结合多尺度特征融合机制与自适应超参数Soft-NMS算法提升遮挡场景下的检测精度,通过改进自适应Focal Loss重构置信度损失函数,缓解正负样本不平衡及噪声标签敏感性问题. 针对雾天图像中目标模糊的问题提出改进轻量化AOD-Net去雾方法,通过构建基于深度可分离卷积的多尺度特征提取网络,优化跨层连接结构并引入边界增强模块,有效提升图像对比度、增强纹理细节. 通过联合损失函数对去雾网络与检测网络进行端到端协同优化,为有雾图像中的遮挡目标检测任务提供更可靠的优化路径. 实验结果表明,联合优化模型提升了雾天遮挡场景下的目标检测性能,以93.85%的准确率和47.61帧/s的检测速度实现了高效检测,并表现出优异的模型鲁棒性.

关键词: 被遮挡目标雾天图像特征融合目标检测特征提取    
Abstract:

Two key technical improvements were proposed to address the issue of decreased target detection accuracy caused by target occlusion and fog interference in intelligent autonomous driving scenarios. Firstly, an improved detection method was proposed for the problem of target occlusion. A lightweight MobileNetV3_small with an integrated enhanced attention mechanism was used as the SSD backbone feature extraction network. The detection accuracy in occlusion scenarios was enhanced by combining a multi-scale feature fusion mechanism and an adaptive hyperparameter Soft-NMS algorithm. The confidence loss function was reconstructed by improving the adaptive Focal Loss to alleviate the imbalance between positive and negative samples and the sensitivity to noisy labels. Secondly, an improved lightweight AOD-Net defogging method was proposed to address the problem of target blurriness in foggy images. The image contrast was effectively enhanced and texture details were improved, by constructing a multi-scale feature extraction network based on depthwise separable convolution, optimizing the cross-layer connection structure, and introducing a boundary enhancement module. Finally, an end-to-end collaborative optimization of the defogging network and the detection network was achieved through a joint loss function, providing a more reliable optimization path for the detection task of occluded targets in foggy images. Experimental results showed that the joint optimization model improved the target detection performance in foggy occlusion scenarios, achieving efficient detection with an accuracy of 93.85% and a detection speed of 47.61 frames per second, and demonstrating excellent model robustness.

Key words: occluded target    foggy image    feature fusion    target detection    feature extraction
收稿日期: 2025-07-26 出版日期: 2026-03-19
CLC:  TP 391.41  
作者简介: 于天河(1972—),男,教授,从事图像处理、识别和自动化检测以及智能仪器研究. orcid.org/0000-0001-8121-7620. E-mail:ythaa@163.com
服务  
把本文推荐给朋友
加入引用管理器
E-mail Alert
作者相关文章  
于天河
王文龙
刘镛
杨壮壮
侯善冲

引用本文:

于天河,王文龙,刘镛,杨壮壮,侯善冲. 改进的有雾图像中被遮挡车辆及行人识别算法[J]. 浙江大学学报(工学版), 2026, 60(4): 738-750.

Tianhe YU,Wenlong WANG,Yong LIU,Zhuangzhuang YANG,Shanchong HOU. Improved algorithm for identifying occluded vehicles and pedestrians in foggy images. Journal of ZheJiang University (Engineering Science), 2026, 60(4): 738-750.

链接本文:

https://www.zjujournals.com/eng/CN/10.3785/j.issn.1008-973X.2026.04.006        https://www.zjujournals.com/eng/CN/Y2026/V60/I4/738

图 1  AOD-Net去雾算法流程
图 2  ABMD-Net去雾网络模型
图 3  边界增强模块
图 4  SSD目标检测网络图
图 5  MsF-SSD-Net网络结构图
图 6  特征提取网络结构图
图 7  特征提取网络检测对比图
图 8  普通-深度可分离卷积双分支池化拼接融合模块
不同场景$ \beta $V/km
薄雾0.0050.6≤V<1.0
中雾0.0100.3≤V<0.6
重雾0.020V<0.3
表 1  去雾评价指标
模型薄雾中雾重雾FPS/(帧·s?1)
PSNRSSIMPSNRSSIMPSNRSSIM
AOD-Net19.490.882818.150.803217.790.766438.46
MSCNN[25]19.800.783418.150.821916.920.793927.78
GCANet[25]22.430.928218.110.757516.940.76557.46
GFN[25]21.150.923917.330.760916.930.738714.70
DehazeNet[25]20.460.875319.180.836518.890.827523.81
ABMD-Net24.020.939120.320.846120.030.835186.96
Improve AOD-Net[25]21.800.922817.790.851014.140.8024113.64
表 2  有雾图像上的PSNR和SSIM
图 9  网络模型去雾效果对比
模型FPS/
(帧·s?1)
F1PrRemAP
SoPoSe
SSD15.380.81350.83050.75630.77570.53250.3883
Faster
R-CNN
8.130.84130.87280.78230.80490.69260.5264
YOLOv1055.560.87680.91020.82260.83360.75540.5727
MsF-SSD-Net74.070.85450.89010.81890.92410.81360.5812
DCT-
YOLO[26]
0.84000.91600.77500.8870
improved
YOLOv7[27]
252.000.77000.78200.76900.8220
表 3  基于CityPersons数据集的各目标检测模型在遮挡场景下的性能对比
图 10  不同模型基于CityPersons 数据集的训练评价指标对比
模型mAPFPS/(帧·s?1)
清晰图像薄雾中雾重雾SoPoSe
SSD0.78190.71230.65560.53450.72850.47840.345414.92
Faster R-CNN0.81470.74870.68900.57890.75310.61320.48238.04
YOLOv100.83380.76340.70120.60560.77920.65730.495652.62
MsF-SSD-Net0.88760.84010.79340.66780.80520.71490.503471.43
lightweight YOLOv8[28]0.8100166.00
表 4  各目标检测模型在Foggy Cityscapes有雾数据集上的量化性能对比
图 11  经ABMD-Net去雾后的图像检测结果可视化对比图
模型mAPFPS/(帧·s?1)
薄雾中雾重雾SoPoSe
ABMD-Net+SSD0.75270.69560.59670.76450.52840.37639.74
ABMD-Net+Faster R-CNN0.80750.73900.62580.79840.68420.51665.38
ABMD-Net+YOLOv100.82340.78120.67340.80260.71440.542832.79
AOD-Net+MsF-SSD-Net0.88370.83650.71490.83320.74830.555625.03
ABMD-Net+MsF-SSD-Net0.91920.85340.73780.87780.76340.563345.21
Defog YOLO[29]0.8670
YOLOv5-Transformer[30]0.8280
AO YOLO[31]0.8910
ABMD-Net-MsF-SSD-Net0.93850.86830.78240.89520.78930.574847.61
表 5  经ABMD-Net去雾处理后的各模型检测性能量化对比表
图 12  Foggy Cityscapes 数据集下不同网络模型的训练损失对比
算法改动薄雾中雾重雾FPS/(帧·s?1)
PSNRSSIMPSNRSSIMPSNRSSIM
AOD-Net20.310.852118.150.803217.790.766438.17
AOD-Net120.820.865918.870.805317.920.798437.88
AOD-Net1(DW)20.730.855818.780.791317.860.7965117.65
AOD-Net1(DW)(Pyramid)22.430.897419.440.824618.840.8147107.53
ABMD-Net24.020.939120.320.846120.030.835186.96
表 6  AOD-Net算法消融实验结果
图 13  AOD-Net消融实验效果对比图
算法改动mAPF1Time/msFPS/(帧·s?1)PrReParams/106
SSD0.78140.813565.115.360.83050.7563138.36
MobileNetV3+SSD0.77830.79748.5117.650.82270.739425.54
MobileNetV3-EC+SSD0.85450.830213.275.760.85790.762826.73
MobileNetV3-EC+SSD10.87320.841714.867.570.86710.779328.47
MobileNetV3-EC+SSD1+Soft-NMS0.88270.854315.365.360.88280.792428.47
MsF-SSD-Net0.88750.855413.574.070.89010.808928.47
L-SSD[32]0.7380106.00
表 7  SSD算法消融实验结果表
1 TAHIR N U A, ZHANG Z, ASIM M, et al Object detection in autonomous vehicles under adverse weather: a review of traditional and deep learning approaches[J]. Algorithms, 2024, 17 (3): 103
doi: 10.3390/a17030103
2 ABDEL-ATY M, CAI Q, WU Y, et al Evaluation of automated emergency braking system’s avoidance of pedestrian crashes at intersections under occluded conditions within a virtual simulator[J]. Accident Analysis and Prevention, 2022, 176: 106797
doi: 10.1016/j.aap.2022.106797
3 SHETTY A, YU M, KURZHANSKIY A, et al Safety challenges for autonomous vehicles in the absence of connectivity[J]. Transportation Research Part C: Emerging Technologies, 2021, 128: 103133
doi: 10.1016/j.trc.2021.103133
4 MO Y, VIJAY R, RUFUS R, et al Enhanced perception for autonomous vehicles at obstructed intersections: an implementation of vehicle to infrastructure (V2I) collaboration[J]. Sensors, 2024, 24 (3): 936
doi: 10.3390/s24030936
5 ZHOU J, ZHANG D, ZOU P, et al Retinex-based Laplacian pyramid method for image defogging[J]. IEEE Access, 2019, 7: 122459- 122472
doi: 10.1109/ACCESS.2019.2934981
6 HE K, SUN J, TANG X Single image haze removal using dark channel prior[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010, 33 (12): 2341- 2353
doi: 10.18535/ijecs/v5i1.12
7 CAI B, XU X, JIA K, et al DehazeNet: an end-to-end system for single image haze removal[J]. IEEE Transactions on Image Processing, 2016, 25 (11): 5187- 5198
doi: 10.1109/TIP.2016.2598681
8 LIU W, YAO R, QIU G A physics based generative adversarial network for single image defogging[J]. Image and Vision Computing, 2019, 92: 103815
doi: 10.1016/j.imavis.2019.10.001
9 CAI D, BAMISILE O, ZHANG W, et al Anti-occlusion multi-object surveillance based on improved deep learning approach and multi-feature enhancement for unmanned smart grid safety[J]. Energy Reports, 2023, 9: 594- 603
doi: 10.1016/j.egyr.2023.01.074
10 TANG T, ZHOU S, DENG Z, et al Vehicle detection in aerial images based on region convolutional neural networks and hard negative example mining[J]. Sensors, 2017, 17 (2): 336
doi: 10.3390/s17020336
11 SU J, WANG F, ZHUANG W An improved YOLOv7 tiny algorithm for vehicle and pedestrian detection with occlusion in autonomous driving[J]. Chinese Journal of Electronics, 2025, 34 (1): 282- 294
doi: 10.23919/cje.2023.00.256
12 AGHAEE F, FAZL-ERSI E, NOORI H MDSSD-MobV2: an embedded deconvolutional multispectral pedestrian detection based on SSD-MobileNetV2[J]. Multimedia Tools and Applications, 2024, 83 (15): 43801- 43829
doi: 10.1007/s11042-023-17188-7
13 ZHANG C, WANG H, CAI Y, et al TransFusion: multi-modal robust fusion for 3D object detection in foggy weather based on spatial vision transformer[J]. IEEE Transactions on Intelligent Transportation Systems, 2024, 25 (9): 10652- 10666
doi: 10.1109/TITS.2024.3420432
14 LI B, PENG X, WANG Z, et al. AOD-net: all-in-one dehazing network [C]// Proceedings of the IEEE International Conference on Computer Vision. Venice: IEEE, 2017: 4780–4788.
15 LIU W, ANGUELOV D, ERHAN D, et al. SSD: single shot MultiBox detector [C]// Computer Vision – ECCV 2016. Cham: Springer, 2016: 21–37.
16 CAO Z, LI J, FANG L, et al Research on efficient classification algorithm for coal and gangue based on improved MobilenetV3-small[J]. International Journal of Coal Preparation and Utilization, 2025, 45 (2): 437- 462
doi: 10.1080/19392699.2024.2353128
17 HE H, LI Z, TIAN G, et al Towards accurate dense pedestrian detection via occlusion-prediction aware label assignment and hierarchical-NMS[J]. Pattern Recognition Letters, 2023, 174: 78- 84
doi: 10.1016/j.patrec.2023.08.019
18 LIU Y, LI D AdaXod: a new adaptive and momental bound algorithm for training deep neural networks[J]. The Journal of Supercomputing, 2023, 79 (15): 17691- 17715
doi: 10.1007/s11227-023-05338-5
19 ROLNICK D, VEIT A, BELONGIE S, et al. Deep learning is robust to massive label noise [EB/OL]. (2017−05−30) [2025−08−05]. https://doi.org/10.48550/arXiv.1705.10694.
20 ARAZO E, ORTEDO D, ALBERT P, et al. Unsupervised label noise modeling and loss cor-rection [C]// International Conference on Machine Learning. Long Beach: Curran Associates, Inc. 2019: 312–321.
21 ZHANG Z, SABUNCU M. Generalized cross entropy loss for training deep neural networks with noisy labels [EB/OL]. (2018−12−03) [2025−08−05]. https://proceedings.neurips.cc/paper/2018/hash/f2925f97bc13ad2852a7a551802feea0-Abstract.html.
22 LI B, REN W, FU D, et al. REalistic single image dehazing (RESIDE): a benchmark for single image dehazing [DS/OL]. (2017−12−01) [2025−08−05]. https://github.com/Boyiliee/RESIDE-dataset-link.
23 ZHANG S, BENENSON R, SCHIELE B. CityPersons: a diverse dataset for pedestrian detection [DS/OL]. (2017−08−09) [2025−08−05]. https://www.cityscapes-dataset.com/citypersonsdownload/.
24 CORDTS M, OMRAN M, RAMOS S. The cityscapes dataset for semantic urban scene understanding [DS/OL]. (2016−06−01) [2025−08−05]. https://www.cityscapes-dataset.com/.
25 都雪静, 刘世新. 基于改进AOD-Net的交通场景去雾算法 [EB/OL]. (2025−06−12). https://kns.cnki.net/kcms/detail/23.1498.N.20250611.1938.008.html.
26 张壮, 孟祥瑞 基于DCT-YOLO的密集遮挡行人检测算法[J]. 宁夏师范大学学报, 2025, 46 (7): 73- 85
ZHANG Zhuang, MENG Xiangrui Dense occlusion pedestrian detection algorithm based on DCT-YOLO[J]. Journal of Ningxia Normal University, 2025, 46 (7): 73- 85
doi: 10.3969/j.issn.1674-1331.2025.07.007
27 SU J, WANG F, ZHUANG W An improved YOLOv7 tiny algorithm for vehicle and pedestrian detection with occlusion in autonomous driving[J]. Chinese Journal of Electronics, 2025, 34 (1): 282- 294
doi: 10.23919/cje.2023.00.256
28 CHENG L, ZHANG D, ZHENG Y Road object detection in foggy complex scenes based on improved YOLOv8[J]. IEEE Access, 2024, 12: 107420- 107430
doi: 10.1109/ACCESS.2024.3438612
29 SHI X, SONG A Defog YOLO for road object detection in foggy weather[J]. The Computer Journal, 2024, 67 (11): 3115- 3127
doi: 10.1093/comjnl/bxae074
30 LIU Z, YAN J, ZHANG J Research on a recognition algorithm for traffic signs in foggy environments based on image defogging and transformer[J]. Sensors, 2024, 24 (13): 4370
doi: 10.3390/s24134370
31 SU J, MAO S, ZHUANG W AOYOLO algorithm oriented vehicle and pedestrian detection in foggy weather[J]. Chinese Journal of Electronics, 2025, 34 (2): 661- 672
doi: 10.23919/cje.2023.00.280
[1] 包晓安,陈恩琳,张娜,涂小妹,吴彪,张庆琪. 基于多尺度编码器融合的三维人体姿态估计算法[J]. 浙江大学学报(工学版), 2026, 60(3): 565-573.
[2] 李国燕,李鹏辉,刘榕,梅玉鹏,张明辉. 融合多尺度分辨率和带状特征的遥感道路提取[J]. 浙江大学学报(工学版), 2026, 60(3): 585-593.
[3] 雒伟群,陆敬蔚,吴佳缔,梁钰迎,申传鹏,朱睿. 藏区高原典型环境地形目标的轻量化检测模型[J]. 浙江大学学报(工学版), 2026, 60(3): 594-603.
[4] 孟昱煜,孔垂乐,火久元,武泽宇. 重构YOLOv11的无人机小目标检测算法[J]. 浙江大学学报(工学版), 2026, 60(2): 303-312.
[5] 肖剑,何昕泽,程鸿亮,杨小苑,胡欣. 基于多尺度特征增强的航拍小目标检测算法[J]. 浙江大学学报(工学版), 2026, 60(1): 19-31.
[6] 包晓安,彭书友,张娜,涂小妹,张庆琪,吴彪. 基于多方位感知深度融合检测头的目标检测算法[J]. 浙江大学学报(工学版), 2026, 60(1): 32-42.
[7] 周思瑶,夏楠,江佳鸿. 姿态引导的双分支换装行人重识别网络[J]. 浙江大学学报(工学版), 2026, 60(1): 71-80.
[8] 董超群,汪战,廖平,谢帅,荣玉杰,周靖淞. 轻量化YOLOv5s-OCG的轨枕裂纹检测算法[J]. 浙江大学学报(工学版), 2025, 59(9): 1838-1845.
[9] 程文鑫,闫光辉,常文文,吴佰靖,黄亚宁. 基于通道加权的多模态特征融合用于EEG疲劳驾驶检测[J]. 浙江大学学报(工学版), 2025, 59(9): 1775-1783.
[10] 周著国,鲁玉军,吕利叶. 基于改进YOLOv5s的印刷电路板缺陷检测算法[J]. 浙江大学学报(工学版), 2025, 59(8): 1608-1616.
[11] 翟亚红,陈雅玲,徐龙艳,龚玉. 改进YOLOv8s的轻量级无人机航拍小目标检测算法[J]. 浙江大学学报(工学版), 2025, 59(8): 1708-1717.
[12] 付家瑞,李兆飞,周豪,黄惟. 基于Convnextv2与纹理边缘引导的伪装目标检测[J]. 浙江大学学报(工学版), 2025, 59(8): 1718-1726.
[13] 徐慧智,王秀青. 基于车辆图像特征的前车距离与速度感知[J]. 浙江大学学报(工学版), 2025, 59(6): 1219-1232.
[14] 李沈崇,曾新华,林传渠. 基于轴向注意力的多任务自动驾驶环境感知算法[J]. 浙江大学学报(工学版), 2025, 59(4): 769-777.
[15] 张振利,胡新凯,李凡,冯志成,陈智超. 基于CNN和Efficient Transformer的多尺度遥感图像语义分割算法[J]. 浙江大学学报(工学版), 2025, 59(4): 778-786.