|
|
Deep 3D point cloud classification network based on competitive attention fusion |
Han-juan CHEN1,2(),Fei-peng DA1,2,3,*(),Shao-yan GAI1,2 |
1. School of Automation, Southeast University, Nanjing 210096, China 2. Key Laboratory of Measurement and Control of Complex Systems of Engineering, Ministry of Education, Southeast University, Nanjing 210096, China 3. Shenzhen Research Institute, Southeast University, Shenzhen 518063, China |
|
|
Abstract A competitive attention fusion block that can be transferred to different classification networks was proposed, in order to improve the 3D point cloud deep network classification model’s ability to extract and express global features, and enhance the model’s robustness to noise interference. The global representation of multi-hierarchical features and the internal similarity of intermediate features were learned. The weights of the intermediate feature channels were re-allocated. The proposed block was embedded in the benchmark networks Pointnet++ and PointASNL for experiments. Results show that the proposed block is independent and transferable, focusing on the core and backbone features that are more conducive to 3D point cloud shape classification. Compared with the benchmark network, the proposed block enhances the model’s anti-interference ability to point cloud disturbance noise, outlier noise and random noise without decreasing the classification accuracy. The proposed method achieves the accuracy was 93.2%, 92.9%, 85.7%, 78.2%, 63.5% in the case of the number of random noises was 0, 10, 50, 100, 200, respectively. Compared with the traditional filtering method, end-to-end learning reduces the pre-processing steps and manual intervention process, and has better anti-noise performance.
|
Received: 04 January 2021
Published: 31 December 2021
|
|
Fund: 国家自然科学基金资助项目(51475092);江苏省前沿引领技术基础研究专项资助项目(BK20192004C);江苏省自然基金资助项目(BK20181269);深圳市科技创新委员会资助项目(JCYJ20180306174455080) |
Corresponding Authors:
Fei-peng DA
E-mail: 220181486@seu.edu.cn;dafp@seu.edu.cn
|
基于竞争注意力融合的深度三维点云分类网络
为了提高三维点云深度网络分类模型对全局特征的提取与表达能力,增强模型对噪声干扰的鲁棒性,提出可迁移应用于不同分类网络的竞争性注意力融合模块,学习多层级特征的全局表征和中间特征内在相似度,对中间特征通道权值重分配. 在基准网络Pointnet++和PointASNL中嵌入所提模块并进行实验,结果显示:所提模块具有独立性和可迁移性,聚焦更利于三维点云形状分类的核心骨干特征. 与基准网络相比,所提模块在保持分类精度稳定不下降的情况下,模型对点云扰动噪声、离群点噪声和随机噪声的抗干扰能力增强,在随机噪声数分别为0、10、50、100、200的情况下,准确度分别达到93.2%、92.9%、85.7%、78.2%、63.5%. 与传统滤波方法相比,端到端的学习减少预处理步骤和人工干预过程,同时具有更优的抗噪性能.
关键词:
点云物体分类,
三维点云,
深度学习,
神经网络,
注意力机制,
竞争性融合
|
|
[1] |
BU S, LIU Z, HAN J, et al Learning high-level feature by deep belief networks for 3-D model retrieval and recognition[J]. IEEE Transactions on Multimedia, 2014, 16 (8): 2154- 2167
doi: 10.1109/TMM.2014.2351788
|
|
|
[2] |
SU H, MAJI S, KALOGERAKIS E, et al. Multi-view convolutional neural networks for 3D shape recognition [C]// 2015 IEEE International Conference on Computer Vision. Santiago: IEEE, 2015: 945-953.
|
|
|
[3] |
WU Z, SONG S, KHOSLA A, et al. 3D shapeNets: a deep representation for volumetric shapes [C]// 2015 IEEE Conference on Computer Vision and Pattern Recognition. Boston: IEEE, 2015: 1912-1920.
|
|
|
[4] |
QI C R, SU H, MO K, et al. PointNet: deep learning on point sets for 3D classification and segmentation [C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition. Honolulu: IEEE, 2017: 77-85.
|
|
|
[5] |
QI C R, YI L, SU H, et al. PointNet++: deep hierarchical feature learning on point sets in a metric space [C]// Advances in Neural Information Processing Systems. Long Beach: MIT Press, 2017: 5099-5108.
|
|
|
[6] |
GUERRERO P, KLEIMAN Y, OVSJANIKOV M, et al PCPNET learning local shape properties from raw point clouds[J]. Computer Graphics Forum, 2018, 37 (2): 75- 85
doi: 10.1111/cgf.13343
|
|
|
[7] |
SHEN Y, FENG C, YANG Y, et al. Mining point cloud local structures by kernel correlation and graph pooling [C]// 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE, 2018: 4548-4557.
|
|
|
[8] |
LI J, CHEN B M, LEE G H. SO-Net: self-organizing network for point cloud analysis [C]// 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE, 2018: 9397-9406.
|
|
|
[9] |
QI C R, LIU W, WU C, et al. Frustum PointNets for 3D object detection from RGB-D data [C]// 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE, 2018: 918-927.
|
|
|
[10] |
LI Y, BU R, SUN M, et al. PointCNN: convolution on Χ-transformed points [C]// Advances in Neural Information Processing Systems. Montreal: MIT Press, 2018: 828-838.
|
|
|
[11] |
LIU Y, FAN B, MENG G, et al. DensePoint: learning densely contextual representation for efficient point cloud processing [C]// 2019 IEEE/CVF International Conference on Computer Vision. Seoul: IEEE, 2019: 5238-5247.
|
|
|
[12] |
LIU X, HAN Z, LIU Y S, et al. Point2Sequence: learning the shape representation of 3D point clouds with an attention-based sequence to sequence network [C]// Proceedings of the AAAI conference on Artificial Intelligence. Honolulu: AAAI, 2019: 8778-8785.
|
|
|
[13] |
KOMARICHEV A, ZHONG Z, HUA J. A-CNN: annularly convolutional neural networks on point clouds [C]// 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Long Beach: IEEE, 2019: 7413-7422.
|
|
|
[14] |
ZHAO H, JIANG L, FU C W, et al. PointWeb: enhancing local neighborhood features for point cloud processing [C]// 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Long Beach: IEEE, 2019: 5560-5568.
|
|
|
[15] |
WANG C, SAMARI B, SIDDIQI K. Local spectral graph convolution for point set feature learning [C]// 15th European Conference on Computer Vision. Munich: Springer, 2018: 56-71.
|
|
|
[16] |
TE G, HU W, GUO Z, et al. RGCNN: regularized graph CNN for point cloud segmentation [C]// Proceedings of the 26th ACM international conference on Multimedia. Seoul: ACM, 2018: 746-754.
|
|
|
[17] |
LANDRIEU L, SIMONOVSKY M. Large-scale point cloud semantic segmentation with superpoint graphs [C]// 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE, 2018: 4558-4567.
|
|
|
[18] |
VASWANI A, SHAZEER N, PARMAR N, et al. Attention is all you need [C]// Advances in Neural Information Processing Systems. Long Beach: MIT Press, 2017: 5998-6008.
|
|
|
[19] |
HU J, SHEN L, SUN G. Squeeze-and-excitation networks [C]// 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE, 2018: 7132-7141.
|
|
|
[20] |
YOU H, FENG Y, JI R, et al. PVNet: a joint convolutional network of point cloud and multi-view for 3D shape recognition [C]// Proceedings of the 26th ACM international conference on Multimedia. Seoul: ACM, 2018: 1310-1318.
|
|
|
[21] |
YAN X, ZHENG C, LI Z, et al. PointASNL: robust point clouds processing using nonlocal neural networks with adaptive sampling [C]// 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Seattle: IEEE, 2020: 5588-5597.
|
|
|
[22] |
HU Y, WEN G, LUO M, et al. Competitive inner-imaging squeeze and excitation for residual network [EB/OL]. (2018-12-23)[2020-12-29]. https://arxiv.org/abs/1807.08920.
|
|
|
[23] |
YI L, KIM V G, CEYLAN D, et al A scalable active framework for region annotation in 3D shape collections[J]. ACM Transactions on Graphics, 2016, 35 (6): 210
|
|
|
|
Viewed |
|
|
|
Full text
|
|
|
|
|
Abstract
|
|
|
|
|
Cited |
|
|
|
|
|
Shared |
|
|
|
|
|
Discussed |
|
|
|
|