首页 | 本学科首页   官方微博 | 高级检索  
     

多特征融合与几何卷积的机载LiDAR点云地物分类
引用本文:戴莫凡,邢帅,徐青,李鹏程,陈坤. 多特征融合与几何卷积的机载LiDAR点云地物分类[J]. 中国图象图形学报, 2022, 27(2): 574-585
作者姓名:戴莫凡  邢帅  徐青  李鹏程  陈坤
作者单位:战略支援部队信息工程大学地理空间信息学院, 郑州 450001
基金项目:国家自然科学基金项目(41876105,41371436)
摘    要:目的点云分类传统方法中大量依赖人工设计特征,缺乏深层次特征,难以进一步提高精度,基于深度学习的方法大部分利用结构化网络,转化为其他表征造成了3维空间结构信息的丢失,部分利用局部结构学习多层次特征的方法也因为忽略了机载数据的几何信息,难以实现精细分类。针对上述问题,本文提出了一种基于多特征融合几何卷积神经网络(multi-feature fusion and geometric convolutional neural network,MFFGCNN)的机载Li DAR(light detection and ranging)点云地物分类方法。方法提取并融合有效的浅层传统特征,并结合坐标尺度等预处理方法,称为APD模块(airporne laser scanning point cloud design module),在输入特征层面对典型地物有针对性地进行信息补充,来提高网络对大区域、低密度的机载Li DAR点云原始数据的适应能力和基础分类精度,基于多特征融合的几何卷积模块,称为FGC(multi-feature fusion and geometric convolution)算子,...

关 键 词:点云分类  机载Li DAR  PointNet++  深度学习  多特征融合  几何卷积网络(GCN)
收稿时间:2021-07-05
修稿时间:2021-08-19

Semantic segmentation of airborne LiDAR point cloud based on multi-feature fusion and geometric convolution
Dai Mofan,Xing Shuai,Xu Qing,Li Pengcheng,Chen Kun. Semantic segmentation of airborne LiDAR point cloud based on multi-feature fusion and geometric convolution[J]. Journal of Image and Graphics, 2022, 27(2): 574-585
Authors:Dai Mofan  Xing Shuai  Xu Qing  Li Pengcheng  Chen Kun
Affiliation:Institute of Geospatial Information, Strategic Support Force Information Engineering University, Zhengzhou 450001, China
Abstract:Objective Airborne laser scanning (ALS) offers a mature structure of point cloud data, which can represent complicated geometric information of the real world. Point cloud classification is a critical task in airborne laser detection and ranging applications, such as topographic mapping, power line detection, building reconstruction, etc. However, unavoidable reasons, such as complicated topographic conditions, sensor noise and sparse point cloud density, make ALS point cloud classification very difficult. Original point cloud classification is dependent on manual features, targeted classification conditions and parameter designation. Deep learning-based methods transform 3D point cloud into other representations such as 3D voxels, 2D images and octree structures based on structured networks at the cost of 3D spatial structure information loss. As the requirement for high point cloud density, the lack of adaptability for data and deep features caused in accuracy. Moreover, some networks make multi-modal data fusion or learn multi-level feature representations of points via local structures exploration, but the applications of geometric information of airborne data is constraint to achieve fine-grained classification for geometrically prominent and diverse ALS point cloud. In this paper, we propose a multi-feature fusion and geometric convolutional neural network (MFFGCNN) consisting of ALS data processing, multi-feature fusion and deep geometric features aggregation for point cloud classification. Method First, an ALS point cloud design module, called APD module, is constructed to organize point cloud structure by balancing classes and scale, partitioning point cloud and processing raw coordinates. Next, discriminative typical features are used to supplement the point cloud information at the input feature level. The applications of echo and intensity has been demonstrated as input along with the coordinates into the point-based network, which can preserve the 3D spatial characteristics while making full use of the effective point cloud features. Then, four types of geometric features of points and their K-nearest neighborhoods are calculated by dividing the neighborhood region at three different scales, with local spatial information. A geometric convolution module based on multi-feature fusion, called multi-feature fusion and geometric convolution(FGC) operator, is to encode the global and local spatial geometric structure of points. This design can obtain the hierarchical geometric structure of large area point clouds. At the end, our method aggregates the discriminative deep global and local geometric features in different levels and the input multi-class features into a hierarchical advanced semantic feature, which enables semantic segmentation of airborne LiDAR(light detection and ranging) point clouds by spatial up-sampling. Result The comparative analyses are based on the International Society for Photogrammetry and Remote Sensing (ISPRS) 3D labeling benchmark dataset. For the three typical ground features of buildings, ground and trees, the dataset is further divided on initial contained nine categories such as car, facade and shrub. The ALS point cloud is segmented into blocks of 50 m as a batch input into the network. Each feature is extracted from the layer and then processed using batch normalization. The batch number for training and evaluation is 16 and the maximum epoch during training is 200. The test is implemented via a NVIDIA GTX 2060Ti GPU. Mean intersection-over-union (mIoU), mean accuracy (mAcc), and overall accuracy (OA) are evaluated. The results of ablation experiments demonstrate that the FGC module can improve the global accuracy by 8%, which can effectively extract local geometric features. Compared with the 3D spatial coordinate-based method, the overall classification accuracy can improve by 15%. The relative elevation of the neighborhood can reflect elevation directional heterogeneity and ground points can achieve clear classification. Echo features can be validated for advantage in vegetation classification. The introduction of geometric features facilitates the distinction between building points and background points while ensuring the consistency of the building''s main body and clear contours. The four classes of geometric features are targeted at the curvature variation, edges, all consistency of the feature and the unique spherical characteristics of the vegetation. It illustrated further that the involvement of multiple classes of features in the semantic segmentation of airborne LiDAR point clouds. The visualization results also demonstrate a stronger model, especially in difficult situations such as buildings surrounded by tall trees and dense buildings with complex roof structures still achieving excellent performance, but there is a potential for improvement in the edges of features, and in the detailed parts of complex scenes. Conclusion The proposed MFFCGNN network combines the advantages of initial features and deep learning-based models. The demonstrated model can be implemented in 3D city modelling.
Keywords:point cloud classification  airborne LiDAR  PointNet++  deep learning  multi-feature fusion  geometric convolutional network(GCN)
点击此处可从《中国图象图形学报》浏览原始摘要信息
点击此处可从《中国图象图形学报》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号