[go: up one dir, main page]

CN113239829B - Cross-dimension remote sensing data target identification method based on space occupation probability characteristics - Google Patents

Cross-dimension remote sensing data target identification method based on space occupation probability characteristics Download PDF

Info

Publication number
CN113239829B
CN113239829B CN202110550692.7A CN202110550692A CN113239829B CN 113239829 B CN113239829 B CN 113239829B CN 202110550692 A CN202110550692 A CN 202110550692A CN 113239829 B CN113239829 B CN 113239829B
Authority
CN
China
Prior art keywords
feature
point cloud
dimensional
input
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110550692.7A
Other languages
Chinese (zh)
Other versions
CN113239829A (en
Inventor
闫奕名
王文轩
宿南
汪子璐
冯收
赵春晖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Engineering University
Original Assignee
Harbin Engineering University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Engineering University filed Critical Harbin Engineering University
Priority to CN202110550692.7A priority Critical patent/CN113239829B/en
Publication of CN113239829A publication Critical patent/CN113239829A/en
Application granted granted Critical
Publication of CN113239829B publication Critical patent/CN113239829B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/176Urban or other man-made structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/08Indexing scheme for image data processing or generation, in general involving all processing steps from image acquisition to 3D model generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Geometry (AREA)
  • Computer Graphics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

基于空间占用概率特征的跨维度遥感数据目标识别方法,属于遥感图像目标识别技术领域。本发明是为了解决目前没有一种有效的跨维度特征描述方法能够实现遥感数据中同一地物目标的二维图像数据与三维点云数据的关联问题。本发明首先训练二维图像的空间占用概率特征提取网络和三维点云的空间占用概率特征提取网络,然后对遥感数据中的二维图像数据和三维点云数据进行同类特征提取,即空间占用概率特征,最终基于空间占用概率特征实现遥感数据中同一地物目标的二维和三维数据之间的跨维度目标识别,即实现同一地物目标的二维图像数据与三维点云数据的关联。主要用于遥感数据的目标识别。

Figure 202110550692

A cross-dimensional remote sensing data target recognition method based on space occupancy probability features belongs to the technical field of remote sensing image target recognition. The present invention is to solve the problem that there is currently no effective cross-dimensional feature description method that can realize the association between two-dimensional image data and three-dimensional point cloud data of the same object in remote sensing data. The invention firstly trains the space occupancy probability feature extraction network of two-dimensional images and the space occupancy probability feature extraction network of three-dimensional point cloud, and then extracts the same kind of features for the two-dimensional image data and three-dimensional point cloud data in the remote sensing data, namely the space occupancy probability Finally, based on the space occupancy probability feature, the cross-dimensional target recognition between the two-dimensional and three-dimensional data of the same object in the remote sensing data is realized, that is, the association between the two-dimensional image data of the same object and the three-dimensional point cloud data is realized. Mainly used for target recognition of remote sensing data.

Figure 202110550692

Description

基于空间占用概率特征的跨维度遥感数据目标识别方法Target recognition method of cross-dimensional remote sensing data based on spatial occupancy probability feature

技术领域technical field

本发明涉及遥感数据目标识别方法,属于遥感图像目标识别技术领域。The invention relates to a remote sensing data target recognition method, and belongs to the technical field of remote sensing image target recognition.

背景技术Background technique

长期以来,基于遥感图像的地物信息提取和识别技术在民生和国防领域有着广泛的应用需求。随着遥感领域的技术发展及应用需求提高,相较于传统可见光波段的相机所获取的二维图像数据,激光雷达等遥感扫描设备所获取的三维点云数据在立体空间中对遥感目标能够进行更加完善的三维描述,不管在民事还是军事领域都具有更重大的意义。因此,为充分利用激光雷达等设备及传统可见光波段的相机所获得的全部地物信息,需要将同一地物目标的二维图像数据与三维点云数据进行关联。由于二维图像和三维点云具有不同的维度特征,导致目前并没有有效的跨维度特征描述方法用于实现两者的关联,因此对上述跨维度遥感数据进行同类特征提取并依据该特征进行跨维度目标识别具有重要意义。For a long time, the extraction and recognition technology of ground object information based on remote sensing images has a wide range of application requirements in the fields of people's livelihood and national defense. With the development of technology in the field of remote sensing and the improvement of application requirements, compared with the two-dimensional image data obtained by traditional visible light band cameras, the three-dimensional point cloud data obtained by remote sensing scanning equipment such as lidar can perform remote sensing targets in three-dimensional space. A more complete three-dimensional description has greater significance in both civil and military fields. Therefore, in order to make full use of all the ground object information obtained by equipment such as lidars and cameras in the traditional visible light band, it is necessary to associate the two-dimensional image data of the same ground object with the three-dimensional point cloud data. Since two-dimensional images and three-dimensional point clouds have different dimensional features, there is currently no effective cross-dimensional feature description method to achieve the correlation between the two. Dimensional target recognition is of great significance.

发明内容SUMMARY OF THE INVENTION

本发明是为了解决目前没有一种有效的跨维度特征描述方法能够实现遥感数据中二维图像数据与三维点云数据的关联的问题。The present invention is to solve the problem that there is currently no effective cross-dimensional feature description method that can realize the association between two-dimensional image data and three-dimensional point cloud data in remote sensing data.

基于空间占用概率特征的跨维度遥感数据目标识别方法,包括以下步骤:The target recognition method of cross-dimensional remote sensing data based on space occupancy probability features includes the following steps:

S1:对二维图像遥感数据进行预处理:将二维图像遥感数据输入实例分割网络,并依据实例分割结果对遥感数据进行目标提取;S1: Preprocess the two-dimensional image remote sensing data: input the two-dimensional image remote sensing data into the instance segmentation network, and perform target extraction on the remote sensing data according to the instance segmentation result;

S2:对三维点云遥感数据进行预处理:将三维点云输入点云目标检测网络,依据目标检测结果对三维点云遥感数据进行目标分割;S2: Preprocess the 3D point cloud remote sensing data: input the 3D point cloud into the point cloud target detection network, and perform target segmentation on the 3D point cloud remote sensing data according to the target detection result;

S3:将S1处理后的图像输入二维图像的空间占用概率特征的深度学习网络,提取二维图像的空间占用概率Ftest-2DS3: Input the image processed by S1 into the deep learning network of the space occupancy probability feature of the two-dimensional image, and extract the space occupancy probability F test-2D of the two-dimensional image;

S4:将S2处理后的三维点云输入三维点云的空间占用概率特征的深度学习网络;提取三维点云的空间占用概率Ftest-3DS4: input the 3D point cloud processed by S2 into a deep learning network of the space occupancy probability feature of the 3D point cloud; extract the space occupancy probability F test-3D of the 3D point cloud;

S5:将S3和S4中得到的空间占用概率特征Ftest-2D和Ftest-3D输入分类器进行目标识别,实现二维图像数据与三维点云数据的关联。S5: Input the space occupancy probability features F test-2D and F test-3D obtained in S3 and S4 into the classifier for target recognition, so as to realize the association between the two-dimensional image data and the three-dimensional point cloud data.

进一步地,S1所述的实例分割网络采用PANET。Further, the instance segmentation network described in S1 adopts PANET.

进一步地,S2所述的点云目标检测网络采用3D-BONET。Further, the point cloud target detection network described in S2 adopts 3D-BONET.

进一步地,S3所述的二维图像的空间占用概率特征的深度学习网络为OccupancyNetwork-2D网络,即Onet-2D,其训练过程包括以下步骤:Further, the deep learning network of the space occupancy probability feature of the two-dimensional image described in S3 is the OccupancyNetwork-2D network, namely Onet-2D, and its training process includes the following steps:

S301、构建二维图像数据集Mpre-2D,二维图像数据集Mpre-2D包括一个二维图像训练数据集Mtr-2D和一个二维图像测试数据集Mtest-2DS301, constructing a two-dimensional image data set M pre-2D , the two-dimensional image data set M pre-2D includes a two-dimensional image training data set M tr-2D and a two-dimensional image test data set M test-2D ;

S302、训练Onet-2D:S302, training Onet-2D:

将二维图像训练数据集Mtr-2D中的二维图像数据输入Onet-2D,Onet-2D首先采用带有超强通道注意力模块ECA的RESNET18残差网络对输入的二维图像数据进行特征提取,得到1*256的特征F1The two-dimensional image data in the two-dimensional image training dataset M tr-2D is input into Onet-2D, and Onet-2D first uses the RESNET18 residual network with the super channel attention module ECA to characterize the input two-dimensional image data. Extraction to obtain a feature F 1 of 1*256;

其次,随机生成一个单位体积的采样点云立方体,将点云立方体中每个点的x、y、z坐标输入一个三层的mlp多层神经网络,并转置,得到256*N的特征F2Second, randomly generate a sampling point cloud cube of unit volume, input the x, y, and z coordinates of each point in the point cloud cube into a three-layer mlp multi-layer neural network, and transpose to obtain a feature F of 256*N 2 ;

然后将F1和F2分别输入至少一个条件批量标准化模块,所述的条件批量标准化模块即CBN模块;具体过程包括以下步骤:Then F 1 and F 2 are respectively input into at least one conditional batch normalization module, the conditional batch normalization module is the CBN module; the specific process includes the following steps:

将从二维图像提取到的1*256的特征F1输入mlp多层神经网络,得到N*256的特征F3,并与从三维点云中提取到的特征F2进行.*运算,得到特征F4,再将F4与二维图像特征F1通过mlp多层神经网络后得到N*256的特征F3进行相加运算,得到N*256的特征F5;再将F5进行最大池化操作,得到1*256的特征F6,再进行卷积操作和sigmoid操作得到最终的N*256的特征F7The 1*256 feature F 1 extracted from the 2D image is input into the mlp multi-layer neural network to obtain the N*256 feature F 3 , and the .* operation is performed with the feature F 2 extracted from the 3D point cloud to obtain Feature F 4 , then F 4 and two-dimensional image feature F 1 are passed through the mlp multi-layer neural network to obtain N*256 feature F 3 for addition operation to obtain N*256 feature F 5 ; then F 5 is maximized The pooling operation is performed to obtain a 1*256 feature F 6 , and then the convolution operation and the sigmoid operation are performed to obtain the final N*256 feature F 7 ;

当条件批量标准化模块大于一个时,将从二维图像提取到的1*256的特征F1输入mlp多层神经网络,得到N*256的特征F3,并与从前一个条件批量标准化模块得到的特征F7进行.*运算,得到特征F4,再将F4与二维图像特征F1通过mlp多层神经网络得到N*256的特征F3进行相加运算,得到N*256的特征F5;再将F5进行最大池化操作,得到1*256的特征F6,再进行卷积操作和sigmoid得到新的N*256的特征F7When the conditional batch normalization module is larger than one, the 1 *256 feature F1 extracted from the two-dimensional image is input into the mlp multi - layer neural network, and the N*256 feature F3 is obtained, which is compared with the one obtained from the previous conditional batch normalization module. The feature F 7 performs the .* operation to obtain the feature F 4 , and then the F 4 and the two-dimensional image feature F 1 are added through the mlp multi-layer neural network to obtain the N*256 feature F 3 , and the N*256 feature F is obtained. 5 ; then perform the maximum pooling operation on F 5 to obtain a 1*256 feature F 6 , and then perform a convolution operation and sigmoid to obtain a new N*256 feature F 7 ;

最后,将F7输入mlp多层神经网络即可得到N*3的点云形式的预测结果;将训练集中的根据地物目标实际尺寸绘制的三维模型进行处理,得到真值点云,将真值点云与预测结果进行比对并计算loss值;Finally, input F 7 into the mlp multi-layer neural network to obtain the prediction result in the form of N*3 point cloud; process the three-dimensional model drawn according to the actual size of the object in the training set, and obtain the true value point cloud. Compare the point cloud with the prediction result and calculate the loss value;

经过迭代最终完成训练得到训练好的Onet-2D。After iteration, the training is finally completed to obtain the trained Onet-2D.

进一步地,构建二维图像数据集Mpre-2D的过程包括以下步骤:Further, the process of constructing a two-dimensional image dataset M pre-2D includes the following steps:

将获得的二维图像遥感数据输入实例分割网络,并依据实例分割结果对遥感数据进行目标提取,得到每幅图中只有一个对象的二维图像数据集Mpre-2D,将二维图像数据集Mpre-2D分为二维图像训练数据集Mtr-2D和二维图像测试数据集Mtest-2D;训练数据集Mtr-2D中包含地物目标的二维图像以及对应的根据该地物目标实际尺寸绘制的三维模型,测试数据集Mtest-2D中只包含地物目标的二维图像。Input the obtained two-dimensional image remote sensing data into the instance segmentation network, and extract objects from the remote sensing data according to the instance segmentation results, and obtain a two-dimensional image data set M pre-2D with only one object in each image. M pre-2D is divided into a two-dimensional image training data set M tr-2D and a two-dimensional image test data set M test-2D ; the training data set M tr-2D contains two-dimensional images of ground objects and corresponding The three-dimensional model drawn by the actual size of the object target, the test data set M test-2D only contains the two-dimensional image of the object target.

进一步地,S4所述的三维点云的空间占用概率特征的深度学习网络为OccupancyNetwork-3D网络,即Onet-3D,其训练过程包括以下步骤:Further, the deep learning network of the space occupancy probability feature of the three-dimensional point cloud described in S4 is the OccupancyNetwork-3D network, namely Onet-3D, and its training process includes the following steps:

S401、构建三维点云数据集Mpre-3D,三维点云数据集Mpre-3D包括一个三维点云训练数据集Mtr-3D和一个三维点云测试数据集Mtest-3DS401, constructing a three-dimensional point cloud data set M pre-3D , the three-dimensional point cloud data set M pre-3D includes a three-dimensional point cloud training data set M tr-3D and a three-dimensional point cloud test data set M test-3D ;

S402、训练Onet-3D:S402. Training Onet-3D:

将三维点云训练数据集Mtr-3D中的三维点云数据输入Onet-3D;Onet-3D首先采用pointnet点云特征提取网络对输入的三维点云数据进行特征提取,得到得到1*256的特征f1Input the 3D point cloud data in the 3D point cloud training data set M tr-3D into Onet-3D; Onet-3D first uses the pointnet point cloud feature extraction network to extract features from the input 3D point cloud data, and obtain 1*256 feature f 1 ;

其次,随机生成一个单位体积的采样点云立方体,将点云立方体中每个点的x、y、z坐标输入一个三层的mlp多层神经网络,并转置,得到256*N的特征f2Second, randomly generate a sampling point cloud cube of unit volume, input the x, y, and z coordinates of each point in the point cloud cube into a three-layer mlp multi-layer neural network, and transpose to obtain a feature f of 256*N 2 ;

然后将f1和f2分别输入至少一个条件批量标准化模块,所述的条件批量标准化模块即CBN模块;具体过程包括以下步骤:Then f 1 and f 2 are respectively input into at least one conditional batch normalization module, the conditional batch normalization module is the CBN module; the specific process includes the following steps:

将从二维图像提取到的1*256的特征f1输入mlp多层神经网络,得到N*256的特征f3,并与从三维点云中提取到的特征f2进行.*运算,得到特征f4,再将f4与二维图像特征f1通过mlp多层神经网络得到N*256的特征f3进行相加运算,得到N*256的特征f5;再将f5进行最大池化操作,得到1*256的特征f6,再进行卷积操作和sigmoid操作得到最终的N*256的特征f7The 1*256 feature f 1 extracted from the two-dimensional image is input into the mlp multi-layer neural network to obtain the N*256 feature f 3 , and the .* operation is performed with the feature f 2 extracted from the three-dimensional point cloud to obtain Feature f 4 , and then add f 4 and two-dimensional image feature f 1 to obtain N*256 feature f 3 through the mlp multi-layer neural network to obtain N*256 feature f 5 ; then perform maximum pooling on f 5 1*256 feature f 6 is obtained, and then the convolution operation and sigmoid operation are performed to obtain the final N*256 feature f 7 ;

当条件批量标准化模块大于一个时,将从二维图像提取到的1*256的特征f1输入mlp多层神经网络,得到N*256的特征f3,并与从前一个条件批量标准化模块得到的特征f7进行.*运算,得到特征f4,再将f4与二维图像特征f1通过mlp多层神经网络得到N*256的特征f3进行相加运算,得到N*256的特征f5;再将f5进行最大池化操作,得到1*256的特征f6,再进行卷积操作和sigmoid得到新的N*256的特征f7When the conditional batch normalization module is larger than one, the 1 *256 feature f1 extracted from the two-dimensional image is input into the mlp multi-layer neural network, and the N*256 feature f3 is obtained, which is compared with the one obtained from the previous conditional batch normalization module. The feature f 7 performs the .* operation to obtain the feature f 4 , and then the f 4 and the two-dimensional image feature f 1 are added to the N*256 feature f 3 through the mlp multi-layer neural network to obtain the N*256 feature f 5 ; then perform the maximum pooling operation on f 5 to obtain a feature f 6 of 1*256, and then perform a convolution operation and sigmoid to obtain a new N*256 feature f 7 ;

最后,将f7输入mlp多层神经网络即可得到N*3的点云形式的预测结果;将训练集中的根据该地物目标实际尺寸绘制的三维模型进行处理,得到真值点云,将真值点云与预测结果进行比对得到loss值;Finally, input f 7 into the mlp multi-layer neural network to obtain the prediction result in the form of N*3 point cloud; process the three-dimensional model drawn according to the actual size of the object in the training set to obtain the true value point cloud, The loss value is obtained by comparing the true point cloud with the prediction result;

经过迭代最终完成训练得到训练好的Onet-3D。After iteration, the training is finally completed to obtain the trained Onet-3D.

进一步地,Occupancy Network-3D网络采用pointnet点云特征提取网络对输入的三维点云数据进行特征提取的过程包括以下步骤:Further, the process that the Occupancy Network-3D network adopts the pointnet point cloud feature extraction network to perform feature extraction on the input 3D point cloud data includes the following steps:

将输入网络的N*3真实点云数据通过input transform模块,然后再通过一个二层的mlp多层神经网络,得到N*64的特征F'1Pass the N*3 real point cloud data of the input network through the input transform module, and then pass through a two-layer mlp multi-layer neural network to obtain the N*64 feature F'1;

将特征F'1输入feature transform模块,然后再通过一个三层的mlp多层神经网络,得到N*1024的特征F'2Input the feature F' 1 into the feature transform module, and then pass through a three-layer mlp multi-layer neural network to obtain the feature F' 2 of N*1024;

对N*1024的特征F'2进行最大池化操作,得到1*1024的特征F'3,再通过一个二层的mlp多层神经网络,得到1*256的特征f1Perform a maximum pooling operation on the N*1024 feature F' 2 to obtain a 1*1024 feature F' 3 , and then pass a two-layer mlp multi-layer neural network to obtain a 1*256 feature f 1 .

进一步地,构建三维点云数据集Mpre-3D的过程包括以下步骤:Further, the process of constructing the 3D point cloud dataset M pre-3D includes the following steps:

将获得的三维点云遥感数据输入点云目标检测网络,依据目标检测结果对三维点云遥感数据进行目标分割,得到每个文件中只包含一个对象的三维点云数据集Mpre-3D,三维点云数据集Mpre-3D分为三维点云训练数据集Mtr-3D和三维点云测试数据集Mtest-3D。训练数据集Mtr-3D中包含地物目标的三维点云以及对应的根据该地物目标实际尺寸绘制的三维模型,测试数据集Mtest-3D中只包含地物目标的三维点云。Input the obtained 3D point cloud remote sensing data into the point cloud target detection network, and perform target segmentation on the 3D point cloud remote sensing data according to the target detection result, and obtain a 3D point cloud data set M pre-3D that contains only one object in each file. The point cloud dataset M pre-3D is divided into a 3D point cloud training dataset M tr-3D and a 3D point cloud testing dataset M test-3D . The training data set M tr-3D contains the 3D point cloud of the object and the corresponding 3D model drawn according to the actual size of the object, and the test data set M test-3D only contains the 3D point cloud of the object.

进一步地,S5所述的分类器采用pointnet++点云分类网络,采用pointnet++点云分类网络进行目标识别的过程包括以下步骤:Further, the classifier described in S5 adopts the pointnet++ point cloud classification network, and the process of using the pointnet++ point cloud classification network to perform target recognition includes the following steps:

(1)、将S3和S4得到的数组形式的空间占用概率转换成点云形式,并设置点云中点的数量为定值m,若点云中的点数>m则进行下采样操作,若点云中的点数<m则进行上采样操作。(1) Convert the space occupation probability of the array form obtained by S3 and S4 into a point cloud form, and set the number of points in the point cloud to a fixed value m, if the number of points in the point cloud > m, perform downsampling operation, if The upsampling operation is performed when the number of points in the point cloud < m.

(2)、将上述预处理后的点云数据输入到pointnet++点云分类网络中;依据最终提取到的1*k的特征进行分类,即实现了二维图像数据与三维点云数据的关联。(2) Input the preprocessed point cloud data into the pointnet++ point cloud classification network; classify according to the finally extracted 1*k features, that is, to realize the association between the two-dimensional image data and the three-dimensional point cloud data.

进一步地,所述的分类器的训练过程包括以下步骤:Further, the training process of the classifier includes the following steps:

S501、提取二维图像数据集Mpre-2D中的二维图像测试数据集Mtest-2D,以及三维点云数据集Mpre-3D中的三维点云测试数据集Mtest-3DS501, extracting the two-dimensional image test data set M test- 2D in the two-dimensional image data set M pre- 2D, and the three-dimensional point cloud test data set M test- 3D in the three-dimensional point cloud data set M pre- 3D;

S502、将二维图像测试数据集Mtest-2D输入到Onet-2D中,提取二维图像数据的空间占用概率特征,得到数组形式的空间占用概率Ftest-2DS502: Input the two-dimensional image test data set M test-2D into Onet-2D, extract the space occupation probability feature of the two-dimensional image data, and obtain the space occupation probability F test-2D in the form of an array.

将三维点云测试数据集Mtest-3D输入到Onet-3D中,提取三维点云数据的空间占用概率特征,得到数组形式的空间占用概率Ftest-3DInput the three-dimensional point cloud test data set M test-3D into Onet-3D, extract the space occupation probability feature of the three-dimensional point cloud data, and obtain the space occupation probability F test-3D in the form of an array;

S503、将二维图像中提取到的空间占用概率特征对应的点云作为输入数据输入到pointnet++网络中,并将提取到的特征用作类特征;将三维点云中提取到的空间占用概率点云作为目标数据输入到pointnet++网络中,将提取到的特征与类特征进行匹配,并计算准确率,反复迭代实现分类器的训练;S503. Input the point cloud corresponding to the space occupancy probability feature extracted from the two-dimensional image as input data into the pointnet++ network, and use the extracted feature as a class feature; use the space occupancy probability point extracted from the three-dimensional point cloud The cloud is input into the pointnet++ network as the target data, the extracted features are matched with the class features, the accuracy is calculated, and the classifier is trained iteratively;

或者,or,

将三维点云中提取到的空间占用概率特征对应的点云作为输入数据输入到pointnet++网络中,并将提取到的特征用作类特征;将二维图像中提取到的空间占用概率点云作为目标数据输入到pointnet++网络中,将提取到的特征与类特征进行匹配,并计算准确率,反复迭代实现分类器的训练。The point cloud corresponding to the space occupancy probability feature extracted from the 3D point cloud is input into the pointnet++ network as input data, and the extracted features are used as class features; the space occupancy probability point cloud extracted from the 2D image is used as the input data. The target data is input into the pointnet++ network, the extracted features are matched with the class features, the accuracy is calculated, and the classifier is trained iteratively.

有益效果:Beneficial effects:

本发明首先训练二维图像的空间占用概率特征提取网络和三维点云的空间占用概率特征提取网络,然后对遥感数据中的二维图像数据和三维点云数据进行同类特征提取,即空间占用概率特征,最终基于空间占用概率特征实现遥感数据中同一地物目标的二维和三维数据之间的跨维度目标识别,即实现同一地物目标的二维图像数据与三维点云数据的关联。本发明可以很好的解决目前不能对二维图像数据与三维点云数据进行有效关联的问题,关联准确率可以达到80%。The invention firstly trains the space occupancy probability feature extraction network of two-dimensional images and the space occupancy probability feature extraction network of three-dimensional point cloud, and then extracts the same kind of features for the two-dimensional image data and three-dimensional point cloud data in the remote sensing data, namely the space occupancy probability Finally, based on the space occupancy probability feature, the cross-dimensional target recognition between the two-dimensional and three-dimensional data of the same object in the remote sensing data is realized, that is, the association between the two-dimensional image data of the same object and the three-dimensional point cloud data is realized. The invention can well solve the problem that the two-dimensional image data and the three-dimensional point cloud data cannot be effectively correlated at present, and the correlation accuracy rate can reach 80%.

附图说明Description of drawings

图1为具体实施方式一的流程示意图;1 is a schematic flow chart of Embodiment 1;

图2是二维图像实例分割方法PANET的示意图;Fig. 2 is the schematic diagram of two-dimensional image instance segmentation method PANET;

图3是三维点云目标检测方法3D-BONET的示意图;3 is a schematic diagram of a three-dimensional point cloud target detection method 3D-BONET;

图4是基于二维图像的空间占用概率特征提取方法Occupancy Network-2D的网络示意图;Fig. 4 is the network schematic diagram of Occupancy Network-2D, a space occupancy probability feature extraction method based on two-dimensional images;

图5是基于三维点云的空间占用概率特征提取方法Occupancy Network-3D的网络示意图;Fig. 5 is the network schematic diagram of Occupancy Network-3D, a space occupancy probability feature extraction method based on three-dimensional point cloud;

图6是点云分类网络pointnet++的网络示意图。Figure 6 is a network diagram of the point cloud classification network pointnet++.

具体实施方式Detailed ways

具体实施方式一:结合图1说明本实施方式,Embodiment 1: This embodiment is described with reference to FIG. 1 ,

本实施方式所述的基于空间占用概率特征的跨维度遥感数据目标识别方法,包括以下步骤:The cross-dimensional remote sensing data target recognition method based on the space occupancy probability feature described in this embodiment includes the following steps:

步骤一:对二维图像遥感数据进行预处理。Step 1: Preprocess the two-dimensional image remote sensing data.

首先对获得的二维图像遥感数据进行预处理:将二维图像遥感数据输入实例分割网络,并依据实例分割结果对遥感数据进行目标提取,得到每幅图中只有一栋楼的二维图像数据集Mpre-2D,其中又分为二维图像训练数据集Mtr-2D和二维图像测试数据集Mtest-2D。训练数据集Mtr-2D中包含地物目标的二维图像以及对应的根据该建筑物实际尺寸绘制的三维模型,测试数据集Mtest-2D中只包含地物目标的二维图像。Firstly, preprocess the obtained two-dimensional image remote sensing data: input the two-dimensional image remote sensing data into the instance segmentation network, and extract the target from the remote sensing data according to the instance segmentation result, and obtain the two-dimensional image data of only one building in each picture. The set M pre-2D is further divided into a two-dimensional image training data set M tr-2D and a two-dimensional image testing data set M test-2D . The training data set M tr-2D contains the two-dimensional images of the ground objects and the corresponding three-dimensional models drawn according to the actual size of the building, and the test data set M test-2D only contains the two-dimensional images of the ground objects.

步骤二:对三维点云遥感数据进行预处理。Step 2: Preprocess the 3D point cloud remote sensing data.

对通过激光雷达等途径获得的三维点云遥感数据进行预处理:将三维点云输入点云目标检测网络,依据目标检测结果对三维点云遥感数据进行目标分割,得到每个文件中只包含一栋楼的点云数据的三维点云数据集Mpre-3D,其中又分为三维点云训练数据集Mtr-3D和三维点云测试数据集Mtest-3D。训练数据集Mtr-3D中包含地物目标的三维点云以及对应的根据该建筑物实际尺寸绘制的三维模型,测试数据集Mtest-3D中只包含地物目标的三维点云。Preprocess the 3D point cloud remote sensing data obtained through lidar and other means: input the 3D point cloud into the point cloud target detection network, and perform target segmentation on the 3D point cloud remote sensing data according to the target detection result, and obtain that each file contains only one file. The three-dimensional point cloud data set M pre-3D of the point cloud data of the building is further divided into a three-dimensional point cloud training data set M tr-3D and a three-dimensional point cloud test data set M test-3D . The training data set M tr-3D contains the 3D point cloud of the object and the corresponding 3D model drawn according to the actual size of the building, and the test data set M test-3D only contains the 3D point cloud of the object.

步骤三:训练二维图像的空间占用概率特征提取网络。Step 3: Train the spatial occupancy probability feature extraction network of the two-dimensional image.

将二维图像训练数据集Mtr-2D中的图像输入深度学习网络中,训练用于提取二维图像的空间占用概率特征的深度学习网络;Input the images in the two-dimensional image training dataset M tr-2D into the deep learning network, and train the deep learning network for extracting the spatial occupancy probability features of the two-dimensional images;

所述的二维图像训练数据集的训练数据包括建筑物的真实二维遥感图像以及根据该建筑物实际尺寸绘制的三维仿真模型。The training data of the two-dimensional image training data set includes the real two-dimensional remote sensing image of the building and the three-dimensional simulation model drawn according to the actual size of the building.

步骤四:训练三维点云的空间占用概率特征提取网络。Step 4: Train the spatial occupancy probability feature extraction network of the 3D point cloud.

将三维点云训练数据集Mtr-3D中的点云文件输入深度学习网络中,训练用于提取三维点云的空间占用概率特征的深度学习网络;Input the point cloud files in the 3D point cloud training data set M tr-3D into the deep learning network, and train the deep learning network for extracting the spatial occupancy probability features of the 3D point cloud;

所述的三维点云训练数据集的训练数据包括建筑物的真实三维点云数据以及根据该建筑物实际尺寸绘制的三维仿真模型;The training data of the three-dimensional point cloud training data set includes the real three-dimensional point cloud data of the building and the three-dimensional simulation model drawn according to the actual size of the building;

步骤五:基于二维图像的空间占用概率特征提取。Step 5: Feature extraction of space occupancy probability based on two-dimensional images.

将二维图像测试数据集Mtest-2D输入到步骤三中训练好的网络中,提取二维图像数据的空间占用概率特征,得到数组形式的空间占用概率Ftest-2DInput the two-dimensional image test data set M test-2D into the network trained in step 3, extract the space occupation probability feature of the two-dimensional image data, and obtain the space occupation probability F test-2D in the form of an array.

步骤六:基于三维点云的空间占用概率特征提取。Step 6: Feature extraction of space occupancy probability based on 3D point cloud.

将三维点云测试数据集Mtest-3D输入到步骤四中训练好的网络中,提取三维点云数据的空间占用概率特征,得到数组形式的空间占用概率Ftest-3DInput the 3D point cloud test data set M test-3D into the network trained in step 4, extract the space occupancy probability feature of the 3D point cloud data, and obtain the space occupancy probability F test-3D in the form of an array.

步骤七:跨维度目标识别。将步骤五和步骤六中得到的来自不同维度遥感数据的空间占用概率特征Ftest-2D和Ftest-3D输入分类器进行目标识别。Step 7: Cross-dimensional target recognition. The spatial occupancy probability features F test-2D and F test-3D obtained from remote sensing data of different dimensions obtained in steps 5 and 6 are input into the classifier for target recognition.

实际上,遥感数据预处理环节包括步骤一和步骤二两步:步骤一为二维图像遥感数据的预处理步骤;步骤二为三维点云遥感数据的预处理步骤。In fact, the preprocessing link of remote sensing data includes step 1 and step 2: step 1 is the preprocessing step of 2D image remote sensing data; step 2 is the preprocessing step of 3D point cloud remote sensing data.

步骤一中采用PANET方法对二维图像的实例进行分割。如图2所示,PANET方法的结构主要分为特征金字塔模块、动态特征池化模块和全连接层模块,以二维图像数据作为输入进行实例分割。实例分割网络使用的损失函数L包括分类误差、检测误差和分割误差:In step 1, the PANET method is used to segment the instances of the two-dimensional image. As shown in Figure 2, the structure of the PANET method is mainly divided into a feature pyramid module, a dynamic feature pooling module and a fully connected layer module, which takes two-dimensional image data as input for instance segmentation. The loss function L used by the instance segmentation network includes classification error, detection error and segmentation error:

L=Lcls+Lbox+Lmask L=L cls +L box +L mask

其中,对于每一个ROI,mask分支定义一个K*m*2维的矩阵表示K个不同的分类对于每一个m*m的区域,对于每一个类都有一个。对于每一个像素,都是用sigmod函数进行求相对熵,得到平均相对熵误差Lmask。对于每一个ROI,如果检测得到ROI属于哪一个分类,就只使用哪一个分支的相对熵误差作为误差值进行计算。Among them, for each ROI, the mask branch defines a K*m*2-dimensional matrix representing K different categories. For each m*m region, there is one for each class. For each pixel, the relative entropy is calculated using the sigmod function to obtain the average relative entropy error Lmask. For each ROI, if it is detected to which category the ROI belongs, only the relative entropy error of which branch is used as the error value for calculation.

步骤二中采用3D-BONET方法对三维点云进行目标检测。如图3所示,3D-BONET为对三维点云数据进行实例分割提供了一个新框架,以三维点云为输入进行目标检测。3D-BONET的损失函数定义为L:In step 2, the 3D-BONET method is used to detect objects on the 3D point cloud. As shown in Figure 3, 3D-BONET provides a new framework for instance segmentation on 3D point cloud data, and takes 3D point cloud as input for object detection. The loss function of 3D-BONET is defined as L:

L=Lsem+Lbbox+Lbbs+Lpmask L=L sem +L bb o x +L bbs +L pmask

其中in

Figure BDA0003069729370000071
Figure BDA0003069729370000071

Figure BDA0003069729370000072
Figure BDA0003069729370000072

前期网络训练包括步骤三和步骤四两步:步骤三为训练基于二维图像的空间占用概率特征提取网络的过程;步骤四为训练基于三维点云的空间占用概率特征提取网络的过程。The preliminary network training includes step 3 and step 4: step 3 is the process of training the network for spatial occupancy probability feature extraction based on 2D images; step 4 is the process of training the spatial occupancy probability feature extraction network based on 3D point cloud.

步骤三中采用Occupancy Network-2D(Onet-2D)方法,使用步骤一中得到的训练数据Mtr-2D来训练基于二维图像的空间占用概率特征提取网络,如图4所示。空间占用概率为理想情况下每个点是否为模型内的点的概率,用3D物体的占用函数fθ(pij,xi):R3→{0,1}来表示,并通过Onet-2D网络来得到这个3D函数。该神经网络给每个位置p分配一个在0到1之间的占用概率,相当于一个用于二分类的神经网络,而本发明关注的是对象表面的决策边界。Onet-2D参数训练的具体步骤为:In step 3, the Occupancy Network-2D (Onet-2D) method is adopted, and the training data M tr-2D obtained in step 1 is used to train a two-dimensional image-based spatial occupancy probability feature extraction network, as shown in Figure 4. The space occupancy probability is the probability of whether each point is a point in the model under ideal conditions. 2D network to get this 3D function. The neural network assigns each position p an occupancy probability between 0 and 1, which is equivalent to a neural network for binary classification, while the present invention focuses on the decision boundary of the object surface. The specific steps of Onet-2D parameter training are:

首先,将训练集中的二维图像数据输入网络,使用带有超强通道注意力模块ECA的RESNET18残差网络对输入的二维图像数据进行特征提取,得到1*256的特征F1。超强通道注意力模块ECA是一种避免了维度缩减、并有效捕获了跨通道交互的模块:将RESNET18残差网络提取到的H*W*C的特征在不降低维度的情况下进行逐通道全局平均池化,然后通过大小为k1的快速一维卷积实现每个通道及其k1个近邻的跨通道交互,其中内核大小k1代表本地跨通道交互的覆盖范围,即有多少个相近邻参与一个通道的注意力预测,输出同样为H*W*C的特征。First, input the 2D image data in the training set into the network, and use the RESNET18 residual network with the super channel attention module ECA to perform feature extraction on the input 2D image data to obtain a 1*256 feature F 1 . The super channel attention module ECA is a module that avoids dimension reduction and effectively captures cross-channel interactions: the H*W*C features extracted by the RESNET18 residual network are processed channel by channel without reducing the dimension. Global average pooling followed by fast 1D convolutions of size k 1 to achieve cross-channel interactions for each channel and its k 1 nearest neighbors, where the kernel size k 1 represents the coverage of local cross-channel interactions, i.e. how many Nearby neighbors participate in the attention prediction of a channel, and the output is also the feature of H*W*C.

其次,随机生成一个单位体积的采样点云立方体(这一步首先随机生成一个采样点云立方体,然后通过后续的训练步骤为每个采样点配置空间占用概率,逐渐将其训练为“三维仿真模型”的形状),将点云立方体中每个点的x、y、z坐标(N*3)输入一个三层的mlp多层神经网络(3→64→256),并转置,得到256*N的特征F2Second, randomly generate a sampling point cloud cube of unit volume (this step first randomly generates a sampling point cloud cube, and then configure the space occupancy probability for each sampling point through the subsequent training steps, and gradually train it as a "3D simulation model" shape), input the x, y, z coordinates (N*3) of each point in the point cloud cube into a three-layer mlp multi-layer neural network (3→64→256), and transpose to get 256*N the feature F 2 .

然后将上述两个特征分别输入至少一个条件批量标准化模块(CBN模块),优选为5个CBN。条件批量标准化模块(CBN模块):将从二维图像提取到的1*256的特征F1输入mlp多层神经网络,得到N*256的特征F3,并与从三维点云中提取到的特征F2进行.*运算(两个矩阵对应元素相乘),得到特征F4,再将F4与二维图像特征F1通过mlp多层神经网络得到N*256的特征F3进行相加运算,得到N*256的特征F5;再将F5进行最大池化操作,得到1*256的特征F6,再进行卷积操作和sigmoid操作得到最终的N*256的特征F7Then the above two features are respectively input into at least one conditional batch normalization module (CBN module), preferably 5 CBNs. Conditional batch normalization module (CBN module): The 1 *256 feature F1 extracted from the 2D image is input into the mlp multi - layer neural network, and the N*256 feature F3 is obtained, which is compared with the feature extracted from the 3D point cloud. The feature F 2 performs the .* operation (multiplying the corresponding elements of the two matrices) to obtain the feature F 4 , and then adds the F 4 and the two-dimensional image feature F 1 through the mlp multi-layer neural network to obtain the N*256 feature F 3 for addition. operation to obtain N*256 feature F 5 ; then perform maximum pooling operation on F 5 to obtain 1*256 feature F 6 , and then perform convolution operation and sigmoid operation to obtain final N*256 feature F 7 ;

当条件批量标准化模块大于一个时,将从二维图像提取到的1*256的特征F1输入mlp多层神经网络,得到N*256的特征F3,并与从前一个条件批量标准化模块得到的特征F7进行.*运算,得到特征F4,再将F4与二维图像特征F1通过mlp多层神经网络得到N*256的特征F3进行相加运算,得到N*256的特征F5;再将F5进行最大池化操作,得到1*256的特征F6,再进行卷积操作和sigmoid得到新的N*256的特征F7When the conditional batch normalization module is larger than one, the 1 *256 feature F1 extracted from the two-dimensional image is input into the mlp multi - layer neural network, and the N*256 feature F3 is obtained, which is compared with the one obtained from the previous conditional batch normalization module. The feature F 7 performs the .* operation to obtain the feature F 4 , and then the F 4 and the two-dimensional image feature F 1 are added through the mlp multi-layer neural network to obtain the N*256 feature F 3 , and the N*256 feature F is obtained. 5 ; then perform the maximum pooling operation on F 5 to obtain a 1*256 feature F 6 , and then perform a convolution operation and sigmoid to obtain a new N*256 feature F 7 ;

最后,将F7输入mlp多层神经网络即可得到N*3的点云形式的预测结果。将训练集中的根据该建筑物实际尺寸绘制的三维模型进行处理,得到该建筑物的真值点云,与预测结果进行比对即可得到loss值。Finally, input F 7 into the mlp multi-layer neural network to get the prediction result in the form of N*3 point cloud. The three-dimensional model drawn according to the actual size of the building in the training set is processed to obtain the true point cloud of the building, and the loss value can be obtained by comparing with the prediction result.

为了学习神经网络的参数,考虑在生成的单位体积点云立方体中随机采样点,对于第i个样本,采样K个点,然后评估这些位置的小批量损失LB(θ)如下所示:To learn the parameters of the neural network, consider randomly sampling points in the generated point cloud cube of unit volume, for the ith sample, sample K points, and then evaluate the mini-batch loss L B (θ) for these positions as follows:

Figure BDA0003069729370000081
Figure BDA0003069729370000081

其中fθ(pij,xi)是空间占用概率函数,以xi和pij作为输入,通过设定的阈值,判断第i个样本的第j个采样点是否为模型内的点,xi是第i个样本的观测值,pij是第i个样本的第j个采样点为模型内的点的概率,oij是点云的真实位置,L是计算交叉熵损失。where f θ (p ij , x i ) is the space occupancy probability function, with x i and p ij as inputs, through the set threshold, it is judged whether the j-th sampling point of the i-th sample is a point in the model, x i is the observed value of the i-th sample, p ij is the probability that the j-th sampling point of the i-th sample is a point in the model, o ij is the true position of the point cloud, and L is the calculated cross-entropy loss.

步骤四中采用Occupancy Network-3D(Onet-3D)方法,使用步骤二中得到的训练数据Mtr-3D来训练基于三维点云的空间占用概率特征提取网络,如图5所示。步骤四大体上的训练步骤与步骤三相同,不同之处在于,由于网络输入由二维图像变为三维点云,因此输入编码部分发生了变化:在步骤四中采用pointnet点云特征提取方法:In step 4, the Occupancy Network-3D (Onet-3D) method is adopted, and the training data M tr-3D obtained in step 2 is used to train the spatial occupancy probability feature extraction network based on 3D point cloud, as shown in Figure 5. The training steps in step four are the same as step three, the difference is that since the network input is changed from two-dimensional image to three-dimensional point cloud, the input coding part has changed: in step four, the pointnet point cloud feature extraction method is used:

首先,将输入网络的N*3真实点云数据通过input transform模块,然后再通过一个二层的mlp多层神经网络(3→64),得到N*64的特征F'1First, pass the N*3 real point cloud data of the input network through the input transform module, and then pass through a two-layer mlp multi-layer neural network (3→64) to obtain the N*64 feature F' 1 .

然后,将特征F'1输入feature transform模块,然后再通过一个三层的mlp多层神经网络(64→128→1024),得到N*1024的特征F'2。input transform模块和featuretransform模块的结构如图5所示,先将输入通过一个T-Net网络得到一个3*3或64*64的矩阵,然后与输入做矩阵乘法,得到N*3或N*64的特征,其功能均为加强特征提取能力。对N*1024的特征F'2进行最大池化操作,得到1*1024的特征F'3,再通过一个二层的mlp多层神经网络(512→256),得到1*256的特征。Then, the feature F' 1 is input into the feature transform module, and then through a three-layer mlp multi-layer neural network (64→128→1024), the N*1024 feature F' 2 is obtained. The structure of the input transform module and featuretransform module is shown in Figure 5. First, pass the input through a T-Net network to obtain a 3*3 or 64*64 matrix, and then perform matrix multiplication with the input to obtain N*3 or N*64 Its functions are to enhance the feature extraction ability. The maximum pooling operation is performed on the N*1024 feature F' 2 to obtain the 1*1024 feature F' 3 , and then a two-layer mlp multi-layer neural network (512→256) is used to obtain the 1*256 feature.

步骤五和步骤六分别使用步骤三和步骤四中训练好的Occupancy Network-2D和Occupancy Network-3D网络对不同维度的遥感数据进行空间占用概率特征提取,以二维图像遥感数据和三维点云遥感数据分别作为输入,输出数组形式的二维图像的空间占用概率特征和数组形式的三维点云的空间占用概率特征。Steps 5 and 6 use the Occupancy Network-2D and Occupancy Network-3D networks trained in Steps 3 and 4, respectively, to extract the spatial occupancy probability feature of remote sensing data of different dimensions, and use 2D image remote sensing data and 3D point cloud remote sensing. The data are used as input, respectively, and the space occupancy probability feature of the two-dimensional image in the form of an array and the space occupancy probability feature of the three-dimensional point cloud in the form of an array are output.

步骤七的分类器采用pointnet++点云分类网络实现:The classifier in step 7 is implemented by the pointnet++ point cloud classification network:

(1)、首先对待分类点云进行预处理。将步骤五和步骤六中得到的数组形式的不同维度数据的空间占用概率转换成点云形式,并设置点云中点的数量为定值m,若点云中的点数>m则进行下采样操作,若点云中的点数<m则进行上采样操作(将点云中的点单纯进行复制,直至达到目标数量m,对点云信息无影响)。(1) First, preprocess the point cloud to be classified. Convert the space occupancy probability of data of different dimensions in the form of arrays obtained in steps 5 and 6 into point cloud form, and set the number of points in the point cloud to a fixed value m, if the number of points in the point cloud > m, perform downsampling operation, if the number of points in the point cloud is less than m, the upsampling operation is performed (the points in the point cloud are simply copied until the target number m is reached, which has no effect on the point cloud information).

(2)、将上述预处理后的点云数据输入到图6所示的点云分类网络中。点云分类网络结构与步骤四中的点云编码器结构相同,依据最终提取到的1*k的特征进行分类,即实现了二维图像数据与三维点云数据的关联。(2) Input the above preprocessed point cloud data into the point cloud classification network shown in FIG. 6 . The structure of the point cloud classification network is the same as that of the point cloud encoder in step 4, and the classification is performed according to the finally extracted 1*k features, that is, the association between the two-dimensional image data and the three-dimensional point cloud data is realized.

该过程将将二维图像中提取到的空间占用概率特征对应的点云作为输入数据输入到点云分类网络pointnet++网络中,并将提取到的特征用作类特征;将三维点云中提取到的空间占用概率点云作为目标数据输入到pointnet++网络中,将提取到的特征与类特征进行匹配,并计算准确率,反复迭代实现分类器的训练。(这一步骤中二维图像中提取到的空间占用概率特征和三维点云中提取到的空间占用概率特征位置可互换,即均可作为输入数据和目标数据);In this process, the point cloud corresponding to the space occupancy probability feature extracted from the two-dimensional image is input as input data into the point cloud classification network pointnet++ network, and the extracted features are used as class features; The space occupancy probability point cloud is input into the pointnet++ network as the target data, the extracted features are matched with the class features, and the accuracy rate is calculated, and the classifier is trained iteratively. (In this step, the spatial occupancy probability feature extracted from the 2D image and the spatial occupancy probability feature extracted from the 3D point cloud can be interchanged, that is, both can be used as input data and target data);

分类器中使用的pointnet++点云特征提取网络对输入的三维点云数据进行特征提取的过程包括以下步骤:The pointnet++ point cloud feature extraction network used in the classifier performs feature extraction on the input 3D point cloud data, including the following steps:

将输入网络的N*3点云数据通过input transform模块,然后再通过一个二层的mlp多层神经网络,得到N*64的特征F”1Pass the N*3 point cloud data of the input network through the input transform module, and then pass through a two-layer mlp multi-layer neural network to obtain the N*64 feature F"1;

将特征F”1输入feature transform模块,然后再通过一个三层的mlp多层神经网络,得到N*1024的特征F”2Input the feature F" 1 into the feature transform module, and then pass through a three-layer mlp multi-layer neural network to obtain the feature F" 2 of N*1024;

对N*1024的特征F”2进行最大池化操作,得到1*1024的特征F”3,再通过一个三层的mlp多层神经网络,得到1*k的特征F”4;k为点云分类的类别数。Perform a maximum pooling operation on the N*1024 feature F" 2 to obtain a 1*1024 feature F" 3 , and then pass a three-layer mlp multi-layer neural network to obtain a 1*k feature F"4; k is a point The number of categories for cloud classification.

上述基于pointnet++点云分类网络的分类器通过对不同维度遥感数据进行跨维度特征提取得到的点云形式的空间占用概率进行特征提取,而后通过特征相似度排序实现遥感数据中同一地物不同维度数据的识别。在实际使用时,将二维图像数据经过步骤一的预处理,然后输入二维图像的空间占用概率特征提取网络提取二维图像的空间占用概率;同时将三维点云数据经过步骤二的预处理,然后输入三维点云的空间占用概率特征提取网络提取三维点云的空间占用概率;然后将二维图像的空间占用概率和三维点云的空间占用概率转换成点云形式,送入分类器进行分类,实现二维图像数据与三维点云数据的关联。The above classifier based on the pointnet++ point cloud classification network performs feature extraction on the space occupancy probability in the form of point clouds obtained by extracting cross-dimensional features from remote sensing data of different dimensions, and then realizes the data of different dimensions of the same object in the remote sensing data through feature similarity ranking. identification. In actual use, the two-dimensional image data is preprocessed in step 1, and then the spatial occupancy probability feature extraction network of the two-dimensional image is input to extract the space occupation probability of the two-dimensional image; at the same time, the three-dimensional point cloud data is preprocessed in step two. , and then input the space occupancy probability of the 3D point cloud. The feature extraction network extracts the space occupancy probability of the 3D point cloud; Classification to realize the association of 2D image data and 3D point cloud data.

本发明还可有其它多种实施例,在不背离本发明精神及其实质的情况下,本领域技术人员当可根据本发明作出各种相应的改变和变形,但这些相应的改变和变形都应属于本发明所附的权利要求的保护范围。The present invention can also have other various embodiments. Without departing from the spirit and essence of the present invention, those skilled in the art can make various corresponding changes and deformations according to the present invention, but these corresponding changes and deformations are all It should belong to the protection scope of the appended claims of the present invention.

Claims (9)

1.基于空间占用概率特征的跨维度遥感数据目标识别方法,其特征在于,包括以下步骤:1. A cross-dimensional remote sensing data target recognition method based on space occupancy probability feature, is characterized in that, comprises the following steps: S1:对二维图像遥感数据进行预处理:将二维图像遥感数据输入实例分割网络,并依据实例分割结果对遥感数据进行目标提取;S1: Preprocess the two-dimensional image remote sensing data: input the two-dimensional image remote sensing data into the instance segmentation network, and perform target extraction on the remote sensing data according to the instance segmentation result; S2:对三维点云遥感数据进行预处理:将三维点云输入点云目标检测网络,依据目标检测结果对三维点云遥感数据进行目标分割;S2: Preprocess the 3D point cloud remote sensing data: input the 3D point cloud into the point cloud target detection network, and perform target segmentation on the 3D point cloud remote sensing data according to the target detection result; S3:将S1处理后的图像输入二维图像的空间占用概率特征的深度学习网络,提取二维图像的空间占用概率Ftest-2DS3: Input the image processed by S1 into the deep learning network of the space occupancy probability feature of the two-dimensional image, and extract the space occupancy probability F test-2D of the two-dimensional image; S4:将S2处理后的三维点云输入三维点云的空间占用概率特征的深度学习网络;提取三维点云的空间占用概率Ftest-3DS4: input the 3D point cloud processed by S2 into a deep learning network of the space occupancy probability feature of the 3D point cloud; extract the space occupancy probability F test-3D of the 3D point cloud; S5:将S3和S4中得到的空间占用概率特征Ftest-2D和Ftest-3D输入分类器进行目标识别,实现二维图像数据与三维点云数据的关联;所述的分类器采用pointnet++点云分类网络,采用pointnet++点云分类网络进行目标识别的过程包括以下步骤:S5: Input the space occupancy probability features F test-2D and F test-3D obtained in S3 and S4 into the classifier for target recognition, so as to realize the association between the two-dimensional image data and the three-dimensional point cloud data; the classifier uses pointnet++ points Cloud classification network, the process of using pointnet++ point cloud classification network for target recognition includes the following steps: (1)、将S3和S4得到的数组形式的空间占用概率转换成点云形式,并设置点云中点的数量为定值m,若点云中的点数>m则进行下采样操作,若点云中的点数<m则进行上采样操作;(1) Convert the space occupation probability of the array form obtained by S3 and S4 into a point cloud form, and set the number of points in the point cloud to a fixed value m, if the number of points in the point cloud > m, perform downsampling operation, if If the number of points in the point cloud < m, the upsampling operation is performed; (2)、将上述预处理后的点云数据输入到pointnet++点云分类网络中;依据最终提取到的1*k的特征进行分类,即实现了二维图像数据与三维点云数据的关联。(2) Input the preprocessed point cloud data into the pointnet++ point cloud classification network; classify according to the finally extracted 1*k features, that is, to realize the association between the two-dimensional image data and the three-dimensional point cloud data. 2.根据权利要求1所述的基于空间占用概率特征的跨维度遥感数据目标识别方法,其特征在于,S1所述的实例分割网络采用PANET。2. The cross-dimensional remote sensing data target identification method based on the space occupancy probability feature according to claim 1, wherein the instance segmentation network described in S1 adopts PANET. 3.根据权利要求2所述的基于空间占用概率特征的跨维度遥感数据目标识别方法,其特征在于,S2所述的点云目标检测网络采用3D-BONET。3. The cross-dimensional remote sensing data target recognition method based on the space occupancy probability feature according to claim 2, wherein the point cloud target detection network described in S2 adopts 3D-BONET. 4.根据权利要求1、2或3所述的基于空间占用概率特征的跨维度遥感数据目标识别方法,其特征在于,S3所述的二维图像的空间占用概率特征的深度学习网络为OccupancyNetwork-2D网络,即Onet-2D,其训练过程包括以下步骤:4. the cross-dimensional remote sensing data target recognition method based on the space occupancy probability feature according to claim 1, 2 or 3, is characterized in that, the deep learning network of the space occupancy probability feature of the described two-dimensional image of S3 is OccupancyNetwork- 2D network, namely Onet-2D, its training process includes the following steps: S301、构建二维图像数据集Mpre-2D,二维图像数据集Mpre-2D包括一个二维图像训练数据集Mtr-2D和一个二维图像测试数据集Mtest-2DS301, constructing a two-dimensional image data set M pre-2D , the two-dimensional image data set M pre-2D includes a two-dimensional image training data set M tr-2D and a two-dimensional image test data set M test-2D ; S302、训练Onet-2D:S302, training Onet-2D: 将二维图像训练数据集Mtr-2D中的二维图像数据输入Onet-2D,Onet-2D首先采用带有超强通道注意力模块ECA的RESNET18残差网络对输入的二维图像数据进行特征提取,得到1*256的特征F1The two-dimensional image data in the two-dimensional image training dataset M tr-2D is input into Onet-2D, and Onet-2D first uses the RESNET18 residual network with the super channel attention module ECA to characterize the input two-dimensional image data. Extraction to obtain a feature F 1 of 1*256; 其次,随机生成一个单位体积的采样点云立方体,将点云立方体中每个点的x、y、z坐标输入一个三层的mlp多层神经网络,并转置,得到256*N的特征F2Second, randomly generate a sampling point cloud cube of unit volume, input the x, y, and z coordinates of each point in the point cloud cube into a three-layer mlp multi-layer neural network, and transpose to obtain a feature F of 256*N 2 ; 然后将F1和F2分别输入至少一个条件批量标准化模块,所述的条件批量标准化模块即CBN模块;具体过程包括以下步骤:Then F 1 and F 2 are respectively input into at least one conditional batch normalization module, the conditional batch normalization module is the CBN module; the specific process includes the following steps: 将从二维图像提取到的1*256的特征F1输入mlp多层神经网络,得到N*256的特征F3,并与从三维点云中提取到的特征F2进行.*运算,得到特征F4,再将F4与二维图像特征F1通过mlp多层神经网络后得到N*256的特征F3进行相加运算,得到N*256的特征F5;再将F5进行最大池化操作,得到1*256的特征F6,再进行卷积操作和sigmoid操作得到最终的N*256的特征F7The 1*256 feature F 1 extracted from the 2D image is input into the mlp multi-layer neural network to obtain the N*256 feature F 3 , and the .* operation is performed with the feature F 2 extracted from the 3D point cloud to obtain Feature F 4 , then F 4 and two-dimensional image feature F 1 are passed through the mlp multi-layer neural network to obtain N*256 feature F 3 for addition operation to obtain N*256 feature F 5 ; then F 5 is maximized The pooling operation is performed to obtain a 1*256 feature F 6 , and then the convolution operation and the sigmoid operation are performed to obtain the final N*256 feature F 7 ; 当条件批量标准化模块大于一个时,将从二维图像提取到的1*256的特征F1输入mlp多层神经网络,得到N*256的特征F3,并与从前一个条件批量标准化模块得到的特征F7进行.*运算,得到特征F4,再将F4与二维图像特征F1通过mlp多层神经网络得到N*256的特征F3进行相加运算,得到N*256的特征F5;再将F5进行最大池化操作,得到1*256的特征F6,再进行卷积操作和sigmoid得到新的N*256的特征F7When the conditional batch normalization module is larger than one, the 1 *256 feature F1 extracted from the two-dimensional image is input into the mlp multi - layer neural network, and the N*256 feature F3 is obtained, which is compared with the one obtained from the previous conditional batch normalization module. The feature F 7 performs the .* operation to obtain the feature F 4 , and then the F 4 and the two-dimensional image feature F 1 are added through the mlp multi-layer neural network to obtain the N*256 feature F 3 , and the N*256 feature F is obtained. 5 ; then perform the maximum pooling operation on F 5 to obtain a 1*256 feature F 6 , and then perform a convolution operation and sigmoid to obtain a new N*256 feature F 7 ; 最后,将F7输入mlp多层神经网络即可得到N*3的点云形式的预测结果;将训练集中的根据地物目标实际尺寸绘制的三维模型进行处理,得到真值点云,将真值点云与预测结果进行比对并计算loss值;Finally, input F 7 into the mlp multi-layer neural network to obtain the prediction result in the form of N*3 point cloud; process the three-dimensional model drawn according to the actual size of the object in the training set, and obtain the true value point cloud. Compare the point cloud with the prediction result and calculate the loss value; 经过迭代最终完成训练得到训练好的Onet-2D。After iteration, the training is finally completed to obtain the trained Onet-2D. 5.根据权利要求4所述的基于空间占用概率特征的跨维度遥感数据目标识别方法,其特征在于,构建二维图像数据集Mpre-2D的过程包括以下步骤:5. the cross-dimensional remote sensing data target recognition method based on space occupancy probability feature according to claim 4, is characterized in that, the process of constructing two-dimensional image data set M pre-2D comprises the following steps: 将获得的二维图像遥感数据输入实例分割网络,并依据实例分割结果对遥感数据进行目标提取,得到每幅图中只有一个对象的二维图像数据集Mpre-2D,将二维图像数据集Mpre-2D分为二维图像训练数据集Mtr-2D和二维图像测试数据集Mtest-2D;训练数据集Mtr-2D中包含地物目标的二维图像以及对应的根据该地物目标实际尺寸绘制的三维模型,测试数据集Mtest-2D中只包含地物目标的二维图像。Input the obtained two-dimensional image remote sensing data into the instance segmentation network, and extract objects from the remote sensing data according to the instance segmentation results, and obtain a two-dimensional image data set M pre-2D with only one object in each image. M pre-2D is divided into a two-dimensional image training data set M tr-2D and a two-dimensional image test data set M test-2D ; the training data set M tr-2D contains two-dimensional images of ground objects and corresponding The three-dimensional model drawn by the actual size of the object target, the test data set M test-2D only contains the two-dimensional image of the object target. 6.根据权利要求4所述的基于空间占用概率特征的跨维度遥感数据目标识别方法,其特征在于,S4所述的三维点云的空间占用概率特征的深度学习网络为OccupancyNetwork-3D网络,即Onet-3D,其训练过程包括以下步骤:6. the cross-dimensional remote sensing data target identification method based on space occupancy probability feature according to claim 4, is characterized in that, the deep learning network of the space occupancy probability feature of the described three-dimensional point cloud of S4 is OccupancyNetwork-3D network, namely Onet-3D, its training process includes the following steps: S401、构建三维点云数据集Mpre-3D,三维点云数据集Mpre-3D包括一个三维点云训练数据集Mtr-3D和一个三维点云测试数据集Mtest-3DS401 , constructing a three-dimensional point cloud data set M pre-3D , where the three-dimensional point cloud data set M pre-3D includes a three-dimensional point cloud training data set M tr-3D and a three-dimensional point cloud test data set M test-3D ; S402、训练Onet-3D:S402. Training Onet-3D: 将三维点云训练数据集Mtr-3D中的三维点云数据输入Onet-3D;Onet-3D首先采用pointnet点云特征提取网络对输入的三维点云数据进行特征提取,得到1*256的特征f1Input the 3D point cloud data in the 3D point cloud training data set M tr-3D into Onet-3D; Onet-3D first uses the pointnet point cloud feature extraction network to extract features from the input 3D point cloud data, and obtain 1*256 features f 1 ; 其次,随机生成一个单位体积的采样点云立方体,将点云立方体中每个点的x、y、z坐标输入一个三层的mlp多层神经网络,并转置,得到256*N的特征f2Second, randomly generate a sampling point cloud cube of unit volume, input the x, y, and z coordinates of each point in the point cloud cube into a three-layer mlp multi-layer neural network, and transpose to obtain a feature f of 256*N 2 ; 然后将f1和f2分别输入至少一个条件批量标准化模块,所述的条件批量标准化模块即CBN模块;具体过程包括以下步骤:Then f 1 and f 2 are respectively input into at least one conditional batch normalization module, the conditional batch normalization module is the CBN module; the specific process includes the following steps: 将从二维图像提取到的1*256的特征f1输入mlp多层神经网络,得到N*256的特征f3,并与从三维点云中提取到的特征f2进行.*运算,得到特征f4,再将f4与二维图像特征f1通过mlp多层神经网络得到N*256的特征f3进行相加运算,得到N*256的特征f5;再将f5进行最大池化操作,得到1*256的特征f6,再进行卷积操作和sigmoid操作得到最终的N*256的特征f7The 1*256 feature f 1 extracted from the two-dimensional image is input into the mlp multi-layer neural network, and the N*256 feature f 3 is obtained, and the .* operation is performed with the feature f 2 extracted from the three-dimensional point cloud to obtain Feature f 4 , and then add f 4 and two-dimensional image feature f 1 through the mlp multi-layer neural network to obtain N*256 feature f 3 to obtain N*256 feature f 5 ; then perform maximum pooling on f 5 1*256 feature f 6 is obtained, and then convolution operation and sigmoid operation are performed to obtain the final N*256 feature f 7 ; 当条件批量标准化模块大于一个时,将从二维图像提取到的1*256的特征f1输入mlp多层神经网络,得到N*256的特征f3,并与从前一个条件批量标准化模块得到的特征f7进行.*运算,得到特征f4,再将f4与二维图像特征f1通过mlp多层神经网络得到N*256的特征f3进行相加运算,得到N*256的特征f5;再将f5进行最大池化操作,得到1*256的特征f6,再进行卷积操作和sigmoid得到新的N*256的特征f7When the conditional batch normalization module is larger than one, the 1 *256 feature f1 extracted from the two-dimensional image is input into the mlp multi-layer neural network, and the N*256 feature f3 is obtained, which is compared with the one obtained from the previous conditional batch normalization module. The feature f 7 performs the .* operation to obtain the feature f 4 , and then the f 4 and the two-dimensional image feature f 1 are added to the N*256 feature f 3 through the mlp multi-layer neural network to obtain the N*256 feature f 5 ; then perform the maximum pooling operation on f 5 to obtain a feature f 6 of 1*256, and then perform a convolution operation and sigmoid to obtain a new N*256 feature f 7 ; 最后,将f7输入mlp多层神经网络即可得到N*3的点云形式的预测结果;将训练集中的根据该地物目标实际尺寸绘制的三维模型进行处理,得到真值点云,将真值点云与预测结果进行比对得到loss值;Finally, input f 7 into the mlp multi-layer neural network to obtain the prediction result in the form of N*3 point cloud; process the three-dimensional model drawn according to the actual size of the object in the training set to obtain the true value point cloud, The loss value is obtained by comparing the true point cloud with the prediction result; 经过迭代最终完成训练得到训练好的Onet-3D。After iteration, the training is finally completed to obtain the trained Onet-3D. 7.根据权利要求6所述的基于空间占用概率特征的跨维度遥感数据目标识别方法,其特征在于,Occupancy Network-3D网络采用pointnet点云特征提取网络对输入的三维点云数据进行特征提取的过程包括以下步骤:7. The cross-dimensional remote sensing data target recognition method based on space occupancy probability feature according to claim 6, is characterized in that, Occupancy Network-3D network adopts pointnet point cloud feature extraction network to carry out feature extraction to the input three-dimensional point cloud data. The process includes the following steps: 将输入网络的N*3真实点云数据通过input transform模块,然后再通过一个二层的mlp多层神经网络,得到N*64的特征F'1Pass the N*3 real point cloud data of the input network through the input transform module, and then pass through a two-layer mlp multi-layer neural network to obtain the N*64 feature F'1; 将特征F'1输入feature transform模块,然后再通过一个三层的mlp多层神经网络,得到N*1024的特征F'2Input the feature F' 1 into the feature transform module, and then pass through a three-layer mlp multi-layer neural network to obtain the feature F' 2 of N*1024; 对N*1024的特征F'2进行最大池化操作,得到1*1024的特征F'3,再通过一个二层的mlp多层神经网络,得到1*256的特征f1Perform a maximum pooling operation on the N*1024 feature F' 2 to obtain a 1*1024 feature F' 3 , and then pass a two-layer mlp multi-layer neural network to obtain a 1*256 feature f 1 . 8.根据权利要求7所述的基于空间占用概率特征的跨维度遥感数据目标识别方法,其特征在于,构建三维点云数据集Mpre-3D的过程包括以下步骤:8. the cross-dimensional remote sensing data target identification method based on space occupancy probability feature according to claim 7, is characterized in that, the process of constructing three-dimensional point cloud data set M pre-3D comprises the following steps: 将获得的三维点云遥感数据输入点云目标检测网络,依据目标检测结果对三维点云遥感数据进行目标分割,得到每个文件中只包含一个对象的三维点云数据集Mpre-3D,三维点云数据集Mpre-3D分为三维点云训练数据集Mtr-3D和三维点云测试数据集Mtest-3D;训练数据集Mtr-3D中包含地物目标的三维点云以及对应的根据该地物目标实际尺寸绘制的三维模型,测试数据集Mtest-3D中只包含地物目标的三维点云。Input the obtained 3D point cloud remote sensing data into the point cloud target detection network, and perform target segmentation on the 3D point cloud remote sensing data according to the target detection result, and obtain a 3D point cloud data set M pre-3D that contains only one object in each file. The point cloud data set M pre-3D is divided into a three-dimensional point cloud training data set M tr-3D and a three-dimensional point cloud test data set M test-3D ; the training data set M tr-3D contains the three-dimensional point cloud of the ground object and the corresponding The 3D model drawn according to the actual size of the object, the test data set M test-3D only contains the 3D point cloud of the object. 9.根据权利要求1所述的基于空间占用概率特征的跨维度遥感数据目标识别方法,其特征在于,所述的分类器的训练过程包括以下步骤:9. The cross-dimensional remote sensing data target recognition method based on space occupancy probability feature according to claim 1, is characterized in that, the training process of described classifier comprises the following steps: S501、提取二维图像数据集Mpre-2D中的二维图像测试数据集Mtest-2D,以及三维点云数据集Mpre-3D中的三维点云测试数据集Mtest-3DS501, extracting the two-dimensional image test data set M test- 2D in the two-dimensional image data set M pre- 2D, and the three-dimensional point cloud test data set M test- 3D in the three-dimensional point cloud data set M pre- 3D; S502、将二维图像测试数据集Mtest-2D输入到Onet-2D中,提取二维图像数据的空间占用概率特征,得到数组形式的空间占用概率Ftest-2DS502, input the two-dimensional image test data set M test-2D into Onet-2D, extract the space occupation probability feature of two-dimensional image data, obtain the space occupation probability F test-2D of array form; 将三维点云测试数据集Mtest-3D输入到Onet-3D中,提取三维点云数据的空间占用概率特征,得到数组形式的空间占用概率Ftest-3DInput the three-dimensional point cloud test data set M test-3D into Onet-3D, extract the space occupation probability feature of the three-dimensional point cloud data, and obtain the space occupation probability F test-3D in the form of an array; S503、将二维图像中提取到的空间占用概率特征对应的点云作为输入数据输入到pointnet++网络中,并将提取到的特征用作类特征;将三维点云中提取到的空间占用概率点云作为目标数据输入到pointnet++网络中,将提取到的特征与类特征进行匹配,并计算准确率,反复迭代实现分类器的训练;S503. Input the point cloud corresponding to the space occupancy probability feature extracted from the two-dimensional image as input data into the pointnet++ network, and use the extracted feature as a class feature; use the space occupancy probability point extracted from the three-dimensional point cloud The cloud is input into the pointnet++ network as the target data, the extracted features are matched with the class features, the accuracy is calculated, and the classifier is trained iteratively; 或者,or, 将三维点云中提取到的空间占用概率特征对应的点云作为输入数据输入到pointnet++网络中,并将提取到的特征用作类特征;将二维图像中提取到的空间占用概率点云作为目标数据输入到pointnet++网络中,将提取到的特征与类特征进行匹配,并计算准确率,反复迭代实现分类器的训练。The point cloud corresponding to the space occupancy probability feature extracted from the 3D point cloud is input into the pointnet++ network as input data, and the extracted features are used as class features; the space occupancy probability point cloud extracted from the 2D image is used as the input data. The target data is input into the pointnet++ network, the extracted features are matched with the class features, the accuracy is calculated, and the classifier is trained iteratively.
CN202110550692.7A 2021-05-17 2021-05-17 Cross-dimension remote sensing data target identification method based on space occupation probability characteristics Active CN113239829B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110550692.7A CN113239829B (en) 2021-05-17 2021-05-17 Cross-dimension remote sensing data target identification method based on space occupation probability characteristics

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110550692.7A CN113239829B (en) 2021-05-17 2021-05-17 Cross-dimension remote sensing data target identification method based on space occupation probability characteristics

Publications (2)

Publication Number Publication Date
CN113239829A CN113239829A (en) 2021-08-10
CN113239829B true CN113239829B (en) 2022-10-04

Family

ID=77137890

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110550692.7A Active CN113239829B (en) 2021-05-17 2021-05-17 Cross-dimension remote sensing data target identification method based on space occupation probability characteristics

Country Status (1)

Country Link
CN (1) CN113239829B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113763539B (en) * 2021-09-09 2023-04-07 哈尔滨工业大学(深圳) Implicit function three-dimensional reconstruction method based on image and three-dimensional input
CN114266967B (en) * 2021-12-15 2024-11-15 哈尔滨工程大学 Target recognition method for cross-source remote sensing data based on signed distance feature

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103984936A (en) * 2014-05-29 2014-08-13 中国航空无线电电子研究所 Multi-sensor multi-feature fusion recognition method for three-dimensional dynamic target recognition
CN104881671A (en) * 2015-05-21 2015-09-02 电子科技大学 High resolution remote sensing image local feature extraction method based on 2D-Gabor
CN111027401A (en) * 2019-11-15 2020-04-17 电子科技大学 An end-to-end object detection method for camera and lidar fusion
CN111259733A (en) * 2020-01-02 2020-06-09 武汉理工大学 A method and device for ship recognition based on point cloud images
CN111626217A (en) * 2020-05-28 2020-09-04 宁波博登智能科技有限责任公司 Target detection and tracking method based on two-dimensional picture and three-dimensional point cloud fusion

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8918209B2 (en) * 2010-05-20 2014-12-23 Irobot Corporation Mobile human interface robot
WO2013020143A1 (en) * 2011-08-04 2013-02-07 University Of Southern California Image-based crack quantification
US10192103B2 (en) * 2016-01-15 2019-01-29 Stereovision Imaging, Inc. System and method for detecting and removing occlusions in a three-dimensional image
NL2016542B1 (en) * 2016-04-04 2017-10-10 Fugro N V Spatial data analysis.
CN112287939B (en) * 2020-10-29 2024-05-31 平安科技(深圳)有限公司 Three-dimensional point cloud semantic segmentation method, device, equipment and medium
CN112307992A (en) * 2020-11-04 2021-02-02 北京大学深圳研究生院 An automatic identification method of mangrove plants based on UAV visible light remote sensing

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103984936A (en) * 2014-05-29 2014-08-13 中国航空无线电电子研究所 Multi-sensor multi-feature fusion recognition method for three-dimensional dynamic target recognition
CN104881671A (en) * 2015-05-21 2015-09-02 电子科技大学 High resolution remote sensing image local feature extraction method based on 2D-Gabor
CN111027401A (en) * 2019-11-15 2020-04-17 电子科技大学 An end-to-end object detection method for camera and lidar fusion
CN111259733A (en) * 2020-01-02 2020-06-09 武汉理工大学 A method and device for ship recognition based on point cloud images
CN111626217A (en) * 2020-05-28 2020-09-04 宁波博登智能科技有限责任公司 Target detection and tracking method based on two-dimensional picture and three-dimensional point cloud fusion

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
"A review on deep learning techniques for 3D sensed data classification";David Griffiths 等;《https://www.mdpi.com/2072-4292/11/12/1499》;20190625;1-294 *
"An Improved RANSAC for 3D Point Cloud Plane Segmentation Based on Normal Distribution Transformation Cells";Lin Li 等;《https://www.mdpi.com/2072-4292/9/5/433》;20170303;1-16 *
一种自适应运动目标检测方法;路红等;《电子测量与仪器学报》;20080815(第04期);全文 *
基于点云数据的三维目标识别和模型分割方法;牛辰庚等;《图学学报》;20190415(第02期);全文 *

Also Published As

Publication number Publication date
CN113239829A (en) 2021-08-10

Similar Documents

Publication Publication Date Title
CN112488210B (en) A method for automatic classification of 3D point clouds based on graph convolutional neural networks
CN110135267B (en) A detection method for small objects in large scene SAR images
CN107239751B (en) High-resolution SAR image classification method based on non-subsampled contourlet full convolution network
CN111899172A (en) Vehicle target detection method oriented to remote sensing application scene
CN110189304B (en) On-line fast detection method of optical remote sensing image target based on artificial intelligence
CN111028327A (en) A three-dimensional point cloud processing method, device and equipment
CN111310666A (en) High-resolution image ground feature identification and segmentation method based on texture features
CN113239829B (en) Cross-dimension remote sensing data target identification method based on space occupation probability characteristics
CN117218351A (en) Three-dimensional point cloud semantic segmentation method based on local and global context awareness
CN117456389B (en) YOLOv5 s-based improved unmanned aerial vehicle aerial image dense and small target identification method, system, equipment and medium
CN109740539A (en) 3D object recognition method based on extreme learning machine and fusion convolutional network
CN113780305B (en) Significance target detection method based on interaction of two clues
CN116703996A (en) Monocular 3D Object Detection Algorithm Based on Instance-Level Adaptive Depth Estimation
CN115311502A (en) A small sample scene classification method for remote sensing images based on multi-scale dual-stream architecture
CN117496384A (en) A method for object detection in drone images
CN115527071A (en) SAR image ground object classification method based on self-supervision jigsaw learning
Zuo et al. A remote sensing image semantic segmentation method by combining deformable convolution with conditional random fields
CN112508106B (en) A method for underwater image classification based on convolutional neural network
CN111860668A (en) A point cloud recognition method with deep convolutional network for raw 3D point cloud processing
CN118097162A (en) Optimization of cloth simulation filtering algorithm based on PointNet++ network model
CN111563423A (en) Object detection method and system in UAV image based on deep denoising autoencoder
CN114578307B (en) A radar target fusion recognition method and system
Chitturi Building detection in deformed satellite images using mask r-cnn
CN113192204B (en) Three-dimensional reconstruction method for building in single inclined remote sensing image
CN116343016A (en) A multi-angle sonar image target classification method based on lightweight convolutional network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant