CN114037743A - A Robust Registration Method for 3D Point Clouds of Terracotta Warriors Based on Dynamic Graph Attention Mechanism - Google Patents
A Robust Registration Method for 3D Point Clouds of Terracotta Warriors Based on Dynamic Graph Attention Mechanism Download PDFInfo
- Publication number
- CN114037743A CN114037743A CN202111245398.1A CN202111245398A CN114037743A CN 114037743 A CN114037743 A CN 114037743A CN 202111245398 A CN202111245398 A CN 202111245398A CN 114037743 A CN114037743 A CN 114037743A
- Authority
- CN
- China
- Prior art keywords
- point cloud
- registration
- attention mechanism
- dynamic graph
- warriors
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000007246 mechanism Effects 0.000 title claims abstract description 27
- 238000000034 method Methods 0.000 title claims abstract description 20
- 239000011159 matrix material Substances 0.000 claims abstract description 5
- 230000008859 change Effects 0.000 claims abstract description 3
- 238000000605 extraction Methods 0.000 claims description 2
- 238000005070 sampling Methods 0.000 claims 2
- 230000009191 jumping Effects 0.000 claims 1
- 239000000284 extract Substances 0.000 abstract 1
- 230000006870 function Effects 0.000 description 10
- 238000004364 calculation method Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 5
- 230000009466 transformation Effects 0.000 description 5
- 238000010606 normalization Methods 0.000 description 4
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 2
- 230000002776 aggregation Effects 0.000 description 1
- 238000004220 aggregation Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 230000008034 disappearance Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
- G06T7/344—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Image Analysis (AREA)
- Processing Or Creating Images (AREA)
Abstract
本发明公开了一种基于动态图注意力机制的秦俑三维点云鲁棒配准方法,包括:步骤1,通过三维扫描仪获取不同分辨率的秦俑三维点云;步骤2,在U‑Net网络中,将卷积层替换为B‑NHN‑Conv,将反卷积层替换为B‑NHN‑ConvTr,并将残差模块和动态图注意力机制嵌入到U‑Net网络中,得到点云配准网络;步骤3,将不同分辨率的秦俑三维点云输入到点云配准网络中,在circle loss损失函数和overlap loss损失函数的监督下对其进行训练;步骤4,利用训练完成的点云配准网络提取秦俑三维点云特征,并结合RANSAC算法来估计源点云与目标点云之间的变化矩阵,完成秦俑三维点云的配准。本发明提供的配准方法在点云分辨率不匹配、包含大量噪声的情况下仍能学习鲁棒的特征并较好地完成低重叠度下点云的配准。
The invention discloses a method for robust registration of three-dimensional point clouds of terracotta figurines based on a dynamic graph attention mechanism. Net network, replace the convolutional layer with B‑NHN‑Conv, replace the deconvolutional layer with B‑NHN‑ConvTr, and embed the residual module and dynamic graph attention mechanism into the U‑Net network to get the point Cloud registration network; step 3, input the 3D point cloud of Terracotta Warriors with different resolutions into the point cloud registration network, and train it under the supervision of the circle loss loss function and the overlap loss loss function; step 4, use the training The completed point cloud registration network extracts the 3D point cloud features of the Terracotta Warriors, and combines the RANSAC algorithm to estimate the change matrix between the source point cloud and the target point cloud to complete the registration of the Terracotta Warriors 3D point cloud. The registration method provided by the present invention can still learn robust features under the condition that the resolution of the point cloud does not match and contains a lot of noise, and can better complete the registration of the point cloud under low overlap.
Description
技术领域technical field
本发明涉及三维点云模型配准技术,具体涉及一种基于动态图注意力机制的秦俑三维点云鲁棒配准方法。The invention relates to a three-dimensional point cloud model registration technology, in particular to a method for robust registration of three-dimensional point clouds of Qin figurines based on a dynamic graph attention mechanism.
背景技术Background technique
点云配准技术在秦俑虚拟复原、秦陵智慧博物馆等项目中发挥着重要的作用,点云配准结果的准确是后续三维重建的关键。点云配准的目的在于通过计算得到一个完美的坐标变换,通过旋转平移等刚性变换,将不同视角下的点云数据统一整合到指定的坐标系中。The point cloud registration technology plays an important role in projects such as the virtual restoration of the Terracotta Warriors and the Qinling Wisdom Museum. The accuracy of the point cloud registration results is the key to the subsequent 3D reconstruction. The purpose of point cloud registration is to obtain a perfect coordinate transformation through calculation, and to uniformly integrate point cloud data from different perspectives into a specified coordinate system through rigid transformations such as rotation and translation.
目前,秦俑以及相关文物的配准方法大多是基于优化的传统配准方法,其中较为经典的方法是迭代最近点(Iterative Closest Point)算法。该方法主要包含两个阶段:对应搜索和变换估计。这两个阶段将反复进行,以找到点云间最佳转换。然而在处理初始位置差异大、点云分辨率不匹配、噪声干扰强以及重叠程度小的场景时,上述方法容易陷入局部最优。近年来,有研究人员提出了基于深度学习的方法来学习鲁棒的特征计算对应点,最终通过RANSAC或SVD算法来最终确定变换矩阵,而无需对应估计和变换估计之间的迭代。然而这类算法在面对部分重叠、密度变化、噪声等挑战时不能快速、鲁棒地处理。At present, most of the registration methods of Qin figurines and related cultural relics are traditional registration methods based on optimization, among which the more classic method is the Iterative Closest Point algorithm. The method mainly consists of two stages: correspondence search and transform estimation. These two stages are repeated to find the best transition between point clouds. However, when dealing with scenes with large initial position differences, mismatched point cloud resolutions, strong noise interference, and small overlap, the above methods are prone to fall into local optimum. In recent years, some researchers have proposed methods based on deep learning to learn robust feature calculation corresponding points, and finally determine the transformation matrix through RANSAC or SVD algorithm without iteration between correspondence estimation and transformation estimation. However, such algorithms cannot be fast and robust in the face of challenges such as partial overlap, density variation, and noise.
发明内容SUMMARY OF THE INVENTION
针对现有技术存在的不足,本发明的目的在于提供一种在面对部分重叠和密度变化时能鲁棒地处理的基于动态图注意力机制的秦俑三维点云鲁棒配准方法。In view of the deficiencies in the prior art, the purpose of the present invention is to provide a robust registration method for 3D point clouds of Terracotta Warriors based on dynamic graph attention mechanism, which can robustly handle partial overlap and density changes.
为了实现上述目的,本发明采用以下技术方案予以实现:In order to achieve the above object, the present invention adopts the following technical solutions to realize:
一种基于动态图注意力机制的秦俑三维点云鲁棒配准方法,包括如下步骤:A robust registration method for 3D point cloud of Terracotta Warriors based on dynamic graph attention mechanism, comprising the following steps:
步骤1,通过三维扫描仪获取不同分辨率的秦俑三维点云;
步骤2,在U-Net网络中,将卷积层替换为B-NHN-Conv,将反卷积层替换为B-NHN-ConvTr,并将残差模块和动态图注意力机制嵌入到U-Net网络中,得到点云配准网络;
步骤3,将不同分辨率的秦俑三维点云输入到点云配准网络中,在circle loss损失函数和overlap loss损失函数的共同监督下训练构建的点云配准网络;Step 3: Input the three-dimensional point clouds of the Terracotta Warriors with different resolutions into the point cloud registration network, and train the constructed point cloud registration network under the joint supervision of the circle loss loss function and the overlap loss loss function;
步骤4,利用训练完成的点云配准网络提取秦俑三维点云特征,并结合RANSAC算法来估计源点云与目标点云之间的变化矩阵,完成秦俑三维点云的配准。Step 4: Use the trained point cloud registration network to extract the 3D point cloud features of the Terracotta Warriors, and combine the RANSAC algorithm to estimate the change matrix between the source point cloud and the target point cloud to complete the registration of the Terracotta Warriors 3D point cloud.
进一步地,所述步骤2中的U-Net网络还包括编码器模块和解码器模块,所述编码器模块通过多次的下采样获得多个尺度特征,并将多层感知器中所有全连接层转换为一系列内核大小为1×1×1的全卷积层,通道数为(64,128,256,512);所述解码器模块每上采样一次,就通过跳跃连接结构和特征提取部分对应通道数相同尺度的特征进行融合,还原降采样所带来的信息损失。Further, the U-Net network in the
进一步地,所述步骤2中的残差模块由多个残差块串联而成,用于将输入信息跳跃连接到输出信息。Further, the residual module in the
进一步地,所述步骤2中动态图注意力机制包括多层自注意力模块和交叉注意力模块的组合模块,所述动态图注意力机制基于注意权重aij为每个查询节点找到前k个边,并且仅使用前k个边和相应的节点来构造图,并在动态图注意力机制的每一层都根据变化后的注意权重aij构造一个新的图。Further, in the
本发明与现有技术相比,具有如下技术效果:Compared with the prior art, the present invention has the following technical effects:
本发明使用U-Net作为主干架构,通过嵌入残差块解决梯度消失或梯度爆炸问题;通过嵌入动态图注意力机制来聚合点云的局部特征和上下文特征,可以获得多层次、具有更加丰富的语义表征,也能更好地帮助网络确定点云间可能的重叠区域;并通过B-NHN-Conv卷积操作减少或去除特征均值和标准偏差的潜在变化,提高了三维特征对点密度变化的鲁棒性;这使得本发明构建的点云配准网络在点云分辨率不匹配、包含大量噪声的情况下仍能学习鲁棒的特征并较好地完成低重叠度下点云的配准。The present invention uses U-Net as the backbone architecture, and solves the problem of gradient disappearance or gradient explosion by embedding residual blocks; by embedding dynamic graph attention mechanism to aggregate local features and context features of point clouds, multi-level and richer features can be obtained. Semantic representation can also better help the network determine possible overlapping areas between point clouds; and reduce or remove potential changes in feature mean and standard deviation through the B-NHN-Conv convolution operation, which improves the sensitivity of 3D features to point density changes. Robustness; this enables the point cloud registration network constructed in the present invention to learn robust features and better complete the registration of point clouds under low overlap even when the resolution of point clouds does not match and contains a lot of noise .
进一步地,将多层感知器中的所有全连接层转换为一系列内核大小为1×1×1的全卷积层,可加快网络处理的效率。Further, converting all fully connected layers in a multilayer perceptron into a series of fully convolutional layers with a kernel size of 1 × 1 × 1 can speed up the efficiency of network processing.
附图说明Description of drawings
图1为本发明一种基于动态图注意力机制的秦俑三维点云鲁棒配准方法的流程图;1 is a flowchart of a method for robust registration of three-dimensional point clouds of Terracotta Warriors based on a dynamic graph attention mechanism of the present invention;
图2为点云配准网络模型图;Fig. 2 is the network model diagram of point cloud registration;
图3为动态图注意力机制的结构示意图;Figure 3 is a schematic structural diagram of a dynamic graph attention mechanism;
图4为残差模块的结构示意图;FIG. 4 is a schematic structural diagram of a residual module;
图5为秦俑两头部点云初始位姿图;Figure 5 is the initial pose diagram of the point cloud of the two heads of the Qin figurines;
图6为秦俑两头部点云配准结果图;Figure 6 is the result of point cloud registration of the two heads of the Qin figurines;
图7为秦俑两足部点云初始位姿图;Figure 7 is the initial pose diagram of the point cloud of the two feet of the Qin figurines;
图8为秦俑两足部点云配准结果图。Figure 8 shows the result of point cloud registration on the feet of the Qin figurines.
具体实施方式Detailed ways
以下结合实施例对本发明的具体内容做进一步详细解释说明。The specific content of the present invention will be further explained in detail below in conjunction with the embodiments.
参照图1-4,本实施例提供一种基于动态图注意力机制的秦俑三维点云鲁棒配准方法,包括如下步骤:1-4, this embodiment provides a method for robust registration of 3D point clouds of Terracotta Warriors based on a dynamic graph attention mechanism, including the following steps:
步骤1,通过三维扫描仪获取不同分辨率的秦俑三维点云;
步骤2,在U-Net网络中,将卷积层替换为B-NHN-Conv,将反卷积层替换为B-NHN-ConvTr,并将残差模块和动态图注意力机制嵌入到U-Net网络中,得到点云配准网络;其中,
B-NHN-Conv是结合了B-NHN归一化与后续的三维稀疏卷积的操作,且B-NHN归一化与三维稀疏卷积之间紧密耦合;B-NHN-ConvTr是将B-NHN-Conv卷积操作进行转置,即B-NHN归一化操作与三维稀疏转置卷积函数结合;B-NHN归一化操作可减少或去除特征均值和标准偏差的潜在变化,提高三维特征对点密度变化的鲁棒性。B-NHN-Conv is a combination of B-NHN normalization and subsequent three-dimensional sparse convolution operations, and the B-NHN normalization and three-dimensional sparse convolution are tightly coupled; B-NHN-ConvTr is a combination of B- The NHN-Conv convolution operation is transposed, that is, the B-NHN normalization operation is combined with the three-dimensional sparse transpose convolution function; the B-NHN normalization operation can reduce or remove the potential changes in the feature mean and standard deviation, and improve the three-dimensional Robustness of features to changes in point density.
残差模块由多个残差块串联而成,用于将输入信息跳跃连接到输出信息;可缓解有效特征信息丢失的问题,使得网络更容易得到优化,更有助于建立两点云间的对应关系。The residual module is composed of multiple residual blocks in series, which is used to jump connect the input information to the output information; it can alleviate the problem of the loss of effective feature information, make the network easier to optimize, and help to establish the relationship between the two point clouds. Correspondence.
动态图注意力机制包括多层自注意力模块(self-attention)和交叉注意力模块(cross-attention)的组合模块,也称动态图注意力模块。所述动态图注意力机制基于注意权重aij为每个查询节点找到前k个边,并且仅使用前k个边和相应的节点来构造图,并在动态图注意力机制的每一层都根据变化后的注意权重aij构造一个新的图。The dynamic graph attention mechanism includes a combination of multiple layers of self-attention and cross-attention, also known as dynamic graph attention. The dynamic graph attention mechanism finds the top k edges for each query node based on the attention weights a ij , and uses only the top k edges and corresponding nodes to construct the graph, and uses the dynamic graph attention mechanism at each layer of the dynamic graph attention mechanism. Construct a new graph according to the changed attention weights a ij .
动态图注意力机制首先通过线性投影对图中的特征节点计算查询向量键向量和值向量其中表示实数集,b为特征的维数,用于图结构更新和注意力聚合,如下公式所示:The dynamic graph attention mechanism first calculates the query vector for the feature nodes in the graph through linear projection key vector sum value vector in Represents the set of real numbers, and b is the dimension of the feature, which is used for graph structure update and attention aggregation, as shown in the following formula:
qi=W1 (l)fi Q+b1 q i =W 1 (l) f i Q +b 1
其中,{Q,S}∈{X,Y}2,当Q=S时,表示自注意力;Q≠S时表示交叉注意力;W和b为可学习的投影参数,(l)fi Q表示点云Q中位于动态图注意模块l层的关键点的特征,可表示为查询节点,点云S中所有节点可以表示为源节点。Among them, {Q,S}∈{X,Y} 2 , when Q=S, it means self-attention; when Q≠S, it means cross-attention; W and b are learnable projection parameters, (l) f i Q represents the feature of the key point located at the l layer of the dynamic graph attention module in the point cloud Q, which can be represented as a query node, and all nodes in the point cloud S can be represented as source nodes.
通过计算qi和各个kj的相似性或者相关性,得到每个kj对应vj的权重系数αij,然后对vj进行加权求和,即得到了最终的Attention数值,消息通过加权平均值来计算,计算公式如下:By calculating the similarity or correlation between qi and each k j , the weight coefficient α ij of each k j corresponding to v j is obtained, and then the weighted summation of v j is performed , that is, the final Attention value is obtained. value, the calculation formula is as follows:
其中权重系数表示特征(l)fi Q对特征的关注程度;ε∈{εself,εcross}。where the weight factor Represents feature (l)f i Q pair feature the degree of attention; ε∈{ε self ,ε cross }.
一旦对所有层进行了聚合,节点的最终特征可表示为:Once all layers are aggregated, the final features of nodes can be expressed as:
fi=Wfi+bf i =Wf i +b
所述U-Net网络还包括用于提取特征的编码器模块和对特征进行融合的解码器模块,所述编码器模块通过多次的下采样获得多个尺度特征,并将多层感知器中所有全连接层转换为一系列内核大小为1×1×1的全卷积层,通道数为(64,128,256,512);所述解码器模块每上采样一次,就通过跳跃连接结构和特征提取部分对应通道数相同尺度的特征进行融合,还原降采样所带来的信息损失。The U-Net network also includes an encoder module for extracting features and a decoder module for fusing features, the encoder module obtains multiple scale features through multiple downsampling, and combines the multi-layer perceptron All fully connected layers are converted into a series of fully convolutional layers with a kernel size of 1×1×1, and the number of channels is (64, 128, 256, 512). The features of the same scale are fused to restore the information loss caused by downsampling.
步骤3,将不同分辨率的秦俑三维点云输入到点云配准网络中,在circle loss损失函数(循环损失函数)和overlap loss损失函数(重叠损失函数)的共同监督下训练构建的点云配准网络;Step 3: Input the 3D point clouds of the Terracotta Warriors with different resolutions into the point cloud registration network, and train the constructed points under the joint supervision of the circle loss function (cycle loss function) and the overlap loss function (overlap loss function). cloud registration network;
Circle loss是点云配准中常见的triplet loss的一种变体,损失函数公式如下所示:Circle loss is a variant of triplet loss commonly used in point cloud registration. The loss function formula is as follows:
其中,重叠点云对X、Y已对齐,nx表示从点云X随机采样后的点云数量,εx(xi)表示以点xi为中心,半径rx内的所有属于点云Y的点,εn(xi)表示以点xi为中心,半径rx外的所有属于点云Y的点,表示特征空间中两特征间的距离,Δx和Δn分别表示正负样本间隔,权重由超参数γ、以及样本正样本间隔Δx所决定,同样权重由超参数γ、以及样本负样本间隔Δn所决定,计算方法同理,所以circle loss最终的损失函数为 Among them, the overlapping point cloud pairs X and Y have been aligned, n x represents the number of point clouds randomly sampled from the point cloud X, ε x (x i ) represents the point x i as the center, and all points within the radius r x belong to the point cloud The point of Y, ε n (x i ) represents all the points belonging to the point cloud Y outside the radius r x with the point x i as the center, Represents the distance between two features in the feature space, Δx and Δn represent the positive and negative sample intervals, respectively, and the weight By the hyperparameter γ, And the sample positive sample interval Δ x is determined, the same weight By the hyperparameter γ, And the sample negative sample interval Δn is determined, The calculation method is the same, so the final loss function of circle loss is
重叠区域的估计被转换为二元分类,使用overlap loss进行监督,损失函数如下所示:The estimates of overlapping regions are converted to binary classification, supervised using overlap loss, and the loss function is as follows:
其中,真实的标签表示点xi是否为重叠区域,表示网络预测的标签。计算方法同理。Among them, the true label Indicates whether point xi is an overlapping area, Represents the label predicted by the network. The calculation method is the same.
步骤4,将源点云和目标点云输入到经过训练的点云配准网络中,依次经过编码器模块,动态图注意力模块,最后通过解码器模块输出提取的特征,并结合RANSAC算法计算变换矩阵完成配准。Step 4: Input the source point cloud and the target point cloud into the trained point cloud registration network, pass through the encoder module, the dynamic graph attention module in turn, and finally output the extracted features through the decoder module, and calculate with the RANSAC algorithm The transformation matrix completes the registration.
图5、图7分别为通过三维扫描仪获取的秦俑两头部点云初始位姿图和秦俑两足部点云初始位姿图;图6、图8别人为通过本发明所述方法进行配准后的秦俑两头部点云配准结果图和秦俑两足部点云配准结果图。由图可知本发明所述方法针对点云部分重叠和密度变化时仍能鲁棒地处理。Figure 5 and Figure 7 are respectively the initial pose map of the point cloud of the two heads of the Qin figurines and the initial pose map of the point clouds of the two feet of the Qin figurines obtained by the 3D scanner; The registration result of the point cloud of the two heads of the Qin figurines and the point cloud registration result of the feet of the Qin figurines after the registration. It can be seen from the figure that the method of the present invention can still handle robustly when the point cloud is partially overlapped and the density changes.
Claims (4)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111245398.1A CN114037743B (en) | 2021-10-26 | 2021-10-26 | A robust registration method for 3D point cloud of Qin Terracotta Warriors based on dynamic graph attention mechanism |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111245398.1A CN114037743B (en) | 2021-10-26 | 2021-10-26 | A robust registration method for 3D point cloud of Qin Terracotta Warriors based on dynamic graph attention mechanism |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114037743A true CN114037743A (en) | 2022-02-11 |
CN114037743B CN114037743B (en) | 2024-01-26 |
Family
ID=80135357
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111245398.1A Active CN114037743B (en) | 2021-10-26 | 2021-10-26 | A robust registration method for 3D point cloud of Qin Terracotta Warriors based on dynamic graph attention mechanism |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114037743B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115546482A (en) * | 2022-09-26 | 2022-12-30 | 浙江省测绘科学技术研究院 | Outdoor point cloud semantic segmentation method based on statistical projection |
CN115631341A (en) * | 2022-12-21 | 2023-01-20 | 北京航空航天大学 | A point cloud registration method and system based on multi-scale feature voting |
CN115631221A (en) * | 2022-11-30 | 2023-01-20 | 北京航空航天大学 | Low-overlapping-degree point cloud registration method based on consistency sampling |
CN119180828A (en) * | 2024-11-22 | 2024-12-24 | 温州大学 | Part surface point cloud segmentation method based on multi-scale network of dynamic graph attention |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108596961A (en) * | 2018-04-17 | 2018-09-28 | 浙江工业大学 | Point cloud registration method based on Three dimensional convolution neural network |
US20190147245A1 (en) * | 2017-11-14 | 2019-05-16 | Nuro, Inc. | Three-dimensional object detection for autonomous robotic systems using image proposals |
CN111046781A (en) * | 2019-12-09 | 2020-04-21 | 华中科技大学 | Robust three-dimensional target detection method based on ternary attention mechanism |
CN111882593A (en) * | 2020-07-23 | 2020-11-03 | 首都师范大学 | A point cloud registration model and method combining attention mechanism and 3D graph convolutional network |
CN112837356A (en) * | 2021-02-06 | 2021-05-25 | 湖南大学 | An unsupervised multi-view 3D point cloud co-registration method based on WGAN |
WO2021104056A1 (en) * | 2019-11-27 | 2021-06-03 | 中国科学院深圳先进技术研究院 | Automatic tumor segmentation system and method, and electronic device |
-
2021
- 2021-10-26 CN CN202111245398.1A patent/CN114037743B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190147245A1 (en) * | 2017-11-14 | 2019-05-16 | Nuro, Inc. | Three-dimensional object detection for autonomous robotic systems using image proposals |
CN108596961A (en) * | 2018-04-17 | 2018-09-28 | 浙江工业大学 | Point cloud registration method based on Three dimensional convolution neural network |
WO2021104056A1 (en) * | 2019-11-27 | 2021-06-03 | 中国科学院深圳先进技术研究院 | Automatic tumor segmentation system and method, and electronic device |
CN111046781A (en) * | 2019-12-09 | 2020-04-21 | 华中科技大学 | Robust three-dimensional target detection method based on ternary attention mechanism |
CN111882593A (en) * | 2020-07-23 | 2020-11-03 | 首都师范大学 | A point cloud registration model and method combining attention mechanism and 3D graph convolutional network |
CN112837356A (en) * | 2021-02-06 | 2021-05-25 | 湖南大学 | An unsupervised multi-view 3D point cloud co-registration method based on WGAN |
Non-Patent Citations (2)
Title |
---|
唐灿;唐亮贵;刘波;: "图像特征检测与匹配方法研究综述", 南京信息工程大学学报(自然科学版), no. 03 * |
马福峰;耿楠;张志毅;: "基于邻域几何特征约束的植株三维形态配准方法研究", 计算机应用与软件, no. 09 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115546482A (en) * | 2022-09-26 | 2022-12-30 | 浙江省测绘科学技术研究院 | Outdoor point cloud semantic segmentation method based on statistical projection |
CN115631221A (en) * | 2022-11-30 | 2023-01-20 | 北京航空航天大学 | Low-overlapping-degree point cloud registration method based on consistency sampling |
CN115631341A (en) * | 2022-12-21 | 2023-01-20 | 北京航空航天大学 | A point cloud registration method and system based on multi-scale feature voting |
CN119180828A (en) * | 2024-11-22 | 2024-12-24 | 温州大学 | Part surface point cloud segmentation method based on multi-scale network of dynamic graph attention |
CN119180828B (en) * | 2024-11-22 | 2025-03-25 | 温州大学 | Part surface point cloud segmentation method based on multi-scale network with dynamic graph attention |
Also Published As
Publication number | Publication date |
---|---|
CN114037743B (en) | 2024-01-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114037743B (en) | A robust registration method for 3D point cloud of Qin Terracotta Warriors based on dynamic graph attention mechanism | |
CN109949368B (en) | A 3D Pose Estimation Method of Human Body Based on Image Retrieval | |
CN109948425B (en) | A pedestrian search method and device based on structure-aware self-attention and online instance aggregation and matching | |
CN109840556B (en) | Image classification and identification method based on twin network | |
CN107229757B (en) | Video retrieval method based on deep learning and hash coding | |
CN106055576B (en) | A kind of fast and effectively image search method under large-scale data background | |
CN110659565B (en) | 3D multi-person human body posture estimation method based on porous convolution | |
CN113988147B (en) | Multi-label classification method and device for remote sensing image scene based on graph network, and multi-label retrieval method and device | |
Zhang et al. | DDRNet: Fast point cloud registration network for large-scale scenes | |
CN112396167B (en) | Loop detection method for fusing appearance similarity and spatial position information | |
CN114612660A (en) | Three-dimensional modeling method based on multi-feature fusion point cloud segmentation | |
CN112818920B (en) | A combined spatial and spectral change detection method for dual-phase hyperspectral images | |
CN116228825B (en) | Point cloud registration method based on significant anchor point geometric embedding | |
CN116128944A (en) | Three-dimensional point cloud registration method based on feature interaction and reliable corresponding relation estimation | |
CN111401303A (en) | Cross-visual angle gait recognition method with separated identity and visual angle characteristics | |
CN114663880B (en) | Three-dimensional object detection method based on multi-level cross-modal self-attention mechanism | |
CN113989283B (en) | 3D human body posture estimation method and device, electronic equipment and storage medium | |
CN112036511B (en) | Image retrieval method based on attention mechanism graph convolutional neural network | |
CN115423847A (en) | Twin Multimodal Target Tracking Method Based on Transformer | |
CN113793261A (en) | Spectrum reconstruction method based on 3D attention mechanism full-channel fusion network | |
CN116665114A (en) | Multi-mode-based remote sensing scene identification method, system and medium | |
CN117788810A (en) | A learning system for unsupervised semantic segmentation | |
CN119360383A (en) | Point cloud instance segmentation method based on weakly supervised dynamic convolution | |
CN115861664A (en) | Feature matching method and system based on local feature fusion and self-attention mechanism | |
CN118887264A (en) | Point cloud registration method and system based on embedding local features into global features, and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |