[go: up one dir, main page]

CN114913105A - Laser point cloud fusion method, device, server and computer-readable storage medium - Google Patents

Laser point cloud fusion method, device, server and computer-readable storage medium Download PDF

Info

Publication number
CN114913105A
CN114913105A CN202210512885.8A CN202210512885A CN114913105A CN 114913105 A CN114913105 A CN 114913105A CN 202210512885 A CN202210512885 A CN 202210512885A CN 114913105 A CN114913105 A CN 114913105A
Authority
CN
China
Prior art keywords
processed
point cloud
image
laser
coordinate system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210512885.8A
Other languages
Chinese (zh)
Inventor
崔岩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Germany Zhuhai Artificial Intelligence Institute Co ltd
Guangdong Siwei Kanan Intelligent Equipment Co Ltd
4Dage Co Ltd
Original Assignee
China Germany Zhuhai Artificial Intelligence Institute Co ltd
Guangdong Siwei Kanan Intelligent Equipment Co Ltd
4Dage Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Germany Zhuhai Artificial Intelligence Institute Co ltd, Guangdong Siwei Kanan Intelligent Equipment Co Ltd, 4Dage Co Ltd filed Critical China Germany Zhuhai Artificial Intelligence Institute Co ltd
Priority to CN202210512885.8A priority Critical patent/CN114913105A/en
Publication of CN114913105A publication Critical patent/CN114913105A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Laser Beam Processing (AREA)

Abstract

本申请适用于图像处理技术领域,提供了一种激光点云融合方法、装置、服务器及计算机可读存储介质,该方法包括:获取待处理点云;其中,待处理点云包括第一待处理点云和第二待处理点云;融合第一待处理点云和所述第二待处理点云,得到目标点云。可见,本申请可以提高激光点云融合的精度。

Figure 202210512885

The present application is applicable to the technical field of image processing, and provides a laser point cloud fusion method, device, server, and computer-readable storage medium. The method includes: acquiring a point cloud to be processed; wherein, the point cloud to be processed includes a first to-be-processed point cloud. point cloud and the second point cloud to be processed; fusing the first point cloud to be processed and the second point cloud to be processed to obtain the target point cloud. It can be seen that the present application can improve the accuracy of laser point cloud fusion.

Figure 202210512885

Description

激光点云融合方法、装置、服务器及计算机可读存储介质Laser point cloud fusion method, device, server and computer-readable storage medium

技术领域technical field

本申请属于点云技术领域,尤其涉及一种激光点云融合方法、装置、服务器及计算机可读存储介质。The present application belongs to the technical field of point clouds, and in particular, relates to a laser point cloud fusion method, device, server, and computer-readable storage medium.

背景技术Background technique

随着传感技术和测绘装备的不断发展,数据来源多种多样,在三维重建领域,需要获取全面的点云数据,使得重建效果更佳。但是现有的不同类型的点云数据之间的融合精度较低。With the continuous development of sensing technology and surveying and mapping equipment, there are various data sources. In the field of 3D reconstruction, it is necessary to obtain comprehensive point cloud data to make the reconstruction effect better. However, the fusion accuracy between existing different types of point cloud data is low.

发明内容SUMMARY OF THE INVENTION

本申请实施例提供了一种激光点云融合方法、装置、服务器及计算机可读存储介质,可以解决现有技术中点云融合精度较低的技术问题。Embodiments of the present application provide a laser point cloud fusion method, device, server, and computer-readable storage medium, which can solve the technical problem of low point cloud fusion accuracy in the prior art.

第一方面,本申请实施例提供了一种激光点云融合方法,包括:In a first aspect, an embodiment of the present application provides a laser point cloud fusion method, including:

获取待处理点云;其中,所述待处理点云包括第一待处理点云和第二待处理点云;Obtain a point cloud to be processed; wherein, the point cloud to be processed includes a first point cloud to be processed and a second point cloud to be processed;

融合所述第一待处理点云和所述第二待处理点云,得到目标点云。The target point cloud is obtained by fusing the first point cloud to be processed and the second point cloud to be processed.

在第一方面的一种可能的实施方式中,获取待处理点云,包括:In a possible implementation of the first aspect, acquiring the point cloud to be processed includes:

获取第一待处理图像和第一激光数据;acquiring the first image to be processed and the first laser data;

获取第二待处理图像和第二激光数据;acquiring the second to-be-processed image and the second laser data;

对所述第一激光数据所处得第一坐标系和所述第二激光数据所处得第二坐标系进行坐标系校准;performing coordinate system calibration on a first coordinate system where the first laser data is located and a second coordinate system where the second laser data is located;

匹配所述第一待处理图像和所述第二待处理图像;matching the first image to be processed and the second image to be processed;

根据匹配后的所述第一待处理图像和坐标系校准后的所述第一激光数据生成第一待处理点云;generating a first point cloud to be processed according to the matched first image to be processed and the first laser data after coordinate system calibration;

根据匹配后的所述第二待处理图像和坐标系校准后的所述第二激光数据生成第二待处理点云。A second point cloud to be processed is generated according to the matched second image to be processed and the second laser data after coordinate system calibration.

在第一方面的一种可能的实施方式中,匹配所述第一待处理图像和所述第二待处理图像,包括:In a possible implementation manner of the first aspect, matching the first image to be processed and the second image to be processed includes:

提取所述第一待处理图像的第一特征点;extracting the first feature point of the first image to be processed;

提取所述第二待处理图像的第二特征点;extracting second feature points of the second image to be processed;

筛选出所述第一特征点和第二特征点中共有的目标特征点。The target feature points common to the first feature point and the second feature point are screened out.

在第一方面的一种可能的实施方式中,融合所述第一待处理点云和所述第二待处理点云,得到目标点云,包括:In a possible implementation manner of the first aspect, the first point cloud to be processed and the second point cloud to be processed are fused to obtain a target point cloud, including:

根据所述目标特征点获取位姿信息;Obtain pose information according to the target feature point;

根据所述位姿信息,对所述第一待处理点云和所述第二待处理点云进行融合,得到目标点云。According to the pose information, the first point cloud to be processed and the second point cloud to be processed are fused to obtain a target point cloud.

在第一方面的一种可能的实施方式中,融合所述第一待处理点云和所述第二待处理点云,得到目标点云之后,还包括:In a possible implementation manner of the first aspect, after fusing the first point cloud to be processed and the second point cloud to be processed to obtain the target point cloud, the method further includes:

优化所述目标点云。Optimize the target point cloud.

第二方面,本申请实施例提供了一种激光点云融合装置,包括:In a second aspect, an embodiment of the present application provides a laser point cloud fusion device, including:

获取模块,用于获取待处理点云;其中,所述待处理点云包括第一待处理点云和第二待处理点云;an acquisition module, configured to acquire a point cloud to be processed; wherein, the point cloud to be processed includes a first point cloud to be processed and a second point cloud to be processed;

融合模块,用于融合所述第一待处理点云和所述第二待处理点云,得到目标点云。A fusion module, configured to fuse the first point cloud to be processed and the second point cloud to be processed to obtain a target point cloud.

在第二方面的一种可选的实施方式中,所述获取模块,包括:In an optional implementation manner of the second aspect, the obtaining module includes:

第一获取子模块,用于获取第一待处理图像和第一激光数据;a first acquisition sub-module for acquiring the first image to be processed and the first laser data;

第二获取子模块,用于获取第二待处理图像和第二激光数据;a second acquisition sub-module for acquiring the second to-be-processed image and the second laser data;

校准子模块,用于对所述第一激光数据所处得第一坐标系和所述第二激光数据所处得第二坐标系进行坐标系校准;a calibration sub-module for performing coordinate system calibration on the first coordinate system where the first laser data is located and the second coordinate system where the second laser data is located;

匹配子模块,用于匹配所述第一待处理图像和所述第二待处理图像;a matching submodule for matching the first image to be processed and the second image to be processed;

第一生成子模块,用于根据匹配后的所述第一待处理图像和坐标系校准后的所述第一激光数据生成第一待处理点云;a first generation submodule, configured to generate a first point cloud to be processed according to the matched first image to be processed and the first laser data after coordinate system calibration;

第二生成子模块,用于根据匹配后的所述第二待处理图像和坐标系校准后的所述第二激光数据生成第二待处理点云。The second generation sub-module is configured to generate a second point cloud to be processed according to the matched second image to be processed and the second laser data after coordinate system calibration.

在第二方面的一种可选的实施方式中,所述匹配子模块,包括:In an optional implementation manner of the second aspect, the matching submodule includes:

第一提取单元,用于提取所述第一待处理图像的第一特征点;a first extraction unit, configured to extract the first feature point of the first image to be processed;

第二提取单元,用于提取所述第二待处理图像的第二特征点;a second extraction unit, configured to extract the second feature point of the second to-be-processed image;

筛选单元,用于筛选出所述第一特征点和第二特征点中共有的目标特征点。A screening unit, configured to screen out target feature points common to the first feature point and the second feature point.

在第二方面的一种可选的实施方式中,所述融合模块,包括:In an optional implementation manner of the second aspect, the fusion module includes:

第三获取子模块,用于根据所述目标特征点获取位姿信息;a third acquisition sub-module, configured to acquire pose information according to the target feature point;

融合子模块,用于根据所述位姿信息,对所述第一待处理点云和所述第二待处理点云进行融合,得到目标点云。The fusion sub-module is configured to fuse the first point cloud to be processed and the second point cloud to be processed according to the pose information to obtain a target point cloud.

在第二方面的一种可选的实施方式中,所述装置,还包括:In an optional implementation manner of the second aspect, the device further includes:

优化模块,用于优化所述目标点云。An optimization module for optimizing the target point cloud.

第三方面,本申请实施例提供了一种服务器,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现如上述第一方面所述的方法。In a third aspect, an embodiment of the present application provides a server, including a memory, a processor, and a computer program stored in the memory and executable on the processor, which is implemented when the processor executes the computer program The method as described in the first aspect above.

第四方面,本申请实施例提供了一种可读计算机可读存储介质,所述计算机程序被处理器执行时实现如上述第一方面所述的方法。In a fourth aspect, an embodiment of the present application provides a readable computer-readable storage medium, and when the computer program is executed by a processor, the method described in the first aspect above is implemented.

本申请实施例与现有技术相比存在的有益效果是:The beneficial effects that the embodiments of the present application have compared with the prior art are:

本申请实施例通过获取待处理点云;其中,待处理点云包括第一待处理点云和第二待处理点云;融合第一待处理点云和所述第二待处理点云,得到目标点云,提高激光点云融合的精度。The embodiment of the present application obtains the point cloud to be processed; wherein the point cloud to be processed includes a first point cloud to be processed and a second point cloud to be processed; and the first point cloud to be processed and the second point cloud to be processed are obtained by fusing the first point cloud to be processed and the second point cloud to be processed Target point cloud to improve the accuracy of laser point cloud fusion.

附图说明Description of drawings

为了更清楚地说明本申请实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to illustrate the technical solutions in the embodiments of the present application more clearly, the following briefly introduces the accompanying drawings that need to be used in the description of the embodiments or the prior art. Obviously, the drawings in the following description are only for the present application. In some embodiments, for those of ordinary skill in the art, other drawings can also be obtained according to these drawings without any creative effort.

图1是本申请实施例提供的激光点云融合方法的流程示意图;1 is a schematic flowchart of a laser point cloud fusion method provided by an embodiment of the present application;

图2是本申请实施例提供的激光点云融合装置的结构框图;2 is a structural block diagram of a laser point cloud fusion device provided by an embodiment of the present application;

图3是本申请实施例提供的服务器的结构示意图。FIG. 3 is a schematic structural diagram of a server provided by an embodiment of the present application.

具体实施方式Detailed ways

以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本申请实施例。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本申请。在其它情况中,省略对众所周知的系统、装置、电路以及方法的详细说明,以免不必要的细节妨碍本申请的描述。In the following description, for the purpose of illustration rather than limitation, specific details such as a specific system structure and technology are set forth in order to provide a thorough understanding of the embodiments of the present application. However, it will be apparent to those skilled in the art that the present application may be practiced in other embodiments without these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present application with unnecessary detail.

应当理解,当在本申请说明书和所附权利要求书中使用时,术语“包括”指示所描述特征、整体、步骤、操作、元素和/或组件的存在,但并不排除一个或多个其它特征、整体、步骤、操作、元素、组件和/或其集合的存在或添加。It is to be understood that, when used in this specification and the appended claims, the term "comprising" indicates the presence of the described feature, integer, step, operation, element and/or component, but does not exclude one or more other The presence or addition of features, integers, steps, operations, elements, components and/or sets thereof.

还应当理解,在本申请说明书和所附权利要求书中使用的术语“和/或”是指相关联列出的项中的一个或多个的任何组合以及所有可能组合,并且包括这些组合。It will also be understood that, as used in this specification and the appended claims, the term "and/or" refers to and including any and all possible combinations of one or more of the associated listed items.

如在本申请说明书和所附权利要求书中所使用的那样,术语“如果”可以依据上下文被解释为“当...时”或“一旦”或“响应于确定”或“响应于检测到”。类似地,短语“如果确定”或“如果检测到[所描述条件或事件]”可以依据上下文被解释为意指“一旦确定”或“响应于确定”或“一旦检测到[所描述条件或事件]”或“响应于检测到[所描述条件或事件]”。As used in the specification of this application and the appended claims, the term "if" may be contextually interpreted as "when" or "once" or "in response to determining" or "in response to detecting ". Similarly, the phrases "if it is determined" or "if the [described condition or event] is detected" may be interpreted, depending on the context, to mean "once it is determined" or "in response to the determination" or "once the [described condition or event] is detected. ]" or "in response to detection of the [described condition or event]".

另外,在本申请说明书和所附权利要求书的描述中,术语“第一”、“第二”、“第三”等仅用于区分描述,而不能理解为指示或暗示相对重要性。In addition, in the description of the specification of the present application and the appended claims, the terms "first", "second", "third", etc. are only used to distinguish the description, and should not be construed as indicating or implying relative importance.

在本申请说明书中描述的参考“一个实施例”或“一些实施例”等意味着在本申请的一个或多个实施例中包括结合该实施例描述的特定特征、结构或特点。由此,在本说明书中的不同之处出现的语句“在一个实施例中”、“在一些实施例中”、“在其他一些实施例中”、“在另外一些实施例中”等不是必然都参考相同的实施例,而是意味着“一个或多个但不是所有的实施例”,除非是以其他方式另外特别强调。术语“包括”、“包含”、“具有”及它们的变形都意味着“包括但不限于”,除非是以其他方式另外特别强调。References in this specification to "one embodiment" or "some embodiments" and the like mean that a particular feature, structure or characteristic described in connection with the embodiment is included in one or more embodiments of the present application. Thus, appearances of the phrases "in one embodiment," "in some embodiments," "in other embodiments," "in other embodiments," etc. in various places in this specification are not necessarily All refer to the same embodiment, but mean "one or more but not all embodiments" unless specifically emphasized otherwise. The terms "including", "including", "having" and their variants mean "including but not limited to" unless specifically emphasized otherwise.

下面将通过具体实施例对本申请实施例提供的技术方案进行介绍。The technical solutions provided by the embodiments of the present application will be introduced below through specific embodiments.

参见图1,为本申请实施例提供的激光点云融合方法的流程示意图,作为示例而非限定,该方法可以应用于服务器,该方法可以包括以下步骤:Referring to FIG. 1, which is a schematic flowchart of a laser point cloud fusion method provided by an embodiment of the present application, as an example and not a limitation, the method can be applied to a server, and the method can include the following steps:

步骤S101,获取待处理点云。Step S101, acquiring a point cloud to be processed.

其中,待处理点云包括第一待处理点云和第二待处理点云。The point cloud to be processed includes a first point cloud to be processed and a second point cloud to be processed.

需说明的是,第一待处理点云和第二待处理点云是根据不同类型的激光扫描设备拍摄同一目标对象生成的。It should be noted that the first point cloud to be processed and the second point cloud to be processed are generated by photographing the same target object by different types of laser scanning devices.

第一待处理点云和第二待处理点云对应同一拍摄对象,第一待处理点云可以是地面激光扫描仪拍摄得到的,第二待处理点云可以是无人机航拍得到的。The first point cloud to be processed and the second point cloud to be processed correspond to the same photographing object, the first point cloud to be processed may be photographed by a ground laser scanner, and the second point cloud to be processed may be photographed by a drone.

具体应用中,获取待处理点云,包括:In specific applications, the point cloud to be processed is obtained, including:

步骤S101-1,获取第一待处理图像和第一激光数据。Step S101-1, acquiring the first image to be processed and the first laser data.

步骤S101-2,获取第二待处理图像和第二激光数据。Step S101-2, acquiring a second image to be processed and second laser data.

其中,第一待处理图像和第一激光数据可以是地面激光扫描仪拍摄得到的,第一待处理图像可以是全景图像,第一激光数据中包含第一待处理图像中每个像素点的深度值;第二待处理图像和第二激光数据可以是无人机航拍得到的,第二待处理图像可以是普通图像,第二激光数据中包含第二待处理图像中每个像素点的深度值。另外,地面激光扫描仪还可以获取目标对象的第一GPS数据,无人机也可以获取目标对象的第二GPS数据。The first image to be processed and the first laser data may be captured by a terrestrial laser scanner, the first image to be processed may be a panoramic image, and the first laser data includes the depth of each pixel in the first image to be processed value; the second image to be processed and the second laser data can be obtained from drone aerial photography, the second image to be processed can be an ordinary image, and the second laser data contains the depth value of each pixel in the second image to be processed . In addition, the ground laser scanner can also obtain the first GPS data of the target object, and the drone can also obtain the second GPS data of the target object.

步骤S101-3,对第一激光数据所处得第一坐标系和所述第二激光数据所处得第二坐标系进行坐标系校准。Step S101-3, performing coordinate system calibration on the first coordinate system where the first laser data is located and the second coordinate system where the second laser data is located.

具体地,第一GPS数据所处得是第一坐标系,第二GPS数据所处的是第二坐标系,第一GPS数据和第二GPS数据进行坐标对齐,即对第一激光数据所处得第一坐标系和所述第二激光数据所处得第二坐标系进行坐标系校准。Specifically, the first GPS data is located in the first coordinate system, the second GPS data is located in the second coordinate system, and the coordinates of the first GPS data and the second GPS data are aligned, that is, the coordinate alignment of the first GPS data and the second GPS data is performed. Coordinate system calibration is performed on the obtained first coordinate system and the second coordinate system where the second laser data is located.

步骤S101-4,匹配第一待处理图像和第二待处理图像。Step S101-4, matching the first image to be processed and the second image to be processed.

示例性地,匹配第一待处理图像和第二待处理图像,包括:Exemplarily, matching the first image to be processed and the second image to be processed includes:

步骤S101-4-1,提取第一待处理图像的第一特征点。Step S101-4-1, extract the first feature point of the first image to be processed.

示例性地,采用特征提取算法提取第一待处理图像的第一特征点。Exemplarily, a feature extraction algorithm is used to extract the first feature point of the first image to be processed.

步骤S101-4-2,提取第二待处理图像的第二特征点。Step S101-4-2, extract the second feature point of the second image to be processed.

示例性地,采用特征提取算法提取第一待处理图像的第一特征点。Exemplarily, a feature extraction algorithm is used to extract the first feature point of the first image to be processed.

需说明的是,特征提取算法是AKAZE算法。AKAZE特征算法是SIFT特征算法的一种改进版本,但不使用高斯模糊来构建尺度空间,因为高斯模糊具有丢失边缘信息的缺点,进而采用非线性扩散滤波来构建尺度空间,从而保留图像更多的边缘特征。It should be noted that the feature extraction algorithm is the AKAZE algorithm. The AKAZE feature algorithm is an improved version of the SIFT feature algorithm, but does not use Gaussian blur to construct the scale space, because the Gaussian blur has the disadvantage of losing edge information, and then uses nonlinear diffusion filtering to construct the scale space, so as to retain more of the image. edge features.

步骤S101-4-3,筛选出第一特征点和第二特征点中共有的目标特征点。Step S101-4-3, screening out target feature points common to the first feature point and the second feature point.

示例性地,将第一特征点和第二特征点输入至预先训练的神经网络,输出共有的目标特征点。Exemplarily, the first feature point and the second feature point are input to a pre-trained neural network, and a common target feature point is output.

可选的,筛选出第一特征点和第二特征点中共有的目标特征点之前,还包包括:训练神经网络。Optionally, before filtering out the common target feature points in the first feature point and the second feature point, the method further includes: training a neural network.

可以理解的是:先用现有的算法提取粗略匹配,sfm算法进行稠密重建,筛取有效点,训练一个神经网络,实现无人机的拍摄与全景图像的匹配。It is understandable that the existing algorithm is used to extract the rough match first, the sfm algorithm is used for dense reconstruction, the effective points are screened, and a neural network is trained to realize the matching between the UAV shooting and the panoramic image.

神经网络的训练过程中包括初始步和迭代步。其中,初始步是普通图像与全景图像的粗匹配,可以把已经在模型中的全景图像切片,用全景图像切片后的普通图像与普通图像的特征点进行匹配。初始步也可以将无人机的图像转成全景图像的一部分,实现全景对全景的匹配。迭代步就直接使用无人机图像与全景图像的匹配做迭代。其中,神经网络的模型可以是对普通图像和全景图像进行特征点匹配的神经网络模型,例如是SuperGlue模型、SuperGlue模型变体或者OANet模型。The training process of a neural network includes an initial step and an iterative step. Among them, the initial step is the rough matching between the ordinary image and the panoramic image. The panoramic image already in the model can be sliced, and the ordinary image after slicing the panoramic image can be matched with the feature points of the ordinary image. The initial step can also convert the image of the UAV into a part of the panoramic image, so as to realize the matching of the panorama to the panorama. The iteration step directly uses the matching of the UAV image and the panoramic image to iterate. The model of the neural network may be a neural network model that performs feature point matching on an ordinary image and a panoramic image, such as a SuperGlue model, a variant of the SuperGlue model, or an OANet model.

步骤S101-5,根据匹配后的第一待处理图像和坐标系校准后的第一激光数据生成第一待处理点云。Step S101-5: Generate a first point cloud to be processed according to the matched first image to be processed and the first laser data after coordinate system calibration.

示例性地,根据下式得到第一待处理点云的三维坐标:Exemplarily, the three-dimensional coordinates of the first point cloud to be processed are obtained according to the following formula:

Figure DEST_PATH_IMAGE001
Figure DEST_PATH_IMAGE001
,

其中,(u,v)为第一待处理图像中每个特征点的像素坐标,d为第一待处理图像中每个像素点的深度值,K为地面激光扫描仪的内参,(X,Y,Z)为第一待处理点云的三维坐标。Among them, (u, v) is the pixel coordinate of each feature point in the first image to be processed, d is the depth value of each pixel point in the first image to be processed, K is the internal parameter of the ground laser scanner, (X, Y, Z) are the three-dimensional coordinates of the first point cloud to be processed.

步骤S101-6,根据匹配后的第二待处理图像和坐标系校准后的第二激光数据生成第二待处理点云。Step S101-6: Generate a second point cloud to be processed according to the matched second image to be processed and the second laser data after coordinate system calibration.

示例性地,根据下式得到第二待处理点云的三维坐标:Exemplarily, the three-dimensional coordinates of the second point cloud to be processed are obtained according to the following formula:

Figure 795590DEST_PATH_IMAGE001
Figure 795590DEST_PATH_IMAGE001
,

其中,(u,v)为第二待处理图像中每个特征点的像素坐标,d为第二待处理图像中每个像素点的深度值,K为无人机的内参,(X,Y,Z)为第二待处理点云的三维坐标。Among them, (u, v) is the pixel coordinate of each feature point in the second image to be processed, d is the depth value of each pixel in the second image to be processed, K is the internal parameter of the drone, (X, Y , Z) are the three-dimensional coordinates of the second point cloud to be processed.

步骤S102,融合第一待处理点云和第二待处理点云,得到目标点云。Step S102, fusing the first point cloud to be processed and the second point cloud to be processed to obtain a target point cloud.

具体应用中,融合第一待处理点云和第二待处理点云,得到目标点云,包括:In a specific application, the first point cloud to be processed and the second point cloud to be processed are fused to obtain the target point cloud, including:

步骤S102-1,根据目标特征点获取位姿信息。Step S102-1, obtaining pose information according to the target feature point.

其中,根据SFM算法对目标特征点进行匹配,得到位姿信息。Among them, the target feature points are matched according to the SFM algorithm to obtain the pose information.

步骤S102-2,根据位姿信息,对第一待处理点云和第二待处理点云进行融合,得到目标点云。Step S102-2, according to the pose information, the first point cloud to be processed and the second point cloud to be processed are fused to obtain a target point cloud.

示例性地,根据位姿信息,通过ICP算法对第一待处理点云和第二待处理点云进行融合,得到目标点云。Exemplarily, according to the pose information, the ICP algorithm is used to fuse the first point cloud to be processed and the second point cloud to be processed to obtain the target point cloud.

在一种可选的实施方式中,融合第一待处理点云和第二待处理点云,得到目标点云之后,还包括:In an optional implementation manner, after fusing the first point cloud to be processed and the second point cloud to be processed to obtain the target point cloud, the method further includes:

优化目标点云。Optimize the target point cloud.

可以理解的是,对第一待处理点云和第二待处理点云进行融合,得到目标点云之后,会有一些点云重叠,首先将合成的点云进行光栅化,对每个栅格计算两部分点云的密度A和可见度B(点云法线与相机位置和点云连线之间的夹角)可信度C = A * (cosB)。哪部分点云可信度越高,最终合成的点云选择哪一部分,从而达到优化目标点云的目的。It can be understood that after the first point cloud to be processed and the second point cloud to be processed are fused, after obtaining the target point cloud, some point clouds will overlap. First, the synthesized point cloud is rasterized, and each grid is rasterized. Calculate the density A and visibility B of the two parts of the point cloud (the angle between the point cloud normal and the camera position and the line connecting the point cloud). The credibility C = A * (cosB). Which part of the point cloud is more credible, and which part is selected for the final synthesized point cloud, so as to achieve the purpose of optimizing the target point cloud.

本申请实施例中,通过获取待处理点云;其中,待处理点云包括第一待处理点云和第二待处理点云;融合第一待处理点云和所述第二待处理点云,得到目标点云,提高激光点云融合的精度。In the embodiment of the present application, the point cloud to be processed is obtained by acquiring the point cloud to be processed; wherein the point cloud to be processed includes a first point cloud to be processed and a second point cloud to be processed; the first point cloud to be processed and the second point cloud to be processed are fused , get the target point cloud and improve the accuracy of laser point cloud fusion.

应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。It should be understood that the size of the sequence numbers of the steps in the above embodiments does not mean the sequence of execution, and the execution sequence of each process should be determined by its function and internal logic, and should not constitute any limitation to the implementation process of the embodiments of the present application.

对应于上文实施例所述的方法,图2示出了本申请实施例提供的激光点云融合装置的结构框图,为了便于说明,仅示出了与本申请实施例相关的部分。Corresponding to the methods described in the above embodiments, FIG. 2 shows a structural block diagram of the laser point cloud fusion apparatus provided by the embodiments of the present application. For convenience of description, only the parts related to the embodiments of the present application are shown.

参照图2,该装置包括:Referring to Figure 2, the device includes:

获取模块21,用于获取待处理点云;其中,所述待处理点云包括第一待处理点云和第二待处理点云;The acquiring module 21 is used for acquiring the point cloud to be processed; wherein, the point cloud to be processed includes a first point cloud to be processed and a second point cloud to be processed;

融合模块22,用于融合所述第一待处理点云和所述第二待处理点云,得到目标点云。The fusion module 22 is configured to fuse the first point cloud to be processed and the second point cloud to be processed to obtain a target point cloud.

在一种可选的实施方式中,所述获取模块,包括:In an optional implementation manner, the acquisition module includes:

第一获取子模块,用于获取第一待处理图像和第一激光数据;a first acquisition sub-module for acquiring the first image to be processed and the first laser data;

第二获取子模块,用于获取第二待处理图像和第二激光数据;a second acquisition sub-module for acquiring the second to-be-processed image and the second laser data;

校准子模块,用于对所述第一激光数据所处得第一坐标系和所述第二激光数据所处得第二坐标系进行坐标系校准;a calibration sub-module for performing coordinate system calibration on the first coordinate system where the first laser data is located and the second coordinate system where the second laser data is located;

匹配子模块,用于匹配所述第一待处理图像和所述第二待处理图像;a matching submodule for matching the first image to be processed and the second image to be processed;

第一生成子模块,用于根据匹配后的所述第一待处理图像和坐标系校准后的所述第一激光数据生成第一待处理点云;a first generation submodule, configured to generate a first point cloud to be processed according to the matched first image to be processed and the first laser data after coordinate system calibration;

第二生成子模块,用于根据匹配后的所述第二待处理图像和坐标系校准后的所述第二激光数据生成第二待处理点云。The second generation sub-module is configured to generate a second point cloud to be processed according to the matched second image to be processed and the second laser data after coordinate system calibration.

在一种可选的实施方式中,所述匹配子模块,包括:In an optional embodiment, the matching submodule includes:

第一提取单元,用于提取所述第一待处理图像的第一特征点;a first extraction unit, configured to extract the first feature point of the first image to be processed;

第二提取单元,用于提取所述第二待处理图像的第二特征点;a second extraction unit, configured to extract the second feature point of the second to-be-processed image;

筛选单元,用于筛选出所述第一特征点和第二特征点中共有的目标特征点。A screening unit, configured to screen out target feature points common to the first feature point and the second feature point.

在一种可选的实施方式中,所述融合模块,包括:In an optional embodiment, the fusion module includes:

第三获取子模块,用于根据所述目标特征点获取位姿信息;a third acquisition sub-module, configured to acquire pose information according to the target feature point;

融合子模块,用于根据所述位姿信息,对所述第一待处理点云和所述第二待处理点云进行融合,得到目标点云。The fusion sub-module is configured to fuse the first point cloud to be processed and the second point cloud to be processed according to the pose information to obtain a target point cloud.

在一种可选的实施方式中,所述装置,还包括:In an optional embodiment, the device further includes:

优化模块,用于优化所述目标点云。An optimization module for optimizing the target point cloud.

需要说明的是,上述装置/单元之间的信息交互、执行过程等内容,由于与本申请方法实施例基于同一构思,其具体功能及带来的技术效果,具体可参见方法实施例部分,此处不再赘述。It should be noted that the information exchange, execution process and other contents between the above-mentioned devices/units are based on the same concept as the method embodiments of the present application. For specific functions and technical effects, please refer to the method embodiments section. It is not repeated here.

图3为本申请实施例提供的服务器的结构示意图。如图3所示,该实施例的服务器3包括:至少一个处理器30、存储器31以及存储在所述存储器31中并可在所述至少一个处理器30上运行的计算机程序32,所述处理器30执行所述计算机程序32时实现上述任意各个方法实施例中的步骤。FIG. 3 is a schematic structural diagram of a server provided by an embodiment of the present application. As shown in FIG. 3, the server 3 of this embodiment includes: at least one processor 30, a memory 31, and a computer program 32 stored in the memory 31 and executable on the at least one processor 30, the processing When the computer 30 executes the computer program 32, the steps in any of the foregoing method embodiments are implemented.

所述服务器3可以是云端服务器等计算设备。该服务器可包括,但不仅限于,处理器30、存储器31。本领域技术人员可以理解,图3仅仅是服务器3的举例,并不构成对服务器3的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如还可以包括输入输出设备、网络接入设备等。The server 3 may be a computing device such as a cloud server. The server may include, but is not limited to, the processor 30 and the memory 31 . Those skilled in the art can understand that FIG. 3 is only an example of the server 3, and does not constitute a limitation on the server 3. It may include more or less components than the one shown in the figure, or combine some components, or different components, such as It may also include input and output devices, network access devices, and the like.

所称处理器30可以是中央处理单元(Central Processing Unit,CPU),该处理器30还可以是其他通用处理器、数字信号处理器 (Digital Signal Processor,DSP)、专用集成电路 (Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA) 或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The so-called processor 30 may be a central processing unit (Central Processing Unit, CPU), and the processor 30 may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSP), application specific integrated circuits (Application Specific Integrated Circuits) , ASIC), off-the-shelf programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.

所述存储器31在一些实施例中可以是所述服务器3的内部存储单元,例如服务器3的硬盘或内存。所述存储器31在另一些实施例中也可以是所述服务器3的外部存储设备,例如所述服务器3上配备的插接式硬盘,智能存储卡(Smart Media Card, SMC),安全数字(Secure Digital, SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器31还可以既包括所述服务器3的内部存储单元也包括外部存储设备。所述存储器31用于存储操作系统、应用程序、引导装载程序(BootLoader)、数据以及其他程序等,例如所述计算机程序的程序代码等。所述存储器31还可以用于暂时地存储已经输出或者将要输出的数据。The memory 31 may be an internal storage unit of the server 3 in some embodiments, such as a hard disk or a memory of the server 3 . In other embodiments, the memory 31 may also be an external storage device of the server 3, such as a plug-in hard disk, a smart memory card (Smart Media Card, SMC), a secure digital (Secure) Digital, SD) card, flash card (Flash Card), etc. Further, the memory 31 may also include both an internal storage unit of the server 3 and an external storage device. The memory 31 is used to store an operating system, an application program, a boot loader (Boot Loader), data, and other programs, such as program codes of the computer program. The memory 31 can also be used to temporarily store data that has been output or will be output.

所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。实施例中的各功能单元、模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中,上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。另外,各功能单元、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述系统中单元、模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for the convenience and simplicity of description, only the division of the above-mentioned functional units and modules is used as an example for illustration. In practical applications, the above-mentioned functions can be allocated to different functional units, Module completion, that is, dividing the internal structure of the device into different functional units or modules to complete all or part of the functions described above. Each functional unit and module in the embodiment may be integrated in one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit, and the above-mentioned integrated units may adopt hardware. It can also be realized in the form of software functional units. In addition, the specific names of the functional units and modules are only for the convenience of distinguishing from each other, and are not used to limit the protection scope of the present application. For the specific working processes of the units and modules in the above-mentioned system, reference may be made to the corresponding processes in the foregoing method embodiments, which will not be repeated here.

本申请实施例还提供了一种可读计算机可读存储介质,所述可读计算机可读存储介质优选为计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现可实现上述各个方法实施例中的步骤。Embodiments of the present application further provide a computer-readable storage medium, where the computer-readable storage medium is preferably a computer-readable storage medium, and the computer-readable storage medium stores a computer program, and the computer program When executed by the processor, the implementation can implement the steps in the above-mentioned various method embodiments.

本申请实施例提供了一种计算机程序产品,当计算机程序产品在移动终端上运行时,使得移动终端执行时实现可实现上述各个方法实施例中的步骤。The embodiments of the present application provide a computer program product, when the computer program product runs on a mobile terminal, the steps in the foregoing method embodiments can be implemented when the mobile terminal executes the computer program product.

所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取计算机可读存储介质中。基于这样的理解,本申请实现上述实施例方法中的全部或部分流程,可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一计算机可读存储介质中,该计算机程序在被处理器执行时,可实现上述各个方法实施例的步骤。其中,所述计算机程序包括计算机程序代码,所述计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质至少可以包括:能够将计算机程序代码携带到拍照装置/终端设备的任何实体或装置、记录介质、计算机存储器、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、电载波信号、电信信号以及软件分发介质。例如U盘、移动硬盘、磁碟或者光盘等。在某些司法管辖区,根据立法和专利实践,计算机可读介质不可以是电载波信号和电信信号。The integrated units, if implemented in the form of software functional units and sold or used as independent products, may be stored in a computer-readable computer-readable storage medium. Based on this understanding, the present application realizes all or part of the processes in the methods of the above embodiments, which can be completed by instructing the relevant hardware through a computer program, and the computer program can be stored in a computer-readable storage medium. When executed by a processor, the steps of each of the above method embodiments can be implemented. Wherein, the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file or some intermediate form, and the like. The computer-readable medium may include at least: any entity or device capable of carrying the computer program code to the photographing device/terminal device, recording medium, computer memory, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), electrical carrier signals, telecommunication signals, and software distribution media. For example, U disk, mobile hard disk, disk or CD, etc. In some jurisdictions, under legislation and patent practice, computer readable media may not be electrical carrier signals and telecommunications signals.

在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the foregoing embodiments, the description of each embodiment has its own emphasis. For parts that are not described or described in detail in a certain embodiment, reference may be made to the relevant descriptions of other embodiments.

本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。Those of ordinary skill in the art can realize that the units and algorithm steps of each example described in conjunction with the embodiments disclosed herein can be implemented in electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are performed in hardware or software depends on the specific application and design constraints of the technical solution. Skilled artisans may implement the described functionality using different methods for each particular application, but such implementations should not be considered beyond the scope of this application.

在本申请所提供的实施例中,应该理解到,所揭露的装置/网络设备和方法,可以通过其它的方式实现。例如,以上所描述的装置/网络设备实施例仅仅是示意性的,例如,所述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,装置或单元的间接耦合或通讯连接,可以是电性,机械或其它的形式。In the embodiments provided in this application, it should be understood that the disclosed apparatus/network device and method may be implemented in other manners. For example, the apparatus/network device embodiments described above are only illustrative. For example, the division of the modules or units is only a logical function division. In actual implementation, there may be other division methods, such as multiple units. Or components may be combined or may be integrated into another system, or some features may be omitted, or not implemented. On the other hand, the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, and may be in electrical, mechanical or other forms.

所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this embodiment.

以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present application, but not to limit them; although the present application has been described in detail with reference to the above-mentioned embodiments, those of ordinary skill in the art should understand that: it can still be used for the above-mentioned implementations. The technical solutions described in the examples are modified, or some technical features thereof are equivalently replaced; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions in the embodiments of the application, and should be included in the within the scope of protection of this application.

Claims (10)

1. A laser point cloud fusion method is characterized by comprising the following steps: a
Acquiring point clouds to be processed; the point clouds to be processed comprise a first point cloud to be processed and a second point cloud to be processed;
and fusing the first point cloud to be processed and the second point cloud to be processed to obtain a target point cloud.
2. The laser point cloud fusion method of claim 1, wherein obtaining the point cloud to be processed comprises:
acquiring a first image to be processed and first laser data;
acquiring a second image to be processed and second laser data;
calibrating a coordinate system of a first coordinate system where the first laser data are located and a second coordinate system where the second laser data are located;
matching the first image to be processed and the second image to be processed;
generating a first point cloud to be processed according to the matched first image to be processed and the first laser data after the coordinate system is calibrated;
and generating a second point cloud to be processed according to the matched second image to be processed and the second laser data after the coordinate system is calibrated.
3. The laser point cloud fusion method of claim 2, wherein matching the first to-be-processed image and the second to-be-processed image comprises:
extracting a first feature point of the first image to be processed;
extracting a second feature point of the second image to be processed;
and screening out target characteristic points which are common to the first characteristic points and the second characteristic points.
4. The laser point cloud fusion method of claim 3, wherein fusing the first point cloud to be processed and the second point cloud to be processed to obtain a target point cloud comprises:
acquiring pose information according to the target feature points;
and fusing the first point cloud to be processed and the second point cloud to be processed according to the pose information to obtain a target point cloud.
5. The laser point cloud fusion method of claim 1, wherein after fusing the first point cloud to be processed and the second point cloud to be processed to obtain a target point cloud, the method further comprises:
optimizing the target point cloud.
6. A laser point cloud fusion device, comprising:
the acquisition module is used for acquiring point clouds to be processed; the point clouds to be processed comprise a first point cloud to be processed and a second point cloud to be processed;
and the fusion module is used for fusing the first point cloud to be processed and the second point cloud to be processed to obtain a target point cloud.
7. The laser point cloud fusion apparatus of claim 6, wherein the acquisition module comprises:
the first acquisition submodule is used for acquiring a first image to be processed and first laser data;
the second acquisition submodule is used for acquiring a second image to be processed and second laser data;
the calibration sub-module is used for calibrating a coordinate system of a first coordinate system where the first laser data are located and a second coordinate system where the second laser data are located;
the matching submodule is used for matching the first image to be processed with the second image to be processed;
the first generation submodule is used for generating a first point cloud to be processed according to the matched first image to be processed and the first laser data after the coordinate system is calibrated;
and the second generation submodule is used for generating a second point cloud to be processed according to the matched second image to be processed and the second laser data after the coordinate system is calibrated.
8. The laser point cloud fusion apparatus of claim 7, wherein the matching sub-module comprises:
a first extraction unit, configured to extract a first feature point of the first image to be processed;
the second extraction unit is used for extracting second feature points of the second image to be processed;
and the screening unit is used for screening out the target characteristic points which are common to the first characteristic points and the second characteristic points.
9. A server comprising a memory, a processor and a computer program stored in the memory and executable on the processor, characterized in that the processor implements the method according to any of claims 1 to 5 when executing the computer program.
10. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the method according to any one of claims 1 to 5.
CN202210512885.8A 2022-05-12 2022-05-12 Laser point cloud fusion method, device, server and computer-readable storage medium Pending CN114913105A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210512885.8A CN114913105A (en) 2022-05-12 2022-05-12 Laser point cloud fusion method, device, server and computer-readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210512885.8A CN114913105A (en) 2022-05-12 2022-05-12 Laser point cloud fusion method, device, server and computer-readable storage medium

Publications (1)

Publication Number Publication Date
CN114913105A true CN114913105A (en) 2022-08-16

Family

ID=82766411

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210512885.8A Pending CN114913105A (en) 2022-05-12 2022-05-12 Laser point cloud fusion method, device, server and computer-readable storage medium

Country Status (1)

Country Link
CN (1) CN114913105A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116758006A (en) * 2023-05-18 2023-09-15 广州广检建设工程检测中心有限公司 Scaffold quality detection method and device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108230379A (en) * 2017-12-29 2018-06-29 百度在线网络技术(北京)有限公司 For merging the method and apparatus of point cloud data
CN111199579A (en) * 2020-01-02 2020-05-26 腾讯科技(深圳)有限公司 Method, device, equipment and medium for building three-dimensional model of target object
CN111369678A (en) * 2018-12-25 2020-07-03 浙江舜宇智能光学技术有限公司 Three-dimensional scene reconstruction method and system
CN112233056A (en) * 2020-10-16 2021-01-15 北京恒华伟业科技股份有限公司 Point cloud fusion method and device for multi-source data and storage medium
US20220050210A1 (en) * 2021-03-18 2022-02-17 Apollo Intelligent Connectivity (Beijing) Technology Co., Ltd. Method, apparatus for superimposing laser point clouds and high-precision map and electronic device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108230379A (en) * 2017-12-29 2018-06-29 百度在线网络技术(北京)有限公司 For merging the method and apparatus of point cloud data
CN111369678A (en) * 2018-12-25 2020-07-03 浙江舜宇智能光学技术有限公司 Three-dimensional scene reconstruction method and system
CN111199579A (en) * 2020-01-02 2020-05-26 腾讯科技(深圳)有限公司 Method, device, equipment and medium for building three-dimensional model of target object
CN112233056A (en) * 2020-10-16 2021-01-15 北京恒华伟业科技股份有限公司 Point cloud fusion method and device for multi-source data and storage medium
US20220050210A1 (en) * 2021-03-18 2022-02-17 Apollo Intelligent Connectivity (Beijing) Technology Co., Ltd. Method, apparatus for superimposing laser point clouds and high-precision map and electronic device

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116758006A (en) * 2023-05-18 2023-09-15 广州广检建设工程检测中心有限公司 Scaffold quality detection method and device
CN116758006B (en) * 2023-05-18 2024-02-06 广州广检建设工程检测中心有限公司 Scaffold quality detection method and device

Similar Documents

Publication Publication Date Title
CN109474780B (en) Method and device for image processing
US10726580B2 (en) Method and device for calibration
CN113379815A (en) Three-dimensional reconstruction method and device based on RGB camera and laser sensor and server
CN113807451B (en) Training method, device and server for panoramic image feature point matching model
CN112435193B (en) Method and device for denoising point cloud data, storage medium and electronic equipment
CN113436338A (en) Three-dimensional reconstruction method and device for fire scene, server and readable storage medium
CN113140036B (en) Three-dimensional modeling method, device, equipment and storage medium
CN111383264B (en) Positioning method, positioning device, terminal and computer storage medium
CN112927306B (en) Calibration method and device of shooting device and terminal equipment
CN113362445B (en) Method and device for reconstructing object based on point cloud data
CN111290684A (en) Image display method, image display device and terminal equipment
CN111754435A (en) Image processing method, apparatus, terminal device, and computer-readable storage medium
CN114413849A (en) Three-dimensional geographic information data processing method and device for power transmission and transformation project
CN115393815A (en) Road information generation method and device, electronic equipment and computer readable medium
CN108492284B (en) Method and apparatus for determining perspective shape of image
CN111340722B (en) Image processing method, processing device, terminal equipment and readable storage medium
CN117671159A (en) Three-dimensional model generation method and device, equipment and storage medium
CN114913105A (en) Laser point cloud fusion method, device, server and computer-readable storage medium
CN114022358A (en) Image splicing method and device for laser camera and dome camera, and server
CN113298122B (en) Target detection method and device and electronic equipment
CN112669432B (en) Volume cloud rendering method and device, electronic equipment and storage medium
CN117635875B (en) Three-dimensional reconstruction method, device and terminal
CN118115659A (en) Multi-view three-dimensional reconstruction method, device, electronic equipment and storage medium
CN111161348A (en) A method, device and device for object pose estimation based on monocular camera
CN114255268A (en) Disparity map processing and deep learning model training method and related equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination