Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a registration method of laser point cloud and sequence images of an unmanned aerial vehicle.
The invention provides a registration method of laser point cloud and sequence image of an unmanned aerial vehicle, which comprises the following steps:
step 1, acquiring LiDAR laser point clouds and sequence images acquired by an unmanned aerial vehicle, and generating MVS image dense point clouds based on the sequence images;
step 2, extracting the roof of the building in the LiDAR laser point cloud to obtain first contour extraction information; extracting the roof of the building from the MVS image dense point cloud to obtain second contour extraction information;
when building roof extraction is carried out in the MVS image dense point cloud, under the condition that POS data exists, a rough registration model is selected as a 2D-3D collinear equation registration model, the first contour extraction information is used for guiding the extraction of building registration primitives on the MVS image dense point cloud, attitude determination positioning data provided by POS are converted into external camera orientation element values, an extracted building outer frame is back-projected onto all images by using a collinear equation, the image with the complete building back-projection outer frame is marked as a key frame, and the key frame is involved in the 2D-3D rough registration model for resolving to obtain second contour extraction information;
when building roof extraction is carried out in the MVS image dense point cloud, under the condition of no POS data, building MVS image point cloud segmentation, outline extraction and regularization processing are carried out in the MVS image dense point cloud to obtain second outline extraction information;
step 3, constructing a first registration primitive image based on the first contour extraction information, and obtaining a first registration primitive image set; constructing a second registration primitive image based on the second contour extraction information, and obtaining a second registration primitive image set; matching the first registration primitive image set and the second registration primitive image set to obtain a conjugate registration primitive pair;
step 4, resolving a coarse registration model according to the conjugate registration element pair to obtain a space coordinate conversion relation between an unmanned aerial vehicle photogrammetry coordinate system and a LiDAR reference coordinate system;
and 5, realizing the registration of the MVS image dense point cloud and the LiDAR laser point cloud based on the space coordinate conversion relation.
Preferably, the specific implementation manner of step 1 is as follows: and after the sequence image is obtained, recovering external orientation elements of the sequence image collected by the calibration camera in an unmanned aerial vehicle photogrammetry coordinate system by a motion structure recovery (SfM) method, and generating the MVS image dense point cloud from the sequence image by a multi-view stereo matching method.
Preferably, in the step 2, the method of a marking point process is adopted to realize the extraction of the building roof in the LiDAR laser point cloud, so as to obtain a primary extraction profile; and regularizing the primary extraction contour by using an iteration minimum outsourcing rectangle RMBR algorithm to obtain the first contour extraction information.
Preferably, in step 3, the first contour extraction information and the second contour extraction information both include extracted outer polygons of buildings, centers of the extracted outer polygons of buildings are used as graph nodes, building roofs are used as building registration primitives, and registration primitive graphs are constructed for the extracted building registration primitives;
when matching the registered primitive images, firstly detecting the local similarity between the first registered primitive image and the second registered primitive image through matching of kernel triangles, and then measuring the global similarity between the first registered primitive image and the second registered primitive image by using GED (global similarity between coordinates); and combining the results of the local similarity and the global similarity to realize the matching of the optimal registration primitive images to obtain the conjugate registration primitive pair.
Preferably, in step 4, when the coarse registration model is a 2D-3D collinearity equation registration model, the calculating the coarse registration model includes:
using the regularized polygon corner points outside the building to solve a 2D-3D collinear equation registration model; recording the image of the back projection frame of the complete building as a key frame, and making an outer polygon of the building in the key frameThe angular point is m ═ u, v, f)TThe coordinate of the corner point of the building corresponding to the point in the LiDAR laser point cloud data is Mlas=(X,Y,Z)TThen, the collinear relationship between two angular points is expressed as:
spnpm=A[Rpnp|tpnp]Mlas
wherein A is a parameter matrix in the camera, spnpAs a proportional parameter, Rpnp、tpnpConstituting an extrinsic parameter matrix of the camera.
Preferably, in step 4, when the coarse registration model is a 3D-3D model, the calculating the coarse registration model includes:
extracting building registration elements from MVS image point cloud, matching the registration element image with registration elements in LiDAR laser point cloud, correspondingly registering a geometric model by 3D-3D space similarity transformation, and solving three-dimensional space similarity transformation by using corner pairs of regular building outer polygons conjugated by 3D as control points; let the unmanned aerial vehicle photogrammetry coordinate system CmvsWherein the outer polygon corner point of the regularized building is MmvsWith its corresponding building corner point in the LiDAR reference frame CwCoordinate in (A) is MlasThen, the conversion relationship between two corner points is defined as:
Mlas=λRMmvs+T;
wherein λ and R, T are scale, rotation and translation parameters in the coordinate transformation process, respectively.
Preferably, in the step 5, the space coordinate conversion relationship obtained by the coarse registration calculation is used as an initial value, and a variant ICP algorithm is used to realize the optimal registration between the MVS image dense point cloud and the LiDAR laser point cloud, so as to obtain the accurate registration parameters of the sequence image.
Preferably, the registration method of the laser point cloud of the unmanned aerial vehicle and the sequence image further includes: and 6, constructing a data fusion result based on the registered LiDAR laser point cloud and the sequence image data set.
Preferably, the data fusion result comprises color laser point cloud generation and real projection image.
One or more technical schemes provided by the invention at least have the following technical effects or advantages:
in the invention, LiDAR laser point clouds and sequence images collected by an unmanned aerial vehicle are obtained, and MVS image dense point clouds are generated based on the sequence images; building roof extraction is carried out in LiDAR laser point cloud to obtain first contour extraction information; extracting the roof of the building from the MVS image dense point cloud to obtain second contour extraction information; constructing a first registration primitive image based on the first contour extraction information, and obtaining a first registration primitive image set; constructing a second registration primitive image based on the second contour extraction information, and obtaining a second registration primitive image set; matching the first registration primitive image set and the second registration primitive image set to obtain a conjugate registration primitive pair; calculating a coarse registration model according to the conjugate registration element pair to obtain a space coordinate conversion relation between an unmanned aerial vehicle photogrammetry coordinate system and a LiDAR reference coordinate system; based on the space coordinate conversion relationship, the registration of the MVS image dense point cloud and the LiDAR laser point cloud is realized. According to the invention, the unmanned aerial vehicle MMS imaging data is taken as a research object, a two-step registration model application strategy is formulated according to the data characteristics of the unmanned aerial vehicle MMS imaging data, the registration element extraction and matching defined in the two-step registration model of the distance imaging and visible light imaging data and the coarse-to-fine registration algorithm are specified, the registration and fusion of the unmanned aerial vehicle MMS imaging data are completed, and the usability of the unmanned aerial vehicle MMS imaging data is improved.
Detailed Description
In order to better understand the technical solution, the technical solution will be described in detail with reference to the drawings and the specific embodiments.
The embodiment provides a registration method of laser point cloud and sequence image of an unmanned aerial vehicle, as shown in fig. 1, comprising the following steps:
step 1, acquiring LiDAR laser point clouds and sequence images acquired by an unmanned aerial vehicle, and generating MVS image dense point clouds based on the sequence images, namely MVS dense image point cloud generation.
SfM (Structure from Motion) is a means to obtain the three-dimensional Structure of an object and the parameters of a camera from a sequence of Motion images, and comprises a series of sub-processes such as homonymous point matching, relative orientation solution, beam adjustment and the like. SfM is highly consistent with the procedures of homonymous point measurement, relative orientation, aerial triangulation and the like in the traditional photogrammetry, and is a different name of the camera parameter and object space structure recovery process in photogrammetry in the field of computer vision. The method realizes camera external parameter calibration and sparse object space structure recovery by using an Incremental beam method Adjustment (SfM) algorithm. In the algorithm, homonymous point matching is realized through Sift, a RANSAC five-point resolving method is adopted in a relative orientation process, and adjustment of an incremental beam method is realized by a sparse adjustment software package. On the basis of camera parameter calibration, an MVS image point cloud is generated by using a Daisy algorithm, and reconstruction of object space dense point cloud is achieved. Fig. 2 is a top and side screenshot of a set of MVS image dense point clouds generated from drone MMS sequence image data. Except noise and uneven point density distribution, the MVS image dense point cloud has data essence and expression form similar to those of LiDAR laser point cloud, and is discrete sampling, namely discrete three-dimensional point cloud, of the three-dimensional surface of an object space object.
Step 2, extracting the roof of the building in the LiDAR laser point cloud to obtain first contour extraction information; and extracting the roof of the building from the MVS image dense point cloud to obtain second contour extraction information, namely extracting the registration primitive of the building.
The MMS imaging data of the unmanned aerial vehicle is divided into LiDAR laser point cloud and sequence images which are collected by the same machine and different machines according to a data collection mode. LiDAR laser point clouds acquired by the same machine of an unmanned aerial vehicle MMS and sequence images have direct geospatial reference as initial registration values, and the registration process of a two-step registration model can be accelerated. Sequence images of unmanned aerial vehicles acquired by the same machine as LiDAR laser point cloud are generally acquired by a small and light system (such as a small fixed wing unmanned aerial vehicle and an eight-rotor unmanned aerial vehicle), the data generally only have navigation-level precision POS or direct geographical orientation data without POS, and under the condition of the data, only artificial structures contained in the data are used as registration primitives to perform two-step registration model registration. The extraction of the registration primitive comprises two parts:
extracting registration elements from LiDAR point clouds.
Referring to fig. 3, the building roof has good distinguishability in both airborne LiDAR laser point clouds and sequence images, and contains rich structural information (points, lines, planes). The method for selecting the marking point process is adopted to realize the extraction of the building point cloud in the LiDAR laser point cloud data, which is shown in (a) in figure 3 and (b) in figure 3. Due to the fact that the laser point cloud is blocked and the like, point cloud scanning is incomplete, and the extracted building point cloud blocks are incomplete. To form meaningful building outer boundary polygons, the building outer polygon extraction and regularization are performed using RMBR (iterative Minimum Bounding Rectangle) algorithm. The RMBR uses rectangles as regularization primitives, and uses a group of rectangle combinations to realize regularization on the original outline. Fig. 3 (c) shows the regularized outer frame of the building, and it can be seen that after contour extraction and regularization, the original cloud of building points is transformed into a geometrically meaningful outer polygon of the building.
And extraction of registration elements in the sequence image.
The selection of the registration model (2D-3D or 3D-3D model) is different according to the existence of POS data of the sequence images and the rough registration process. The extraction of the registration elements in the sequence image can be carried out by adopting different methods, namely an image building element extraction method guided by laser building extraction prior knowledge under the condition of POS data and an MVS image point cloud building registration element extraction method generally used for POS data.
1) And (3) extracting image building elements under the guidance of the prior knowledge of laser building extraction.
Under the condition of POS data, selecting a rough registration model as a 2D-3D collinear equation registration model, guiding the extraction of building registration elements on a sequence image by using a building extraction result (namely the first contour extraction information) corresponding to LiDAR laser point cloud data, converting attitude determination positioning data provided by POS into an external camera orientation element value, back-projecting an extracted building outer frame onto all images by using a collinear equation, recording the image with the complete building back-projection outer frame as a key frame, and participating in the 2D-3D rough registration model for resolving to obtain second contour extraction information. And converting the attitude determination positioning data provided by the POS into an element value of an external orientation of the camera, and back-projecting the extracted building outer frame onto all images by using a collinear equation. And recording the image with the back projection outer frame of the complete building as a key frame, and adding the key frame into a 2D-3D coarse registration model for calculation. The building outer frame after the back projection provides good priori knowledge for the extraction of image building elements based on local saliency. However, due to POS data quality and system calibration and sensor synchronization errors, there is a severe position and direction offset between the building frame after back projection and the real image building position, which requires further correction.
The present invention uses tensor gradient statistics within the R1 buffer region to determine the area in which a building is located in the image. A buffer is created for the backprojection region R1, the buffer width viewing region within which the tensor gradient statistics are performed for the initial backprojection error determination (empirical values 50-200 pixels), as shown in fig. 4.
For a multi-channel image, the structure tensor is defined as:
wherein (represents a Gaussian kernel convolution operation, fx,fyRepresenting the gradient in the horizontal and vertical directions.
The structure tensor describes the local differential structure of the image, and is mostly researched in the detection of the corner points and the edges of the image. For color unmanned aerial vehicle image f ═ (R, G, B)T. After the spatial derivation operation is performed, two eigenvalues of the tensor G are calculated according to the following formula:
λ1denotes the local derived Energy (Derivative Energy.) in the principal direction, λ2Representing the local derivative energy in the vertical main direction. The principal direction of the tensor G is set as the tensor gradient direction (theta), and the corresponding lambda is1The magnitude of the tensor gradient after Non-local Maximum Suppression (Non-Maximum Suppression) is shown in fig. 5.
By the method, the tensor gradient size and direction of each pixel point in the R1 buffer area can be calculated. And counting the tensor gradient direction in the region to obtain a tensor gradient statistical histogram in the buffer region. Fig. 5 (a) and 5 (b) are tensor gradient direction statistical diagrams for a rectangular house and an L-shaped house, respectively. As can be seen from fig. 5, the long edge of a rectangular or L-shaped house will form a certain gradient direction peak and the shorter edge another lower peak.
All LiDAR building outer frames are back-projected on the image, tensor gradient statistical graphs in a back-projection buffer area are analyzed, areas without the statistical characteristics of rectangular and L-shaped tensor gradient directions are removed, and a back-projection area R1 with the characteristics of single-peak and double-peak tensor gradient directions in the areas is reserved. Further refinement of the area coverage of the building is achieved by:
1. rotating the R1 back projection area to the vertical direction of the tensor gradient direction statistical peak, namely the long side direction of the building;
2. and constructing a buffer area by taking the R2 area as a core, performing buffer sliding in the main direction of R2 and the vertical main direction, counting the sum of tensor gradient sizes in the area formed by R2 and the buffer area, stopping sliding at a local maximum response position, and taking the area as an optimal area for extracting the image building elements.
After the area where the image building is located is determined, the building in the local area presents high global significance. The building image is segmented in the R3 Region using a segmentation method (RCC) based on global Contrast saliency detection (see fig. 6 (a), fig. 6 (b)). The segmentation result is subjected to contour extraction (see (c) in fig. 6), and is subjected to regularization by using RMBR algorithm (see (d) in fig. 6), so that a regularized image building top surface extraction result can be obtained, and the process is shown in fig. 6.
2) And extracting registration elements of the MVS image point cloud building.
Under the condition that the sequential image has no POS direct directional data or the POS data precision is poor, the image building extraction method guided according to the LiDAR laser point cloud building extraction result is not applicable. Under the condition, a method for extracting building registration primitives from LiDAR laser point clouds is used, and building MVS image point cloud segmentation, outline extraction and regularization are directly carried out from MVS image dense point cloud data generated by a sequence image. Fig. 7 is a schematic diagram of a process of performing building point cloud extraction and regularization on a set of MVS image dense point cloud data, which includes (a) in fig. 7, (b) in fig. 7, and (c) in fig. 7. As can be seen from (c) in fig. 7, although the dense point cloud of the MVS image has the data degradation problems such as noise and uneven density, the edge of the regularized outline of the building can still summarize the building boundary information.
Step 3, constructing a first registration primitive image based on the first contour extraction information, and obtaining a first registration primitive image set; constructing a second registration primitive image based on the second contour extraction information, and obtaining a second registration primitive image set; and matching the first registration primitive image set and the second registration primitive image set to obtain a conjugate registration primitive pair, namely construction and matching of registration primitive images.
Under the support of POS data, the primitive matching process is completed in the image building primitive extraction process guided by the prior knowledge of laser building extraction. Under the condition of no POS data support, the registration primitive image matching method defined in the two-step registration model is used for realizing the matching of extracting registration primitives, and generating a conjugate registration primitive pair for resolving the coarse registration model.
And according to the composition rule of the registration primitive graph, taking the center of the extracted outer polygon of the building as a graph node, taking the roof of the building as a registration primitive of the building, and respectively constructing the registration primitive graph for the extracted registration primitive of the building. Fig. 8 (a) and 8 (b) are schematic diagrams illustrating the process of registering the primitive map by the sequence image (MVS image dense point cloud) and the LiDAR laser point cloud respectively, wherein the triangle is the kernel triangle corresponding to the current map. The left columns of (a) in fig. 8 and (b) in fig. 8 are both extracted registration primitives, the middle columns of (a) in fig. 8 and (b) in fig. 8 are both generated registration primitive maps, and the right columns of (a) in fig. 8 and (b) in fig. 8 are both registration primitive maps and original registration primitives are displayed in an overlapping manner.
After the registration primitive map set is generated, the MVS image dense point cloud registration primitive map and the LiDAR laser point cloud registration primitive map are registered according to a registration primitive map matching method defined in the two-step registration model. In the registration primitive image matching process, local similarity of the registration primitive images is detected through matching of kernel triangles. And for the registration primitive graphs corresponding to the similar kernel triangles, measuring the global similarity by using GED. The registration pairs with similar kernel triangles (local similarity) and large GED distances (low global similarity) shown in FIG. 9 are removed, and finally the optimal registration primitive image matching is realized.
And 4, resolving a coarse registration model according to the conjugate registration element pair to obtain a space coordinate conversion relation between the unmanned aerial vehicle photogrammetry coordinate system and the LiDAR reference coordinate system.
Different registration element extraction and matching methods correspond to different coarse registration geometric models. And extracting and matching image building elements under the guidance of the prior knowledge of laser building extraction, and directly extracting registration elements on the image to correspond to the 2D-3D collinearity equation registration model. And extracting building registration elements from the MVS image dense point cloud, and matching the registration element map with the registration elements in the LIDAR laser point cloud to correspondingly transform the registration geometric model by 3D-3D space similarity. Two model solution methods are described below:
(1) and (5) solving a 2D-3D model coarse registration model.
And solving the 2D-3D registration model by using the regularized polygon corner points outside the building. Recording the image of the back projection outer frame of the complete building as a key frame, and setting the outer polygon corner point of one building in the key frame as m ═ u, v, f)TThe coordinate of the corner point of the building corresponding to the LiDAR data is Mlas=(X,Y,Z)TThen, the collinear relationship between two corner points can be expressed as:
spnpm=A[Rpnp|tpnp]Mlas
where A is a known in-camera parameter matrix, spnpAs a proportional parameter, Rpnp,tpnpConstituting an extrinsic parameter matrix of the camera. Equation describes a collinear problem of object point and image point, which is solved linearly by the present invention using the EPnP algorithm. The regular outer polygon of the building simultaneously provides geometric characteristics of collinearity, coplanarity and the like, and can be used as a constraint condition to carry out iterative optimization on an EPnP result, such as a line-to-coplanar constraint registration algorithm.
(2) And (4) solving a 3D-3D model coarse registration model.
And (3D) using the corner point pair of the regularized outer polygon of the building, which is subjected to 3D conjugation (conjugation refers to the fact that a registration primitive pair is extracted by two methods under the support of POS data or not), as a control point, and calculating three-dimensional space similarity transformation. In a photogrammetric coordinate system (C)mvs) The angular point of the regular outer polygon of the middle building is MmvsWith its corresponding corner point in the LiDAR reference frame (C)w) Coordinate in (A) is MlasThe conversion relationship therebetween is defined as Mlas=λRMmvs+ T. Where λ, R, T are the scale, rotation and translation parameters, respectively. Using the robust SVD method for each 3 graph node pairs of the n graph matching node pairs, a set (λ, R, T) can be solved.
And 5, realizing the registration of the MVS image dense point cloud and the LiDAR laser point cloud based on the space coordinate conversion relation.
And taking a space coordinate conversion relation obtained by coarse registration calculation as an initial value, and using a variant ICP algorithm to realize the optimal registration between the MVS image dense point cloud and the LiDAR laser point cloud so as to obtain accurate registration parameters of the sequence image.
Among them, the ICP algorithm and its Variants (Variants of the ICP algorithm) can be decomposed into 5 steps (rusinfikiewicz and Levoy,2001), namely: (1) point Selection (Selection of Points); (2) point Matching (Matching Points); (3) matching weights (Weighting of Matches); (4) matching pair generation rules (Rejecting Pairs); (5) error definition and Minimization method (Error Metric and Minimization).
By the data registration method, data registration errors between the sequence images and the LiDAR laser point clouds are eliminated, and coordinate references of different source data are unified.
And 6, constructing a data fusion result based on the registered LiDAR laser point cloud and the sequence image data set.
On the basis of the registration of the laser point cloud of the unmanned aerial vehicle and the sequence image, a series of data fusion processing of the sequence image and the LiDAR point cloud can be expanded, such as color laser point cloud generation and real projective images.
The color laser point cloud generation is to assign the registered image color information to LiDAR point cloud data.
A simple color point generation algorithm is described below:
1) selecting a color-imparting image to ensure that the measuring area is completely covered;
2) cutting point cloud data corresponding to the current coloring image, carrying out shielding detection according to camera parameters of the current coloring image, and determining visible laser point cloud data in the current coloring image;
3) carrying out coloring on the currently visible and non-colored three-dimensional point cloud according to a collinear equation;
4) and (5) circulating the step (2-3) until all the images are traversed or all the point clouds are colored.
Referring to fig. 10 (including fig. 10 (a), (b), and (c)) the unmanned aerial vehicle operates at a high altitude, a projection difference formed by the ground object height in the acquired image data is large, and if no occlusion detection is performed, a Double-Mapping error occurs by directly applying a collinearity equation for color assignment. Double-Mapping refers to an error generated by making a DOM using DSM for ordinary orthorectification. Fig. 10 (a) shows the cause of the above: the ground object fluctuation (2-3) forms shielding on the ground (4-5) area, and color is given by directly using a collinear condition equation under the condition of not carrying out shielding detection, so that the color error in the section (2-3) is copied into the section (4-5), and the color error in the section (4-5) is given.
Occlusion detection is performed before color application, so that the error can be effectively avoided, and the commonly used occlusion detection comprises a Z-buffer method and the like. Considering the sparsity of laser Point clouds, the optimal thickness of Z-buffer light is difficult to determine, the invention directly carries out shielding detection from the laser Point clouds, carries out space closing operation on a visual domain graph output by an HPR (high Point Removal) algorithm, determines a shielding region, and enhances the adaptability of the shielding region on the processing of Point cloud data of a sparse unmanned aerial vehicle. FIG. 11 depicts occlusion detection and color application after detection is complete. Fig. 11 (a) is a color image, and fig. 11 (b) is a visible region detected by the HPR algorithm, where white is the visible region and black is the invisible region. Fig. 11 (b) shows that the output of the original HPR is poor in adaptability to the sparse point cloud due to its algorithm limitation. The invention optimizes the misdetection of the occlusion region by using closed operation, wherein the criterion is a closed operation morphological operation criterion, namely if the occlusion regions are necessarily communicated, the occlusion regions do not exist independently. The independent occlusion region of the HPR output is culled according to this criterion, as shown in (c) of fig. 11. Fig. 11 (d) shows the result of coloring the point cloud after occlusion analysis, and it can be seen that Double-Mapping is well suppressed.
The registration method of the laser point cloud and the sequence image of the unmanned aerial vehicle provided by the embodiment of the invention at least comprises the following technical effects:
(1) LiDAR laser point cloud and sequence images collected by the same machine of an unmanned aerial vehicle MMS are adopted, and the direct geospatial reference is used as an initial registration value, so that the registration process of a registration model is accelerated.
(2) And a conjugate registration element with or without POS data extraction is adopted, and the registration result is optimized due to different registration calculation models.
Finally, it should be noted that the above embodiments are only for illustrating the technical solutions of the present invention and not for limiting, and although the present invention has been described in detail with reference to examples, it should be understood by those skilled in the art that modifications or equivalent substitutions may be made on the technical solutions of the present invention without departing from the spirit and scope of the technical solutions of the present invention, which should be covered by the claims of the present invention.