[go: up one dir, main page]

CN114119866B - A visual evaluation method for urban road intersections based on point cloud data - Google Patents

A visual evaluation method for urban road intersections based on point cloud data Download PDF

Info

Publication number
CN114119866B
CN114119866B CN202111340586.2A CN202111340586A CN114119866B CN 114119866 B CN114119866 B CN 114119866B CN 202111340586 A CN202111340586 A CN 202111340586A CN 114119866 B CN114119866 B CN 114119866B
Authority
CN
China
Prior art keywords
point cloud
point
cloud data
points
dimensional
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111340586.2A
Other languages
Chinese (zh)
Other versions
CN114119866A (en
Inventor
张家钰
程建川
刘佳玲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southeast University
Original Assignee
Southeast University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southeast University filed Critical Southeast University
Priority to CN202111340586.2A priority Critical patent/CN114119866B/en
Publication of CN114119866A publication Critical patent/CN114119866A/en
Application granted granted Critical
Publication of CN114119866B publication Critical patent/CN114119866B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/06Topological mapping of higher dimensional structures onto lower dimensional surfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/60Rotation of whole images or parts thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration using local operators
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration using histogram techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • G06T2207/10044Radar image

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种基于点云数据的城市道路交叉口可视化评价方法,通过采集道路交叉口的点云数据,对点云数据进行路面识别分割处理,将路面部分于其他离散的平面点区分开,以得到完整的路面点云数据,建立平面格网点阵,以识别到的路面点作为基准插值参照,利用邻近点插值算法获取到平面格网点的高度信息,构建数字高程模型,以驾驶员视线起点为原点构建坐标系,将三维坐标系转换成二维坐标系,再将二维直角坐标系转换为极坐标系,确定阻挡驾驶员视线的障碍物位置,本发明利用圆柱形透视投影模型,实现将三维问题转换为二维问题,构建可视域模型,分析确定影响驾驶员可视性的障碍物位置并对交叉口的可视化进行分析。

The present invention discloses a visualization evaluation method for urban road intersections based on point cloud data. The method collects point cloud data of road intersections, performs road surface recognition and segmentation processing on the point cloud data, separates the road surface part from other discrete plane points to obtain complete road surface point cloud data, establishes a plane grid dot matrix, uses the recognized road surface points as benchmark interpolation references, uses a neighboring point interpolation algorithm to obtain height information of the plane grid points, constructs a digital elevation model, constructs a coordinate system with the starting point of the driver's line of sight as the origin, converts the three-dimensional coordinate system into a two-dimensional coordinate system, and then converts the two-dimensional rectangular coordinate system into a polar coordinate system to determine the position of obstacles blocking the driver's line of sight. The present invention uses a cylindrical perspective projection model to realize the conversion of a three-dimensional problem into a two-dimensional problem, constructs a visual domain model, analyzes and determines the position of obstacles that affect the driver's visibility, and analyzes the visualization of the intersection.

Description

Urban road intersection visual evaluation method based on point cloud data
Technical Field
The invention relates to the technical field of road three-dimensional vision distance detection, in particular to a visual evaluation method for an urban road intersection based on point cloud data.
Background
The urban road intersection is a multiple road section of traffic safety accidents, and in the road design stage, the line-of-sight triangle of the intersection is used as an important visibility evaluation index to ensure good intersection visibility. However, the real-time three-dimensional road intersection environment is more complex than the road at the design stage, and many design stages do not consider or consider the wide variety of potential line-of-sight obstacles such as vegetation growth, new facilities, debris accumulation, etc. The current means for acquiring the road infrastructure data such as a mapping-level laser radar, an unmanned aerial vehicle, a communication satellite and the like have the advantages of high data precision and large data volume, and have obvious advantages for acquiring the traffic road data on a macroscopic scale. However, the method has certain limitation on microscopic-scale traffic infrastructures such as intersections and the like, and mainly has the defects of influencing traffic travel, high equipment loss cost and the like. The small laser radar has low cost and simple operation, but the application of the small laser radar in engineering practice is limited due to the excessively small view angle of 38.4 degrees. The three-dimensional vision distance calculation is a basis for visual evaluation of urban road intersections, and the domestic road three-dimensional vision distance calculation method fundamentally simplifies the road three-dimensional environment to different degrees, and is greatly limited in application at road intersections with complex traffic flow composition and more road side barriers. Many methods of three-dimensional viewing distance of foreign roads can only be called 2.5D due to modeling limitations, and cannot handle objects at the same location but with different heights. A few 3D model-based calculation methods have the limitations of complex algorithm and long calculation time, and have a further improvement space. The existing point cloud data acquisition mode and intersection visual evaluation program have great defects in economy and high efficiency, and evaluation consideration for actual driver conditions and road environment conditions is not comprehensive enough.
Disclosure of Invention
Aiming at the defects in the prior art, the invention provides a visual evaluation method for the urban road intersection based on point cloud data, which considers different parameters such as the height, the range, the visual angle position and the like of the sight of a driver.
In order to achieve the above purpose, the present invention adopts the following technical scheme:
in a first aspect, an embodiment of the present invention provides a visual evaluation method for an urban road intersection based on point cloud data, where the three-dimensional line-of-sight calculation method for the road intersection includes the following steps:
S1, collecting point cloud data of a road intersection;
s2, carrying out pavement identification and segmentation processing on the point cloud data, and distinguishing a pavement part from other discrete plane points to obtain complete pavement point cloud data;
S3, establishing a plane grid lattice, taking the identified road surface points as basic interpolation references, acquiring the height information of the plane grid points by utilizing an adjacent point interpolation algorithm, and constructing a digital elevation model;
S4, the visual field concept is quantized into a fan shape formed by countless rays emitted by a sight original point of a driver, and the angle of the fan shape is related to the visual angle range of the driver;
Constructing a coordinate system by taking a sight starting point of a driver as an origin, converting a three-dimensional coordinate system into a two-dimensional coordinate system, converting a two-dimensional rectangular coordinate system into a polar coordinate system, and determining the position of an obstacle blocking the sight of the driver;
S5, analyzing variable factors which possibly influence the visual field of the urban road intersection.
Further, in step S1, the process of collecting the point cloud data of the road intersection includes the following steps:
s11, fixing a small laser radar on a rotary tripod head, placing the small laser radar in the center of an urban road intersection, and keeping a laser radar carrying bracket fixed;
And S12, performing data splicing on the obtained point cloud file through the corresponding characteristic points to obtain complete urban road intersection point cloud data.
Further, in step S2, the process of performing the road surface recognition and segmentation processing on the point cloud data includes the following steps:
S21, rasterizing original three-dimensional point cloud data, wherein each rasterized point of the point cloud data has a corresponding linear index value, data points at the same position of the same matrix have the same linear index value, and when each data point spans one matrix, the linear index value zeta of the data point is stepped;
S22, screening out points with lower elevations in the columnar units by utilizing a height threshold h t, and analyzing the remaining data points by using a principal component analysis method to obtain the principal directions of the columnar units;
S23, sequencing the detected road points by using a height histogram method, and extracting features of the obtained plane point cloud data; the characteristic variables comprise the number of point clouds in a unit area, and standard deviation and median of adjacent term differences along the x and y directions for describing the distribution characteristics of the point clouds on an x-y plane, so as to obtain an observation value array consistent with the number of columnar units; dividing the cylindrical units into three types of uniform plane points, non-uniform plane points and pseudo plane points by using an unsupervised classification K-Means clustering algorithm, carrying out standardization processing on all variables to eliminate the variability of the variables so that the intervals of the variables fall in [0,1 ];
And S24, distinguishing the pavement part from other discrete plane points by using an ultra-voxel clustering method to obtain complete pavement point cloud data.
Further, in step S21, the process of rasterizing the original three-dimensional point cloud data includes the following steps:
S211, assuming that the original three-dimensional point cloud data space coordinates are (X, Y, Z) T, creating a temporary array (X, Y, Z) T:
S212, assuming that (min x, max y) T and (max x, min y) T are the start point and the end point of the linear index calculation, after the point cloud data is rasterized, the horizontal distance D x and the vertical distance D y between the start point and the end point are calculated by a formula, wherein D x and D y are both positive integers, and the calculation formula is as follows:
S213, establishing a zero value matrix ψ and an empty set cell matrix omega with the size of (1+D y)×(1+Dx), calculating to obtain a horizontal distance D x and a vertical distance D y from any grid point to the starting point of a data index, using a row number 1+d y and a column number 1+d x to represent elements in the zero value matrix ψ and the empty set cell matrix omega, and converting the row number and the column number into a linear index represented by ζ through a formula ζ=d x·(1+Dy)+dy +1;
When the point cloud data are segmented, the point cloud data are arranged in an ascending order according to the value of the linear index zeta, whether the point is a step point or not is judged by calculating the difference delta zeta of adjacent data points, if delta zeta is smaller than 1, the point is not the step point, otherwise, the point cloud data between the step point zeta j and the last step point zeta j-1 are stored in a zeta j-1 unit of a cellular omega matrix, and meanwhile, the zeta j-1 element of the psi matrix is assigned to be 1.
Further, in step S23, the processing procedure of the height orthogonal diagram sorting method includes the following steps:
S231, setting the height delta h of a single stripe in the height histogram to be 1-4 m, establishing a frequency histogram along the Z direction, and simultaneously establishing an empty cell array phi;
S232, sorting the frequency numbers according to the frequency number descending order, and setting F { F 1,F2...Fi...Fn|1≤i≤n,i,n∈N+ } as the height value corresponding to each arranged stripe;
S233, establishing a loop i=2, and k=1, and merging and storing the point cloud data corresponding to F i-1 to Φ { k };
S234, if F i-Fi-1≤δh, merging and storing the point cloud data corresponding to F i and F i-1 to phi { k }, otherwise, storing the point cloud data corresponding to F i to phi { k }, wherein k=k+1;
S235, let i=i+1, if i is less than or equal to N h, return to step S234, otherwise end the loop.
Further, in step S24, the processing procedure of the super voxel clustering method is as follows:
And searching to obtain three-dimensional point cloud data in the corresponding cell matrix by utilizing the index position information, splicing the three-dimensional point cloud data to obtain a final road surface super-voxel clustering result to obtain complete road surface point cloud data, and distinguishing a road surface part from other discrete plane points.
Further, in step S4, the processing procedure of converting the three-dimensional coordinate point into the two-dimensional coordinate point and converting the three-dimensional problem into the two-dimensional problem through the cylindrical perspective projection model is as follows:
Setting a viewpoint starting point as an origin, setting coordinates as (X m,ym,zm)T, setting a vehicle advancing direction as a Y 'axis, wherein the X' axis is parallel to an XOY plane and perpendicular to the Y axis, and the Z 'axis is perpendicular to the X' OY plane, and constructing a three-dimensional space coordinate system;
Constructing a corresponding two-dimensional coordinate system u-v, wherein the origin of the two-dimensional coordinate system is arranged on a Y ' axis, the distance from the origin of the sight line is R, the v axis is arranged in parallel with a Z ' axis along a vector, and the u axis is arranged as a clockwise arc section and is in vertical relation with the v axis and the Y ';
in two spatial coordinate systems, the coordinate calculation formula of the two-dimensional point on the cylindrical surface is as follows:
Wherein:
(x, y, z) T represents coordinates of points around the driver in geodetic space;
(x m,ym,zm)T represents line-of-sight origin coordinates;
(X ', Y', Z ') T represents coordinates converted into an X' Y 'Z' coordinate system;
λ represents a scale factor between two reference frames, and has a value of 1.0;
θ represents the rotation matrix of the line-of-sight rigid transformation;
θ xyz represents a rotation matrix around X, Y and the Z axis;
Alpha m represents the rotation angle around the Z axis;
gamma m represents the rotation angle around the X-axis;
R represents the radius of the cylindrical surface and takes a value of 3.0;
(u, v) T represents the coordinates of the cylindrical surface;
Alpha m and gamma m respectively represent the azimuth and vertical angles of the origin of the line of sight, the direction of the Y 'axis being coincident with the direction of advance of the vehicle, so in this case the direction of the Y' axis is coincident with the direction of advance of the vehicle;
In the local three-dimensional space of the local construction, the coordinate calculation formulas of all the sight-line end points are as follows:
η=θ/θres+1
[ρ]={ρ12...ρj...ρη-1ηi=D(1≤j≤η),η=θ/θres+1}
[y′e,x′e]=Polar To Cartesian([θ,ρ])
[z′e]={z1,z2...zj...zη-1,zη|zi=D·tanθv(1≤j≤η),η=θ/θres+1}
Wherein:
[ x' e,y′e,z′e]T represents the three-dimensional coordinates of the line of sight end in the local three-dimensional coordinate system;
θ represents the driver's viewing angle;
θ res represents the angular spacing between view lines;
j represents the line of sight index, and when j is more than or equal to 1 and less than or equal to eta, eta is the line of sight index;
The area where the coordinates of the projection points are located is regarded as a square area with the side length of W i, the coordinates of the projection points are set as the center position of the area Setting collectionsStoring two-dimensional coordinates of all sight-line end points, setting the coordinates of the target point as Ob i(xi,yi,zi)T, searching by using a KD tree algorithm and using a Chebyshev sequence with the parameter of W i/2Points within the neighborhood;
Setting a set psi i to store two-dimensional coordinates of all points in a square area where the projection points are located, wherein all two-dimensional points of the set psi i have three-dimensional coordinates corresponding to the two-dimensional coordinates, setting kappa i to represent the three-dimensional space distance between the sight origin and the target point, and excluding any points with the three-dimensional space distance exceeding kappa i from the sight origin from the set psi i. The beneficial effects of the invention are as follows:
1. According to the invention, the small-sized laser radar is fixed on the holder, the point cloud data are collected in a plurality of rotations and are spliced, so that the problems of small visual angle, incomplete data collection and inaccuracy of the small-sized laser radar are solved.
2. The point cloud data subunit dividing method based on the linear index improves the efficiency of data blocking and indexing.
3. The pavement points in the point cloud data are rapidly and efficiently segmented and extracted by adopting the preliminary filtering method, the height histogram sorting method and the super voxel clustering method.
4. According to the invention, a digital elevation model is adopted to convert a space coordinate system and a local plane coordinate system according to a cylindrical perspective projection model, a visual field model is constructed, the position of an obstacle affecting the visibility of a driver is analyzed and determined, and the visualization of an intersection is analyzed.
5. According to the method for calculating the three-dimensional apparent distance, the processing and calculating efficiency of the point cloud data are improved, and the processing difficulty of mass point cloud data is solved.
6. According to the invention, the transformation from a space three-dimensional coordinate point to a two-dimensional coordinate point is realized through the cylindrical perspective projection model, and the three-dimensional problem is transformed into the two-dimensional problem.
Drawings
Fig. 1 is a schematic structural diagram of an urban road intersection visual evaluation method based on point cloud data.
FIG. 2 is a schematic diagram of a method for constructing a digital elevation model according to the present invention.
FIG. 3 is a schematic diagram showing the effect of the preliminary filtering method of the present invention.
Fig. 4 is a schematic diagram of the structure of the height histogram method of the present invention.
Fig. 5 is a schematic structural diagram of the cylindrical perspective projection calculation of the present invention.
Fig. 6 is a view blind zone structure diagram of the present invention.
Fig. 7 is a schematic diagram showing the correspondence between the visual field and the line-of-sight curve according to the present invention.
Detailed Description
The invention will now be described in further detail with reference to the accompanying drawings.
It should be noted that the terms like "upper", "lower", "left", "right", "front", "rear", and the like are also used for descriptive purposes only and are not intended to limit the scope of the invention in which the invention may be practiced, but rather the relative relationship of the terms may be altered or modified without materially altering the teachings of the invention.
Example 1
Fig. 1 is a schematic structural diagram of an urban road intersection visual evaluation method based on point cloud data. FIG. 2 is a schematic diagram of a method for constructing a digital elevation model according to the present invention. The embodiment is suitable for a three-dimensional line-of-sight calculating method of an urban road intersection of point cloud data, and the three-dimensional line-of-sight calculating method of the road intersection comprises the following steps:
S1, collecting point cloud data of a road intersection.
S2, carrying out pavement identification and segmentation processing on the point cloud data, and distinguishing the pavement part from other discrete plane points to obtain complete pavement point cloud data.
And S3, establishing a plane grid lattice, taking the identified road surface points as basic interpolation references, acquiring the height information of the plane grid points by utilizing an adjacent point interpolation algorithm, and constructing a digital elevation model.
S4, the visual field concept is quantized into a fan shape formed by countless rays emitted by a sight original point of a driver, the angle of the fan shape is related to the visual angle range of the driver, the three-dimensional coordinate point in space is converted into a two-dimensional coordinate point through a cylindrical perspective projection model, and the three-dimensional problem is converted into a two-dimensional problem.
And constructing a coordinate system by taking the starting point of the sight of the driver as an origin, converting the three-dimensional coordinate system into a two-dimensional coordinate system, converting the two-dimensional rectangular coordinate system into a polar coordinate system, and determining the position of the obstacle blocking the sight of the driver.
S5, analyzing variable factors which possibly influence the visual field of the urban road intersection.
1. Collecting point cloud data of road intersections
In the step S1, the process of collecting the point cloud data of the road intersection includes the following sub-steps:
S11, fixing the small-sized laser radar on a rotary tripod head, placing the small-sized laser radar in the center of an urban road intersection, keeping a laser radar carrying support fixed, acquiring subfiles of the overall point cloud data of the intersection through the rotary tripod head for multiple data acquisition, and enabling the overlapping ratio of 2 adjacent point cloud files to be 1/3.
And S12, performing data splicing on the obtained point cloud file through the corresponding characteristic points to obtain complete urban road intersection point cloud data.
2. Road surface recognition and segmentation processing for point cloud data
In step S2, the process of performing the road surface recognition segmentation processing on the point cloud data includes the following sub-steps:
And S21, rasterizing the original three-dimensional point cloud data, wherein each rasterized point of the point cloud data has a corresponding linear index value, the data points at the same position of the same matrix have the same linear index value, and when each data point spans one matrix, the linear index value zeta of the data point is stepped. The specific rasterization processing steps are as follows:
Firstly, the original three-dimensional point cloud data is subjected to rasterization, and assuming that the spatial coordinates of the original three-dimensional point cloud data are (X, Y, Z) T, a temporary array (X, Y, Z) T,(x,y,z)T is created and calculated by the following formula:
Wherein [ (surface ] represents a rounded symbol, ∈ xyz represents grid intervals along the X-axis, Y-axis and Z-axis, respectively, and the smaller the interval is, the smaller the size of the block is, and vice versa. The partitioning process is similar in different dimensions and is therefore explained herein by taking the division of two-dimensional cylindrical cells as an example. Assuming that (min x, max y) T and (min x, min y) T are the start and end points, respectively, of the linear index calculation, after the point cloud data is rasterized, the formula is passed The calculation can obtain that the horizontal distance D x and the vertical distances D y,Dx and D y between the starting point and the ending point are all positive integers. A zero value matrix ψ of (1+d y)×(1+Dx) and an empty set cell matrix Ω are established simultaneously. The horizontal distance d x and the vertical distance d y from any grid point to the beginning of the data index are calculated in the same way. Thus, in a computer memory system, elements in the zero value matrix ψ and the empty set cell matrix Ω can be considered to be represented using row numbers 1+d y and column numbers 1+d x (0<d x≤Dx,0<dy≤Dy). To facilitate data indexing, the row and column numbers can be converted to linear indexes denoted ζ by the formula ζ=d x·(1+Dy)+dy +1. Each grid point after the point cloud data is rasterized has a corresponding linear index value, the data points at the same position of the same matrix have the same linear index value, and when each data point spans across one matrix, the zeta of the data points is stepped, which is called as a step point and is defined by zeta j. When the point cloud data is segmented, the point cloud data can be arranged in ascending order according to the value of the linear index zeta, then whether the point is a step point or not is judged by calculating the difference delta zeta of adjacent data points, if delta zeta is smaller than 1, the point is not the step point, and otherwise, the point is the step point. And (3) carrying out the same processing on all the step points ζ j, namely storing the point cloud data of ζ j-1 between the step points and the last step point into a ζ j-1 unit of a cellular omega matrix, and simultaneously assigning 1 to the ζ j-1 element of the ψ matrix, thereby realizing the blocking of the point cloud data.
S22, screening out points with lower elevations in the columnar units by utilizing a height threshold h t, and analyzing the remaining data points by using a principal component analysis method to obtain the principal directions of the columnar units.
FIG. 3 is a schematic diagram showing the effect of the preliminary filtering method of the present invention. In this embodiment, the unit grid is divided into 0.5m by 0.5m, and in each cylindrical unit, the height range of the road point cloud data is extremely limited and approximately equal, so that the height threshold h t can be utilized to screen the point with the lower height in the cylindrical unit first. The remaining data points are then analyzed using principal component analysis methods to obtain their principal directions. Principal component analysis is one of the common statistical methods, namely, a group of variables possibly with correlation are converted into linear uncorrelated variables through matrix orthotopic transformation, namely, the principal components. Specifically, in the present embodiment, it is assumed that (x, y, z) T is set as any point coordinate in the columnar unit, and (λ, u, v) T is a converted variable thereof. For the point cloud with good plane characteristics, the main direction v is consistent with the plane normal vector formed by the point cloud, namely the point cloud is in a perpendicular relation, so that compared with non-characteristic points, the point cloud has small average absolute error in the upsilon direction. The average absolute error is simply the average of the absolute values of the deviations of all individual data from their overall arithmetic mean. For a set of m data, the average absolute error is as followsThe calculation is performed such that,Since the elevation of the ground point is at a lower position in all the point cloud data, the elevation threshold is set to 0.2m in this embodiment, that is, 0.5×0.5×0.2m per columnar unit size.
S23, sorting and detecting the road surface points by using a height histogram method, carrying out feature extraction on the obtained plane point cloud data, wherein feature variables comprise the number of point clouds in a unit area, standard deviation and median of adjacent term differences along the x and y directions for describing the distribution features of the point clouds on an x-y plane, obtaining an observation value array consistent with the number of cylindrical units, dividing the cylindrical units into three types of uniform plane points, non-uniform plane points and pseudo-plane points by using an unsupervised classification K-Means clustering algorithm, carrying out standardization processing on all variables to eliminate the difference of the cylindrical units, enabling intervals of the variables to fall in [0,1], eliminating the non-uniform plane points and the pseudo-plane points, merging the uniform plane points into the cylindrical units subjected to preliminary filtering processing, and obtaining the optimized plane point cloud data.
Fig. 4 is a schematic diagram of the structure of the height histogram method of the present invention. In step S23 of this embodiment, the processing procedure of the height orthogonal diagram sorting method includes the following steps:
S231, setting the height delta h of a single stripe in the height histogram to be 1-4 m, establishing a frequency histogram along the Z direction, and simultaneously establishing an empty cell array phi.
S232, sorting the frequency numbers according to the frequency number descending order, and setting F { F 1,F2...Fi...Fn|1≤i≤n,i,n∈N+ } as the height value corresponding to each stripe after the arrangement.
S233, a loop i=2 and k=1 is established, and the point cloud data corresponding to F i-1 is merged and stored to Φ { k }.
S234, if F i-Fi-1≤δh, merging and storing the point cloud data corresponding to F i and F i-1 to Φ { k }, otherwise, k=k+1, and storing the point cloud data corresponding to F i to Φ { k }.
S235, let i=i+1, if i is less than or equal to N h, return to step S234, otherwise end the loop.
Considering that the algorithm processes in each columnar unit are similar and mutually independent, the embodiment fully utilizes the parallel computing advantage of the multithreading processor and improves the automatic detection efficiency of the plane points in each columnar unit. In consideration of differences of data points in different cylindrical units, the embodiment also performs feature extraction on the plane point cloud data obtained by the algorithm, the extracted feature variables mainly comprise standard deviation and median of adjacent item differences along the x and y directions, the standard deviation and median are firstly arranged in ascending order during calculation, and the feature variables mainly describe distribution features of the point cloud on the x-y plane. Because the road points are distributed continuously and regularly, rather than unevenly. In addition, considering that the road points in the partial cylindrical units are not completely uniformly distributed, the point cloud density of the number of point clouds in a unit area is also taken as one of characteristic variables. The method for calculating the number of the point clouds in the unit area and the point cloud density adopts a method of dividing and calculating the point clouds and then taking the average value, and each cylindrical unit is divided into 100 subregions of 10 x 10 equally, wherein the point cloud density is equal to the ratio of the subregion area containing the points to the total point number.
After the feature vector is extracted, an observation value array with the same number as that of the columnar units is obtained, and then the columnar units can be divided into three types of uniform plane points, non-uniform plane points and pseudo plane points by using an unsupervised classification K-Means clustering algorithm. The pseudo-planar points are points which are misclassified as planar points due to small height intervals in the Z direction, and therefore, the pseudo-planar points are linear structures. Firstly, all variables are required to be standardized to eliminate the variability, so that the intervals of the variables are all within [0,1 ]. And then eliminating the uneven plane points and the pseudo plane points, and merging the even plane points into the cylindrical units after the preliminary filtering treatment to obtain the optimized plane point cloud data.
And S24, distinguishing the pavement part from other discrete plane points by using an ultra-voxel clustering method to obtain complete pavement point cloud data, obtaining three-dimensional point cloud data in a corresponding cell matrix by searching through index position information, and splicing the three-dimensional point cloud data to obtain a final pavement ultra-voxel clustering result. The clustering method based on the super voxels can effectively distinguish the pavement part from other discrete plane points and obtain complete pavement point cloud data.
3. Construction of digital elevation model
In this embodiment, the digital elevation model construction includes the sub-steps of:
a plane lattice is established on a plane by taking (minx, maxy) T and (maxx, miny) T as angular points and taking 0.2m as an interval;
And taking the identified road surface points as basic interpolation references, and acquiring the height information of the plane lattice points by using a neighboring point interpolation algorithm. When interpolation is carried out under the digital elevation model, adjacent lattice points can be positioned based on the plane positions of the observation points, and the height information of the positions of the observation points can be rapidly acquired, so that subsequent visibility analysis is convenient.
4. Converting a three-dimensional problem to a two-dimensional problem
Fig. 5 is a schematic structural diagram of the cylindrical perspective projection calculation of the present invention. Fig. 6 is a view blind zone structure diagram of the present invention. In step S4 of this embodiment, the processing procedure for converting the three-dimensional problem into the two-dimensional problem by converting the spatial three-dimensional coordinate point into the two-dimensional coordinate point through the cylindrical perspective projection model is as follows:
In a two-dimensional plane, the field of view concept is quantized into a sector consisting of innumerable rays emitted by the origin of the driver's line of sight. The distance between the line of sight initiation points is the sector radius D, the driver's viewing angle is θ, and both D and θ are set as variable parameters for a more comprehensive assessment. In this sector, detection of an obstacle is achieved when the line of sight is obstructed, θ being set to 120 °.
And the transformation from the space three-dimensional coordinate point to the two-dimensional coordinate point is realized through the cylindrical perspective projection model, and the three-dimensional problem is transformed into the two-dimensional problem. Two local coordinate systems are first established to obtain coordinate data of cylindrical surface points. The viewpoint origin is now set as the origin, the coordinates are set (X m,ym,zm)T, the vehicle advancing direction is set as the Y 'axis, the X' axis is parallel to the XOY plane, perpendicular to the Y axis, and the Z 'axis is perpendicular to the X' OY 'plane. On the basis of this three-dimensional space establishment, another two-dimensional coordinate system u-v can be constructed, the origin of which is set on the Y' axis and at a distance R from the line of sight origin, the v axis is parallel to the Z 'axis along the vector, the u axis is set as a clockwise arc, and is perpendicular to both the v axis and the Y', in both spatial coordinate systems, the coordinates of the two-dimensional point on the cylindrical surface can be calculated by:
Wherein (X, Y, Z) T represents coordinates of points around the driver in the geodetic space, (X m,ym,zm)T represents coordinates of the origin of the line of sight, (X ', Y', Z ') T represents coordinates converted into an X' Y 'Z' coordinate system, λ represents a scale factor between two reference coordinate systems, which takes a value of 1.0, θ represents a rotation matrix of the rigid conversion of the line of sight, θ xyz represents a rotation matrix around X, Y and the Z axis, α m represents a rotation angle around the Z axis, γ m represents a rotation angle around the X axis, R represents a radius of the cylindrical surface, which takes a value of 3.0, (u, v) T represents coordinates of the cylindrical surface, α m and γ m represent an azimuth angle and a vertical angle of the origin of the line of sight, respectively, and the direction of the Y 'axis coincides with the advancing direction of the vehicle, so that in this case, the direction of the Y' axis coincides with the advancing direction of the vehicle.
η=θ/θres+1
[ρ]={ρ12…ρj…ρη-1ηi=D(1≤j≤η),η=θ/θres+1}
[y′e,x′e]=Polar To Cartesian([θ,ρ])
[z′e]={z1,z2…zj…zη-1,zη|zi=D·tanθv(1≤j≤η),η=θ/θres+1}
Wherein, each parameter in the formula has the following meaning:
[ x 'e,y′e,z′e]T ] represents the three-dimensional coordinates of the line of sight end in the local three-dimensional coordinate system, θ represents the driver's viewing angle, θ res represents the angular spacing between the lines of sight, j represents the line of sight index, and η is the line of sight index when 1.ltoreq.j.ltoreq.η.
By the above calculation, two-dimensional coordinates of points representing the road environment and all line-of-sight end points can be obtained.
The next step is to obtain the coordinates of the projection point shown in fig. 5, namely, the projection point of the target point on the cylindrical surface, the area where the coordinates of the projection point are located can be regarded as a square area with a side length of W i, the center position of the projection point is located, and the coordinates are set asSetting collectionsStoring two-dimensional coordinates of all sight-line end points, setting the coordinates of the target point as Ob i(xi,yi,zi)T, and searching by using a KD tree algorithm and using a Chebyshev sequence with the parameter of W i/2Points within the neighborhood. The set ψ i is set to store the two-dimensional coordinates of all the points in the square area where the projection points are located, and all the two-dimensional points of the set ψ i have three-dimensional coordinates corresponding to the two-dimensional coordinates. Let κ i denote the three-dimensional spatial distance between the line of sight origin and the target point, any point whose three-dimensional spatial distance from the line of sight origin exceeds κ i will be excluded from the set ψ i, which can greatly improve the efficiency of the search along the line of sight.
In the embodiment, the visual field modeling process mainly comprises the steps of constructing a coordinate system by taking a sight line starting point of a driver as an origin, converting a three-dimensional coordinate system into a two-dimensional coordinate system and converting the two-dimensional rectangular coordinate system into a polar coordinate system, determining the position of an obstacle blocking the sight line of the driver through two times of coordinate conversion, and when the coordinate system is converted based on a specific matrix equation, mapping coordinates in the polar coordinate system into a real three-dimensional space coordinate system, so that the three-dimensional visual field model in the coordinate system can be constructed.
Fig. 7 is a schematic diagram showing the correspondence between the visual field and the line-of-sight curve according to the present invention. According to the visual field modeling and the corresponding schematic diagram of the visual range curve, which are obtained by the embodiment, the position of the obstacle can be rapidly positioned and the visual range size can be quantitatively calculated by the method, so that the visibility evaluation of the urban road intersection is obtained.
The above is only a preferred embodiment of the present invention, and the protection scope of the present invention is not limited to the above examples, and all technical solutions belonging to the concept of the present invention belong to the protection scope of the present invention. It should be noted that modifications and adaptations to the invention without departing from the principles thereof are intended to be within the scope of the invention as set forth in the following claims.

Claims (6)

1.一种基于点云数据的城市道路交叉口可视化评价方法,其特征在于,所述道路交叉口三维视距计算方法包括以下步骤:1. A method for visual evaluation of urban road intersections based on point cloud data, characterized in that the method for calculating the three-dimensional sight distance of the road intersection comprises the following steps: S1,采集道路交叉口的点云数据;S1, collects point cloud data of road intersections; S2,对点云数据进行路面识别分割处理,将路面部分与其他离散的平面点区分开,以得到完整的路面点云数据;S2, performing road surface recognition and segmentation processing on the point cloud data, distinguishing the road surface part from other discrete plane points to obtain complete road surface point cloud data; S3,建立平面格网点阵,以识别到的路面点作为基准插值参照,利用邻近点插值算法获取到平面格网点的高度信息,构建数字高程模型;S3, establishing a plane grid dot matrix, using the identified road surface points as the base interpolation reference, using the neighboring point interpolation algorithm to obtain the height information of the plane grid points, and constructing a digital elevation model; S4,将视域概念量化为由驾驶员视线原点发射出的无数条射线组成的扇形,扇形的角度与驾驶员的视角范围相关;通过圆柱形透视投影模型,将空间三维坐标点转化为二维坐标点,将三维问题转换为二维问题;S4, quantifies the concept of visual field into a fan-shaped shape composed of countless rays emitted from the origin of the driver's line of sight. The angle of the fan-shaped shape is related to the driver's visual range. Through the cylindrical perspective projection model, the three-dimensional coordinate points in space are converted into two-dimensional coordinate points, and the three-dimensional problem is converted into a two-dimensional problem. 以驾驶员视线起点为原点构建坐标系,将三维坐标系转换成二维坐标系,再将二维直角坐标系转换为极坐标系,确定阻挡驾驶员视线的障碍物位置;A coordinate system is constructed with the starting point of the driver's line of sight as the origin, the three-dimensional coordinate system is converted into a two-dimensional coordinate system, and then the two-dimensional rectangular coordinate system is converted into a polar coordinate system to determine the position of the obstacle blocking the driver's line of sight; S5,分析可能影响城市道路交叉口可视域的变量因素;S5, analyze the variable factors that may affect the visibility of urban road intersections; 步骤S4中,所述通过圆柱形透视投影模型,将空间三维坐标点转化为二维坐标点,将三维问题转换为二维问题的处理过程如下:In step S4, the three-dimensional coordinate points in space are converted into two-dimensional coordinate points through the cylindrical perspective projection model, and the processing process of converting the three-dimensional problem into a two-dimensional problem is as follows: 将视点起点设置为原点,坐标设为(xm,ym,zm)T,车辆前进方向设为Y′轴,X′轴与XOY平面平行,与Y轴垂直,而Z′轴与X′OY′平面垂直设置,构建三维空间坐标系;The viewpoint starting point is set as the origin, the coordinates are set as (x m , y m , z m ) T , the vehicle forward direction is set as the Y′ axis, the X′ axis is parallel to the XOY plane and perpendicular to the Y axis, and the Z′ axis is set perpendicular to the X′OY′ plane to construct a three-dimensional space coordinate system; 构建相应的二维坐标系统u-υ,该二维坐标系统原点设置在Y′轴上且距离视线原点距离为R,v轴沿向量与Z′轴平行设置,u轴设置为顺时针弧段,与v轴和Y′均为垂直关系;Construct a corresponding two-dimensional coordinate system u-υ, the origin of which is set on the Y′ axis and at a distance R from the origin of the line of sight, the v axis is set parallel to the Z′ axis along the vector, and the u axis is set as a clockwise arc segment, which is perpendicular to both the v axis and the Y′ axis; 在两个空间坐标系中,圆柱面上二维点的坐标计算公式如下:In the two spatial coordinate systems, the coordinate calculation formula of the two-dimensional point on the cylindrical surface is as follows: 其中:(x,y,z)T代表大地测量空间中驾驶员周围点的坐标;(xm,ym,zm)T代表视线原点坐标;(x′,y′,z′)T代表转换到X′Y′Z′坐标系中的坐标;λ代表两个参考坐标系之间的比例因子,取值为1.0;θ代表视线刚性转换的旋转矩阵;θx,θy,θz代表围绕X、Y和Z轴的旋转矩阵;α代表绕Z轴的旋转角度;γm代表绕X轴的旋转角度;R代表圆柱形表面的半径,取值为3.0;(u,v)T代表圆柱形表面的坐标;αm和γm分别表征视线原点的方位角和竖向角度,Y′轴的方向与车辆前进方向一致,因此在这种情况下,Y′轴的方向与车辆的前进方向一致;Wherein: (x, y, z) T represents the coordinates of points around the driver in geodetic space; (x m , y m , z m ) T represents the coordinates of the origin of the line of sight; (x′, y′, z′) T represents the coordinates converted to the X′Y′Z′ coordinate system; λ represents the scale factor between the two reference coordinate systems, which is 1.0; θ represents the rotation matrix of the rigid transformation of the line of sight; θ x , θ y , θ z represent the rotation matrices around the X, Y and Z axes; α represents the rotation angle around the Z axis; γ m represents the rotation angle around the X axis; R represents the radius of the cylindrical surface, which is 3.0; (u, v) T represents the coordinates of the cylindrical surface; α m and γ m represent the azimuth and vertical angle of the origin of the line of sight, respectively. The direction of the Y′ axis is consistent with the direction of travel of the vehicle, so in this case, the direction of the Y′ axis is consistent with the direction of travel of the vehicle; 在局部构建的局部三维空间中,所有视线终点的坐标计算公式如下:In the locally constructed local three-dimensional space, the coordinate calculation formula of all sight end points is as follows: η=θ/θres+1η=θ/ θres +1 [ρ]={ρ1,ρ2...ρj...ρη-1,ρηi=D,η=θ/θres+1}[ρ]={ρ 1 , ρ 2 ...ρ j ...ρ η-1 , ρ ηi =D, η=θ/θ res +1} [y′e,x′e]=Polar To Cartesian([θ,ρ])[y′ e , x′ e ]=Polar To Cartesian ([θ, ρ]) [z′e]={z1,z2...zj...zη-1,zη|zi=D·tanθv,η=θ/θres+1}[z′ e ]={z 1 , z 2 ...z j ...z η-1 , z η |z i =D·tanθ v , η=θ/θ res +1} 其中:[x′e,y′e,z′e]T代表局部三维坐标系中视线终点的三维坐标;θ代表驾驶员视角;θres代表视线间的角间距;ρ代表坐标转换后极坐标中的距离参数;D代表扇形半径即驾驶员视距;j代表视线指数,当1≤j≤η,则η为视线指数;Where: [x′ e , y′ e , z′ e ] T represents the three-dimensional coordinates of the end point of the line of sight in the local three-dimensional coordinate system; θ represents the driver's viewing angle; θ res represents the angular spacing between lines of sight; ρ represents the distance parameter in polar coordinates after coordinate conversion; D represents the sector radius, i.e., the driver's sight distance; j represents the line of sight index, when 1≤j≤η, then η is the line of sight index; 将投影点坐标所在区域视为边长为Wi的方形区域,投影点位于区域的中心位置,坐标设为设置集合存储所有视线终点的二维坐标,设目标点坐标为Obi(xi,yi,zi)T,通过KD树算法,以参数为Wi/2的Chebyshev序列搜索邻域内的点;The area where the projection point coordinates are located is regarded as a square area with a side length of Wi . The projection point is located at the center of the area, and the coordinates are set to Setting up the collection Store the 2D coordinates of all sight endpoints, set the target point coordinates as Ob i (x i , y i , z i ) T , and use the KD tree algorithm to search for the Chebyshev sequence with parameter Wi /2 Points in the neighborhood; 设置集合Ψi存储投影点所在方形区域内所有点的二维坐标,集合Ψi所有的二维点均有三维坐标与之相对应;设κi表示视线原点与目标点之间的三维空间距离,将任何与视线原点的三维空间距离超过κi的点排除在集合Ψi之外。Set Ψ i to store the two-dimensional coordinates of all points in the square area where the projection point is located. All two-dimensional points in set Ψ i have corresponding three-dimensional coordinates. Let κ i represent the three-dimensional space distance between the origin of the line of sight and the target point, and exclude any point whose three-dimensional space distance from the origin of the line of sight exceeds κ i from the set Ψ i . 2.根据权利要求1所述的一种基于点云数据的城市道路交叉口可视化评价方法,其特征在于,步骤S1中,所述采集道路交叉口的点云数据的过程包括以下步骤:2. The method for visual evaluation of urban road intersections based on point cloud data according to claim 1, characterized in that in step S1, the process of collecting point cloud data of the road intersection comprises the following steps: S11,将小型激光雷达固定在旋转云台上,放置于城市道路交叉口的中心,保持激光雷达搭载支架固定不动;通过旋转云台多次采集数据获取交叉口全貌点云数据的子文件,邻近的2个点云文件重合度为1/3;S11, fix the small laser radar on the rotating gimbal, place it at the center of the urban road intersection, and keep the laser radar mounting bracket fixed; collect data multiple times through the rotating gimbal to obtain the sub-file of the overall point cloud data of the intersection, and the overlap degree of two adjacent point cloud files is 1/3; S12,将获得的点云文件通过对应特征点进行数据拼接,获得完整的城市道路交叉口点云数据。S12, performing data splicing on the obtained point cloud files through corresponding feature points to obtain complete point cloud data of urban road intersections. 3.根据权利要求1所述的一种基于点云数据的城市道路交叉口可视化评价方法,其特征在于,步骤S2中,所述对点云数据进行路面识别分割处理的过程包括以下步骤:3. The method for visual evaluation of urban road intersections based on point cloud data according to claim 1, characterized in that in step S2, the process of performing road surface recognition and segmentation processing on the point cloud data comprises the following steps: S21:对原始的三维点云数据进行栅格化处理,点云数据栅格化后的每一个栅格点都有相对应的线性索引值,同一个矩阵同一位置的数据点会拥有相同的线性索引值,数据点每跨越一个矩阵时,其数据点的线性索引值ζ发生阶跃;S21: Rasterize the original 3D point cloud data. Each grid point of the rasterized point cloud data has a corresponding linear index value. The data points at the same position in the same matrix will have the same linear index value. When a data point crosses a matrix, the linear index value ζ of the data point will jump. S22:利用高度阈值ht筛选出柱形单元内高程较低的点,对剩余数据点使用主成分分析方法进行分析以获得该柱形单元的主方向;S22: using the height threshold ht to filter out points with lower elevations in the columnar unit, and using the principal component analysis method to analyze the remaining data points to obtain the main direction of the columnar unit; S23:使用高度直方图法排序探测路面点,对得到的平面点云数据进行特征提取;特征变量包括单位面积内的点云数量,以及用于描述点云在x-y平面上的分布特征的沿着x和y方向相邻项差值的标准差和中位数,得到与柱形单元数量一致的观测值数组;利用非监督分类K-Means聚类算法将柱形单元分为均匀平面点、不均匀平面点和伪平面点三类;对所有变量进行标准化处理以消除其差异性,使变量的区间均落于[0,1]内;剔除不均匀平面点与伪平面点,将均匀平面点并入初步滤波处理后的柱形单元,得到优化后的平面点云数据;S23: Use the height histogram method to sort the detected road surface points, and extract features from the obtained plane point cloud data; the feature variables include the number of point clouds per unit area, and the standard deviation and median of the difference between adjacent items along the x and y directions used to describe the distribution characteristics of the point cloud on the x-y plane, and obtain an observation value array consistent with the number of cylindrical units; use the unsupervised classification K-Means clustering algorithm to divide the cylindrical units into three categories: uniform plane points, non-uniform plane points, and pseudo-plane points; standardize all variables to eliminate their differences, so that the intervals of the variables all fall within [0, 1]; eliminate non-uniform plane points and pseudo-plane points, and merge the uniform plane points into the cylindrical units after preliminary filtering to obtain optimized plane point cloud data; S24:使用超体素聚类法将路面部分与其他离散的平面点区分开,得到完整的路面点云数据。S24: Use the supervoxel clustering method to distinguish the road surface part from other discrete plane points to obtain the complete road surface point cloud data. 4.根据权利要求3所述的一种基于点云数据的城市道路交叉口可视化评价方法,其特征在于,步骤S21中,对原始的三维点云数据进行栅格化处理的过程包括以下步骤:4. The method for visual evaluation of urban road intersections based on point cloud data according to claim 3 is characterized in that, in step S21, the process of rasterizing the original three-dimensional point cloud data comprises the following steps: S211,假设原始的三维点云数据空间坐标为(X,Y,Z)T,以εx,εy,εz分别代表沿X轴、Y轴与Z轴的栅格间隔,创立临时数组(x,y,z)TS211, assuming that the original three-dimensional point cloud data space coordinates are (X, Y, Z) T , with ε x , ε y , ε z representing the grid intervals along the X axis, Y axis and Z axis respectively, create a temporary array (x, y, z) T : S212,假设(minx,maxy)T和(maxx,miny)T分别为线性索引计算的起点和终点,在点云数据栅格化之后,通过公式计算得出起终点间的水平距离Dx与垂直距离Dy,其中Dx与Dy均为正整数,计算公式如下:S212, assuming that (minx, maxy) T and (maxx, miny) T are the starting point and the end point of the linear index calculation, respectively. After the point cloud data is rasterized, the horizontal distance D x and the vertical distance D y between the starting point and the end point are calculated by the formula, where D x and D y are both positive integers, and the calculation formula is as follows: S213,建立大小为(1+Dy)×(1+Dx)的零值矩阵Ψ与空集元胞矩阵Ω;计算得到任一栅格点到数据索引起点的水平距离Dx和垂直距离Dy;使用行号1+dy和列号1+dx表示零值矩阵Ψ与空集元胞矩阵Ω中的元素;通过公式ζ=dx·(1+Dy)+dy+1将行号与列号转化为以ζ表示的线性索引;S213, establish a zero value matrix Ψ and an empty set cell matrix Ω of size (1+D y )×(1+D x ); calculate the horizontal distance D x and the vertical distance D y from any grid point to the data index starting point; use row number 1+d y and column number 1+d x to represent the elements in the zero value matrix Ψ and the empty set cell matrix Ω; convert the row number and column number into a linear index represented by ζ through the formula ζ=d x ·(1+D y )+d y +1; 在点云数据分块时,按照线性索引ζ的值对点云数据进行升序排列,通过计算相邻数据点的差值Δζ判断该点是否为阶跃点,若Δζ<1则该点不是阶跃点,否则是阶跃点;将阶跃点ζj与上一阶跃点ζj-1之间的点云数据储存至元胞Ω矩阵第ζj-1个单元中,同时给Ψ矩阵第ζj-1个元素赋值为1。When the point cloud data is divided into blocks, the point cloud data is arranged in ascending order according to the value of the linear index ζ. The difference Δζ between adjacent data points is calculated to determine whether the point is a step point. If Δζ<1, the point is not a step point, otherwise it is a step point. The point cloud data between the step point ζ j and the previous step point ζ j-1 is stored in the ζ j-1th unit of the cellular matrix Ω, and the ζ j-1th element of the Ψ matrix is assigned a value of 1. 5.根据权利要求2所述的一种基于点云数据的城市道路交叉口可视化评价方法,其特征在于,步骤S23中,所述高度直角图排序法的处理过程包括以下步骤:5. The method for visual evaluation of urban road intersections based on point cloud data according to claim 2, characterized in that in step S23, the processing process of the height rectangular diagram sorting method includes the following steps: S231,高度直方图中单根条纹的高度δh设定为1m~4m,建立沿Z方向的频数直方图,同时建立空的元胞数组Φ;S231, the height δ h of a single stripe in the height histogram is set to 1m-4m, a frequency histogram along the Z direction is established, and an empty cell array Φ is established at the same time; S232,依频数降序对频数进行排序,设定F{F1,F2...Fi...Fn|1≤i≤n,i,n∈N+}为排列后的各条纹对应的高度值;S232, sorting the frequencies in descending order, setting F{F 1 , F 2 ..F i ..F n |1≤i≤n, i, n∈N + } as the height value corresponding to each stripe after arrangement; S233,建立循环i=2,k=1,将Fi-1对应的点云数据进行合并存储至Φ{k};S233, establish a loop i=2, k=1, merge and store the point cloud data corresponding to F i-1 into Φ{k}; S234,如Fi-Fi-1≤δh,则将Fi与Fi-1对应的点云数据进行合并存储至Φ{k},反之则k=k+1,将Fi对应的点云数据储存至Φ{k};S234, if F i - F i-1 ≤ δ h , the point cloud data corresponding to F i and F i-1 are merged and stored in Φ{k}; otherwise, k=k+1, and the point cloud data corresponding to F i is stored in Φ{k}; S235,令i=i+1,如i≤Nh,则返回步骤S234,反之则结束循环。S235, let i=i+1. If i≤N h , return to step S234. Otherwise, end the loop. 6.根据权利要求3所述的一种基于点云数据的城市道路交叉口可视化评价方法,其特征在于,步骤S24中,所述超体素聚类法的处理过程如下:6. The method for visual evaluation of urban road intersections based on point cloud data according to claim 3, characterized in that in step S24, the processing process of the supervoxel clustering method is as follows: 利用索引位置信息,检索得到对应的元胞矩阵中的三维点云数据,将其拼合得到最终的路面超体素聚类结果中得到完整的路面点云数据,将路面部分与其他离散的平面点区分开。Using the index position information, the corresponding three-dimensional point cloud data in the cell matrix is retrieved, and the three-dimensional point cloud data is spliced to obtain the complete road surface point cloud data in the final road surface supervoxel clustering result, which distinguishes the road surface part from other discrete plane points.
CN202111340586.2A 2021-11-12 2021-11-12 A visual evaluation method for urban road intersections based on point cloud data Active CN114119866B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111340586.2A CN114119866B (en) 2021-11-12 2021-11-12 A visual evaluation method for urban road intersections based on point cloud data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111340586.2A CN114119866B (en) 2021-11-12 2021-11-12 A visual evaluation method for urban road intersections based on point cloud data

Publications (2)

Publication Number Publication Date
CN114119866A CN114119866A (en) 2022-03-01
CN114119866B true CN114119866B (en) 2025-01-03

Family

ID=80379289

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111340586.2A Active CN114119866B (en) 2021-11-12 2021-11-12 A visual evaluation method for urban road intersections based on point cloud data

Country Status (1)

Country Link
CN (1) CN114119866B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114677424B (en) * 2022-05-26 2022-09-06 浙江天新智能研究院有限公司 Point cloud data processing method for unattended screw ship unloader
CN115153483A (en) * 2022-07-01 2022-10-11 山东第一医科大学第一附属医院(山东省千佛山医院) A Coronary Blood Supply Evaluation System Based on Parabolic Stereocoordinate System
CN115712298B (en) * 2022-10-25 2023-05-09 太原理工大学 Autonomous navigation method for robots driving in passageways based on single-line lidar
CN115690773B (en) * 2022-12-26 2023-04-07 武汉天际航信息科技股份有限公司 DEM partitioning and rebuilding method, computing device and storage medium
CN117611759B (en) * 2023-11-30 2024-07-09 博雅达勘测规划设计集团有限公司 Three-dimensional model-based scoring map generation method, device, terminal and storage medium
CN117934749B (en) * 2024-03-25 2024-05-24 成都陆拓信息技术有限公司 Three-dimensional visualization method for vegetation carbon fixation value

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105184852B (en) * 2015-08-04 2018-01-30 百度在线网络技术(北京)有限公司 A kind of urban road recognition methods and device based on laser point cloud
WO2021016891A1 (en) * 2019-07-30 2021-02-04 深圳市大疆创新科技有限公司 Method and apparatus for processing point cloud
CN113569915B (en) * 2021-06-30 2024-04-02 广西大学 Multi-strategy rail transit obstacle recognition method based on laser radar

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于点云数据的城市道路交叉口可视性分析;张家钰;中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑;20230315;全文 *

Also Published As

Publication number Publication date
CN114119866A (en) 2022-03-01

Similar Documents

Publication Publication Date Title
CN114119866B (en) A visual evaluation method for urban road intersections based on point cloud data
CN111985322B (en) Road environment element sensing method based on laser radar
CN112801022B (en) Method for rapidly detecting and updating road boundary of unmanned mining card operation area
CN110781827B (en) A road edge detection system and method based on lidar and fan-shaped space segmentation
US20230186647A1 (en) Feature extraction from mobile lidar and imagery data
CN110570428B (en) Method and system for dividing building roof sheet from large-scale image dense matching point cloud
EP4120123A1 (en) Scan line-based road point cloud extraction method
Yu et al. Semiautomated extraction of street light poles from mobile LiDAR point-clouds
CN112184736B (en) Multi-plane extraction method based on European clustering
Sohn et al. Using a binary space partitioning tree for reconstructing polyhedral building models from airborne lidar data
CN109633674A (en) Three-dimensional Track automatic planning is maked an inspection tour in transmission of electricity based on laser point cloud data
JP6621445B2 (en) Feature extraction device, object detection device, method, and program
CN112232248B (en) Method and device for extracting plane features of multi-line LiDAR point cloud data
CN111325138B (en) Road boundary real-time detection method based on point cloud local concave-convex characteristics
CN113255677B (en) Method, equipment and medium for rapidly extracting rock mass structural plane and occurrence information
CN111861946B (en) Adaptive multi-scale vehicle-mounted laser radar dense point cloud data filtering method
CN114020015A (en) UAV path planning system and method based on two-way search of obstacle map
CN114063107A (en) A method for extracting ground point cloud based on laser beam
Wen et al. Research on 3D point cloud de-distortion algorithm and its application on Euclidean clustering
Yadav et al. Identification of trees and their trunks from mobile laser scanning data of roadway scenes
Schlichting et al. Vehicle localization by lidar point correlation improved by change detection
CN116258857A (en) Outdoor tree-oriented laser point cloud segmentation and extraction method
WO2011085435A1 (en) Classification process for an extracted object or terrain feature
CN115294287A (en) A Laser SLAM Mapping Method for Greenhouse Inspection Robot
CN117593707A (en) Vehicle identification method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant