CN118898838A - Method, device, medium and vehicle for determining three-dimensional shape information of road obstacles - Google Patents
Method, device, medium and vehicle for determining three-dimensional shape information of road obstacles Download PDFInfo
- Publication number
- CN118898838A CN118898838A CN202411379463.3A CN202411379463A CN118898838A CN 118898838 A CN118898838 A CN 118898838A CN 202411379463 A CN202411379463 A CN 202411379463A CN 118898838 A CN118898838 A CN 118898838A
- Authority
- CN
- China
- Prior art keywords
- area
- grid
- obstacle
- information
- vehicle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 105
- 238000001514 detection method Methods 0.000 claims abstract description 51
- 238000012545 processing Methods 0.000 claims abstract description 20
- 235000004522 Pentaglottis sempervirens Nutrition 0.000 claims description 82
- 240000004050 Pentaglottis sempervirens Species 0.000 claims description 81
- 238000004364 calculation method Methods 0.000 claims description 37
- 238000004590 computer program Methods 0.000 claims description 31
- 230000011218 segmentation Effects 0.000 claims description 27
- 230000006870 function Effects 0.000 claims description 19
- 230000033001 locomotion Effects 0.000 claims description 19
- 230000008447 perception Effects 0.000 claims description 16
- 230000009466 transformation Effects 0.000 claims description 15
- 238000012549 training Methods 0.000 claims description 6
- 230000008859 change Effects 0.000 claims description 5
- 230000004927 fusion Effects 0.000 claims description 4
- 238000005516 engineering process Methods 0.000 abstract description 17
- 230000008569 process Effects 0.000 abstract description 12
- 230000000007 visual effect Effects 0.000 abstract description 3
- 238000004422 calculation algorithm Methods 0.000 description 15
- 238000010586 diagram Methods 0.000 description 14
- 238000005259 measurement Methods 0.000 description 7
- 238000013459 approach Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 5
- 238000012795 verification Methods 0.000 description 5
- 238000013528 artificial neural network Methods 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 4
- 230000000877 morphologic effect Effects 0.000 description 4
- 101100233916 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) KAR5 gene Proteins 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 3
- 238000000605 extraction Methods 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 239000000725 suspension Substances 0.000 description 2
- 238000011179 visual inspection Methods 0.000 description 2
- 101001121408 Homo sapiens L-amino-acid oxidase Proteins 0.000 description 1
- 101000827703 Homo sapiens Polyphosphoinositide phosphatase Proteins 0.000 description 1
- 102100026388 L-amino-acid oxidase Human genes 0.000 description 1
- 102100023591 Polyphosphoinositide phosphatase Human genes 0.000 description 1
- 206010039203 Road traffic accident Diseases 0.000 description 1
- 101100012902 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) FIG2 gene Proteins 0.000 description 1
- 230000008094 contradictory effect Effects 0.000 description 1
- 238000000280 densification Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000002068 genetic effect Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 230000008439 repair process Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 239000011800 void material Substances 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/757—Matching configurations of points or features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
技术领域Technical Field
本发明涉及自动驾驶技术领域,特别是涉及一种路面障碍物三维形貌信息确定方法、电子设备、非易失性存储介质、计算机程序产品及无人驾驶车辆。The present invention relates to the field of autonomous driving technology, and in particular to a method for determining three-dimensional shape information of a road obstacle, an electronic device, a non-volatile storage medium, a computer program product, and an unmanned vehicle.
背景技术Background Art
随着人工智能技术和物联网技术的快速发展,自动驾驶技术也得到相应的发展。道路平整度不仅影响自动驾驶车辆的正常运行和乘车人的舒服度,还影响道路及其基础设施的寿命。With the rapid development of artificial intelligence and Internet of Things technologies, autonomous driving technology has also developed accordingly. The smoothness of the road not only affects the normal operation of autonomous vehicles and the comfort of passengers, but also affects the life of the road and its infrastructure.
相关技术在监测道路是否存在坑洼和低矮凸起障碍物时,无法精准识别坑洼区域和凸起障碍物区域,进而无法精准确定道路坑洼和道路低矮凸起障碍物所在位置。When monitoring whether there are potholes and low raised obstacles on the road, the relevant technology cannot accurately identify the pothole areas and raised obstacle areas, and thus cannot accurately determine the locations of potholes and low raised obstacles on the road.
鉴于此,提升道路坑洼区域和道路低矮凸起障碍物区域的三维形貌信息确定精度,是本领域技术人员需要解决的技术问题。In view of this, improving the accuracy of determining three-dimensional topographic information of road potholes and low raised obstacle areas on the road is a technical problem that needs to be solved by technical personnel in this field.
需要说明的是,在上述背景技术部分公开的信息仅用于加强对本申请的背景的理解,因此可以包括不构成对本领域普通技术人员已知的现有技术的信息。It should be noted that the information disclosed in the above background technology section is only used to enhance the understanding of the background of the present application, and therefore may include information that does not constitute the prior art known to ordinary technicians in the field.
发明内容Summary of the invention
本发明提供了一种路面障碍物三维形貌信息确定方法、电子设备、非易失性存储介质、计算机程序产品及无人驾驶车辆,能够精准确定道路坑洼区域和道路低矮凸起障碍物区域的三维形貌信息。The present invention provides a method for determining three-dimensional shape information of road obstacles, an electronic device, a non-volatile storage medium, a computer program product and an unmanned vehicle, which can accurately determine the three-dimensional shape information of road pothole areas and road low raised obstacle areas.
为解决上述技术问题,本发明提供以下技术方案:In order to solve the above technical problems, the present invention provides the following technical solutions:
本发明一方面提供了一种路面障碍物三维形貌信息确定方法,包括:In one aspect, the present invention provides a method for determining three-dimensional shape information of a road obstacle, comprising:
获取栅格化处理的待检测路面区域的点云数据和图像数据在目标视角下的特征;其中,栅格化的待检测路面区域包括多个栅格,各栅格在垂直于路面水平方向上设置空间体素;Acquire the features of the point cloud data and image data of the road surface area to be detected in the target perspective after rasterization; wherein the road surface area to be detected in the rasterization includes a plurality of grids, and each grid is provided with a spatial voxel in a horizontal direction perpendicular to the road surface;
将点云特征和图像特征进行融合,并通过对融合特征进行目标检测处理,得到待检测路面区域内的障碍物的初始三维形貌信息;The point cloud features and image features are fused, and the initial three-dimensional shape information of the obstacles in the road area to be detected is obtained by performing target detection processing on the fused features;
根据车辆前行过程中不断获取的各帧点云数据对应的栅格新高程信息,更新当前时刻的空间体素权重,并基于更新后的空间体素权重更新障碍物区域内栅格的高程信息和几何形态数据,以对所述初始三维形貌信息进行不断更新。According to the new elevation information of the grid corresponding to each frame of point cloud data continuously acquired during the vehicle's forward movement, the spatial voxel weight at the current moment is updated, and the elevation information and geometric morphology data of the grid in the obstacle area are updated based on the updated spatial voxel weight, so as to continuously update the initial three-dimensional shape information.
在第一种示例性的实施方式中,所述获取栅格化处理的待检测路面区域的点云数据和图像数据在目标视角下的特征,包括:In a first exemplary embodiment, the step of acquiring the features of the rasterized point cloud data and image data of the road surface area to be detected under the target viewing angle includes:
当检测到车辆行驶前方区域存在障碍物,则在所述车辆前行过程监视所述障碍物位置;When an obstacle is detected in the area ahead of the vehicle, the position of the obstacle is monitored during the forward movement of the vehicle;
当所述障碍物进入待检测路面区域时,将位于当前车辆前方的待检测路面区域进行栅格化处理,生成包括多个栅格、且各栅格在垂直于路面水平方向上具有空间体素的鸟瞰视图栅格区域;When the obstacle enters the road surface area to be detected, the road surface area to be detected in front of the current vehicle is rasterized to generate a bird's-eye view grid area including a plurality of grids, each grid having spatial voxels in a horizontal direction perpendicular to the road surface;
获取所述鸟瞰视图栅格区域的点云数据和图像数据,并将所述点云数据和图像数据转换至鸟瞰视角,得到相应视角的点云特征和图像特征。The point cloud data and image data of the bird's-eye view grid area are acquired, and the point cloud data and image data are converted to a bird's-eye view to obtain point cloud features and image features of the corresponding view.
其中,所述待检测路面区域在沿车辆行驶方向上与车辆的最大距离小于等于所述车辆行驶前方区域与车辆间的最小距离。The maximum distance between the to-be-detected road surface area and the vehicle along the vehicle's driving direction is less than or equal to the minimum distance between the area in front of the vehicle and the vehicle.
在第二种示例性的实施方式中,所述通过对融合特征进行目标检测处理,得到待检测路面区域内的障碍物的初始三维形貌信息,包括:In a second exemplary embodiment, the initial three-dimensional shape information of obstacles in the road area to be detected is obtained by performing target detection processing on the fused features, including:
预先构建并训练三维目标检测网络模型;所述三维目标检测网络模型包括检测任务头、分类任务头和实例分割任务头,所述检测任务头用于识别障碍物,并输出障碍物位置信息;所述分类任务头用于识别障碍物类别信息;所述实例分割任务头用于对障碍物在目标平面上进行分割,得到实例分割结果;Pre-constructing and training a three-dimensional target detection network model; the three-dimensional target detection network model includes a detection task head, a classification task head and an instance segmentation task head, the detection task head is used to identify obstacles and output obstacle position information; the classification task head is used to identify obstacle category information; the instance segmentation task head is used to segment obstacles on the target plane to obtain instance segmentation results;
将所述融合特征输入至所述三维目标检测网络模型,得到障碍物的位置信息、类别信息和实例分割结果;Inputting the fused features into the three-dimensional target detection network model to obtain the location information, category information and instance segmentation results of the obstacles;
基于所述实例分割结果和障碍物区域内点云数据,通过三维曲面拟合结果确定所述障碍物的几何形态描述数据;Based on the instance segmentation result and the point cloud data in the obstacle area, determining the geometric description data of the obstacle through the three-dimensional surface fitting result;
将所述障碍物的位置信息、类别信息和几何形态描述数据,作为所述障碍物的初始三维形貌信息。The position information, category information and geometric description data of the obstacle are used as the initial three-dimensional shape information of the obstacle.
在第三种示例性的实施方式中,所述基于更新后的空间体素权重更新障碍物区域内栅格的高程信息和几何形态数据之后,还包括:In a third exemplary embodiment, after updating the elevation information and geometric data of the grid in the obstacle area based on the updated spatial voxel weight, the method further includes:
基于车辆坐标系与局部世界坐标系之间的变换关系,将待检测路面区域的障碍物的三维形貌信息转换至局部世界坐标系;Based on the transformation relationship between the vehicle coordinate system and the local world coordinate system, the three-dimensional shape information of the obstacle in the road area to be detected is converted into the local world coordinate system;
其中,所述局部世界坐标系为车辆行驶路面所在的坐标系。The local world coordinate system is the coordinate system of the road surface on which the vehicle is traveling.
在第四种示例性的实施方式中,所述基于车辆坐标系与局部世界坐标系之间的变换关系之前,还包括:In a fourth exemplary embodiment, before the transformation relationship between the vehicle coordinate system and the local world coordinate system is described, the method further includes:
获取当前时刻的点云数据和局部地图点云;Get the point cloud data and local map point cloud at the current moment;
通过对所述点云数据和所述局部地图点云进行点云匹配,确定车辆坐标系与局部世界坐标系之间的变换关系;Determine the transformation relationship between the vehicle coordinate system and the local world coordinate system by performing point cloud matching on the point cloud data and the local map point cloud;
其中,所述局部地图点云为沿着时间轴,对不同时刻的点云数据进行拼接后生成的数据。The local map point cloud is data generated by splicing point cloud data at different times along the time axis.
在第五种示例性的实施方式中,根据车辆前行过程中不断获取的各帧点云数据对应的栅格新高程信息,更新当前时刻的空间体素权重,包括:In a fifth exemplary implementation, the spatial voxel weight at the current moment is updated according to the new grid elevation information corresponding to each frame of point cloud data continuously acquired during the vehicle's forward movement, including:
获取栅格化的待检测路面区域的各栅格的空间体素及对应的权重初始值;Obtaining the spatial voxels of each grid of the rasterized road surface area to be detected and the corresponding initial weight values;
获取当前帧点云数据对应的栅格的高程信息,以作为新高程信息,并利用所述新高程信息更新相应位置的空间体素的权重值。The elevation information of the grid corresponding to the point cloud data of the current frame is obtained as new elevation information, and the weight value of the spatial voxel at the corresponding position is updated using the new elevation information.
在第六种示例性的实施方式中,所述根据车辆前行过程中不断获取的各帧点云数据对应的栅格新高程信息,更新当前时刻的空间体素权重,包括:In a sixth exemplary implementation, the updating of the spatial voxel weight at the current moment according to the new grid elevation information corresponding to each frame of point cloud data continuously acquired during the vehicle's forward movement includes:
获取栅格化的待检测路面区域中各栅格区域类型;Obtain the type of each grid area in the rasterized road surface area to be detected;
当当前栅格属于上一帧识别的第一类障碍物区域,若当前栅格不属于雷达点云投影区域,则基于所述第一类障碍物区域的各栅格的空间体素坐标位置、高程信息和当前栅格的坐标位置,确定所述当前栅格的空间体素的权重,若当前栅格属于雷达点投影区域,则基于所述第一类障碍物区域的各栅格的空间体素坐标位置、高程信息、所述雷达点投影区域的高程信息及当前栅格的坐标位置,确定所述当前栅格的空间体素的权重;When the current grid belongs to the first type of obstacle area identified in the previous frame, if the current grid does not belong to the radar point cloud projection area, the weight of the spatial voxel of the current grid is determined based on the spatial voxel coordinate position, elevation information of each grid in the first type of obstacle area and the coordinate position of the current grid; if the current grid belongs to the radar point projection area, the weight of the spatial voxel of the current grid is determined based on the spatial voxel coordinate position, elevation information of each grid in the first type of obstacle area, the elevation information of the radar point projection area and the coordinate position of the current grid;
当当前栅格属于当前帧识别的第二类障碍物区域,若当前栅格不属于雷达点云投影区域,则基于所述第二类障碍物区域的各栅格的空间体素坐标位置确定所述当前栅格的空间体素的权重,若当前栅格属于雷达点投影区域,则基于所述第二类障碍物区域的各栅格的空间体素坐标位置及所述雷达点投影区域的高程信息,确定所述当前栅格的空间体素的权重。When the current grid belongs to the second type of obstacle area identified by the current frame, if the current grid does not belong to the radar point cloud projection area, the weight of the spatial voxel of the current grid is determined based on the spatial voxel coordinate position of each grid in the second type of obstacle area; if the current grid belongs to the radar point projection area, the weight of the spatial voxel of the current grid is determined based on the spatial voxel coordinate position of each grid in the second type of obstacle area and the elevation information of the radar point projection area.
在第七种示例性的实施方式中,所述获取栅格化的待检测路面区域中各栅格区域类型之前,还包括:In a seventh exemplary implementation, before obtaining the types of each grid area in the gridded road surface area to be detected, the method further includes:
基于一维高斯函数关系式,根据预设高斯曲线控制参数值、由不同栅格区域类型确定的中心位置坐标、栅格空间体素位置信息作为变量,生成权重计算关系式,以通过调用所述权重计算关系式确定各栅格空间体素的权重值。Based on the one-dimensional Gaussian function relationship, a weight calculation relationship is generated according to the preset Gaussian curve control parameter value, the center position coordinates determined by different grid area types, and the grid space voxel position information as variables, so as to determine the weight value of each grid space voxel by calling the weight calculation relationship.
在第八种示例性的实施方式中,基于更新后的空间体素权重更新障碍物区域内栅格的高程信息,包括:In an eighth exemplary implementation, updating elevation information of a grid in an obstacle area based on the updated spatial voxel weight includes:
预先训练高程分类网络,所述高程分类网络用于将深度估计问题转换为分类问题,且所述高程分类网络基于障碍物类型、高程真值编码信息、障碍物区域掩码信息、障碍物区域栅格的点云特征及图像特征、障碍物区域栅格的空间体素权重确定损失函数;Pre-training an elevation classification network, the elevation classification network is used to convert the depth estimation problem into a classification problem, and the elevation classification network determines a loss function based on obstacle type, elevation truth encoding information, obstacle area mask information, point cloud features and image features of the obstacle area grid, and spatial voxel weights of the obstacle area grid;
通过所述高程分类网络得到所述障碍物区域内栅格的高程信息。The elevation information of the grid in the obstacle area is obtained through the elevation classification network.
在第九种示例性的实施方式中,所述根据车辆前行过程中的当前帧点云数据对应的栅格新高程信息更新当前空间体素权重,包括:In a ninth exemplary implementation, the updating of the current spatial voxel weight according to the new grid elevation information corresponding to the current frame point cloud data during the vehicle's forward movement includes:
根据上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息、当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,确定当前帧障碍物区域包含的栅格;Determine the grids included in the obstacle area of the current frame according to the position information of the obstacle identified in the previous frame in the bird's-eye view grid area at the current moment and the position information of the obstacle identified in the current frame in the bird's-eye view grid area at the current moment;
获取所述当前帧障碍物区域包含栅格的高程信息,并对所述当前帧障碍物区域包含栅格的空间体素权重进行更新。The elevation information of the grid included in the obstacle area of the current frame is obtained, and the spatial voxel weights of the grid included in the obstacle area of the current frame are updated.
在第十种示例性的实施方式中,根据上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息、当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,确定当前帧障碍物区域包含的栅格,包括:In a tenth exemplary implementation, determining the grid included in the obstacle area of the current frame according to the position information of the obstacle identified in the previous frame in the bird's-eye view grid area at the current moment and the position information of the obstacle identified in the current frame in the bird's-eye view grid area at the current moment includes:
根据上一帧和当前帧对应的障碍物识别结果,确定对应障碍物区域内栅格;According to the obstacle recognition results corresponding to the previous frame and the current frame, the grid in the corresponding obstacle area is determined;
获取车辆行驶前方区域检测到的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息;所述车辆行驶前方区域在沿车辆行驶方向上与车辆间的最小距离大于等于所述待检测路面区域与车辆的最大距离;Obtaining the position information of the obstacle detected in the area ahead of the vehicle at the current moment in the bird's-eye view grid area; the minimum distance between the area ahead of the vehicle and the vehicle along the vehicle's driving direction is greater than or equal to the maximum distance between the road surface area to be detected and the vehicle;
根据车辆在相邻帧时间段内在局部世界坐标系的位置变化,更新上一帧障碍物区域内栅格的位置信息,以作为上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息;According to the position change of the vehicle in the local world coordinate system in the adjacent frame time period, the position information of the grid in the obstacle area of the previous frame is updated to serve as the position information of the grid area of the bird's-eye view of the obstacle identified in the previous frame at the current moment;
若当前点位置为上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或前方行驶区域识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,则为当前点设置第一预设值,以作为掩码值;若当前点位置不为上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或前方行驶区域识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,则为当前点设置第二预设值,以作为掩码值;If the current point position is the position information of the obstacle identified in the previous frame at the current moment in the bird's-eye view grid area, or the position information of the obstacle identified in the current frame at the current moment in the bird's-eye view grid area, or the position information of the obstacle identified in the front driving area at the current moment in the bird's-eye view grid area, then a first preset value is set for the current point as a mask value; if the current point position is not the position information of the obstacle identified in the previous frame at the current moment in the bird's-eye view grid area, or the position information of the obstacle identified in the current frame at the current moment in the bird's-eye view grid area, or the position information of the obstacle identified in the front driving area at the current moment in the bird's-eye view grid area, then a second preset value is set for the current point as a mask value;
根据所述待检测路面区域的各点的掩码值,生成障碍物区域掩码信息,以用于标识当前帧障碍物区域包含的栅格。Obstacle area mask information is generated according to the mask value of each point in the road area to be detected, so as to identify the grids included in the obstacle area of the current frame.
本发明还提供了一种电子设备,包括处理器,所述处理器用于执行存储器中存储的计算机程序时实现如前任一项所述路面障碍物三维形貌信息确定方法的步骤。The present invention also provides an electronic device, comprising a processor, wherein the processor is used to implement the steps of the method for determining three-dimensional shape information of road obstacles as described in any of the preceding items when executing a computer program stored in a memory.
本发明还提供了一种非易失性存储介质,所述非易失性存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如前任一项所述路面障碍物三维形貌信息确定方法的步骤。The present invention also provides a non-volatile storage medium, on which a computer program is stored. When the computer program is executed by a processor, the steps of the method for determining three-dimensional shape information of road obstacles as described in any of the preceding items are implemented.
本发明还提供了一种计算机程序产品,包括计算机程序/指令,该计算机程序/指令被处理器执行时实现如前任一项所述路面障碍物三维形貌信息确定方法的步骤。The present invention also provides a computer program product, including a computer program/instruction, which, when executed by a processor, implements the steps of the method for determining three-dimensional shape information of road obstacles as described in any of the preceding items.
本发明最后还提供了一种无人驾驶车辆,包括自动驾驶感知系统,所述自动驾驶感知系统用于执行存储器中存储的计算机程序时实现如前任一项所述路面障碍物三维形貌信息确定方法的步骤。Finally, the present invention also provides an unmanned vehicle, including an automatic driving perception system, wherein the automatic driving perception system is used to implement the steps of the method for determining three-dimensional morphological information of road obstacles as described in any of the preceding items when executing a computer program stored in a memory.
本发明提供的技术方案的优点在于,通过目标检测方法识别障碍物,并获取障碍物初始的三维形貌数据。在车辆不断前行过程采集多帧传感器数据,各帧数据均能获得每个栅格新的高程信息,随着障碍物不断靠近车辆,获取传感器数据的精度会越来越高,依据新的高程信息不断更新栅格空间体素的权重,便可以通过权重的更新逐步获取更为精确的高程信息和障碍物几何形态数据,进而能够精准确定道路坑洼区域和道路低矮凸起障碍物区域的三维形貌信息,不仅有利于保障自动驾驶车辆的安全平稳运行,有效提升乘车人的乘车体验,还有利于快速维修道路,延长道路及其基础设施的寿命。The advantage of the technical solution provided by the present invention is that obstacles are identified through target detection methods, and initial three-dimensional shape data of obstacles are obtained. Multiple frames of sensor data are collected as the vehicle moves forward, and each frame of data can obtain new elevation information for each grid. As the obstacle approaches the vehicle, the accuracy of obtaining sensor data will become higher and higher. The weights of the grid space voxels are continuously updated according to the new elevation information, and more accurate elevation information and obstacle geometry data can be gradually obtained through the update of weights, thereby accurately determining the three-dimensional shape information of potholes and low raised obstacle areas on the road, which is not only conducive to ensuring the safe and stable operation of autonomous driving vehicles and effectively improving the riding experience of passengers, but also conducive to rapid road repairs and extending the life of roads and their infrastructure.
此外,本发明还针对路面障碍物三维形貌信息确定方法提供了相应的实现电子设备、非易失性存储介质、计算机程序产品及无人驾驶车辆,进一步使得所述方法更具有实用性,所述电子设备、非易失性存储介质、计算机程序产品及无人驾驶车辆具有相应的优点。In addition, the present invention also provides corresponding electronic devices, non-volatile storage media, computer program products and unmanned vehicles for the method of determining the three-dimensional morphology information of road obstacles, further making the method more practical, and the electronic devices, non-volatile storage media, computer program products and unmanned vehicles have corresponding advantages.
上面已提及的技术特征、下面将要提及的技术特征以及单独地在附图中显示的技术特征可以任意地相互组合,只要被组合的技术特征不是相互矛盾的。所有的可行的特征组合都是在本文中明确地记载的技术内容。在同一个语句中包含的多个分特征之中的任一个分特征可以独立地被应用,而不必一定与其他分特征一起被应用。应当理解的是,以上的一般描述和后文的细节描述仅是示例性的,并不能限制本发明。The technical features mentioned above, the technical features to be mentioned below, and the technical features shown separately in the drawings can be arbitrarily combined with each other, as long as the combined technical features are not contradictory. All feasible feature combinations are technical contents clearly recorded in this article. Any of the multiple sub-features contained in the same sentence can be applied independently, and does not have to be applied together with other sub-features. It should be understood that the above general description and the detailed description below are only exemplary and cannot limit the present invention.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
为了更清楚的说明本发明或相关技术的技术方案,下面将对实施例或相关技术描述中所需要使用的附图作简单的介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions of the present invention or related technologies, the following briefly introduces the drawings required for use in the embodiments or related technical descriptions. Obviously, the drawings described below are only some embodiments of the present invention. For ordinary technicians in this field, other drawings can be obtained based on these drawings without paying creative work.
图1为本发明提供的一种路面障碍物三维形貌信息确定方法的流程示意图;FIG1 is a schematic flow chart of a method for determining three-dimensional shape information of a road obstacle provided by the present invention;
图2为本发明提供的栅格化后的待检测路面区域示意图;FIG2 is a schematic diagram of a road surface area to be detected after rasterization provided by the present invention;
图3为本发明提供一个示例性例子中的坐标系定义示意图;FIG3 is a schematic diagram of a coordinate system definition in an exemplary example of the present invention;
图4为本发明提供的路面纵向剖面上的空间体素示意图;FIG4 is a schematic diagram of space voxels on a longitudinal section of a road surface provided by the present invention;
图5为本发明提供的高程估算侧视示意图;FIG5 is a side view schematic diagram of elevation estimation provided by the present invention;
图6为本发明提供的障碍物的掩码示意图;FIG6 is a schematic diagram of a mask of an obstacle provided by the present invention;
图7为本发明提供的障碍物的空间体素权重示意图;FIG7 is a schematic diagram of the spatial voxel weights of obstacles provided by the present invention;
图8为本发明提供的障碍物的几何形态描述数据示意图;FIG8 is a schematic diagram of geometric description data of an obstacle provided by the present invention;
图9为本发明提供的另一种路面障碍物三维形貌信息确定方法的流程示意图;FIG9 is a schematic flow chart of another method for determining three-dimensional shape information of road obstacles provided by the present invention;
图10为本发明提供的路面障碍物三维形貌信息确定装置的一种具体实施方式结构图;FIG10 is a structural diagram of a specific implementation of a device for determining three-dimensional shape information of road obstacles provided by the present invention;
图11为本发明提供的电子设备的一种具体实施方式结构图;FIG11 is a structural diagram of a specific implementation of an electronic device provided by the present invention;
图12为本发明提供的无人驾驶车辆的障碍物识别示意图。FIG. 12 is a schematic diagram of obstacle recognition for an unmanned vehicle provided by the present invention.
具体实施方式DETAILED DESCRIPTION
为了使本技术领域的人员更好地理解本发明的技术方案,下面结合附图和具体实施方式对本发明作进一步的详细说明。其中,说明书及上述附图中的术语“第一”、“第二”、“第三”、“第四”等是用于区别不同的对象,而不是用于描述特定的顺序。此外术语“包括”和“具有”以及二者的任何变形,意图在于覆盖不排他的包含。术语“示例性”意为“用作例子、实施例或说明性”。这里作为“示例性”所说明的任何实施例不必解释为优于或好于其它实施例。In order to enable those skilled in the art to better understand the technical solution of the present invention, the present invention is further described in detail below in conjunction with the accompanying drawings and specific embodiments. Among them, the terms "first", "second", "third", "fourth", etc. in the specification and the above-mentioned drawings are used to distinguish different objects, rather than to describe a specific order. In addition, the terms "including" and "having" and any variations of the two are intended to cover non-exclusive inclusions. The term "exemplary" means "used as an example, embodiment or illustrative". Any embodiment described here as "exemplary" is not necessarily interpreted as being superior or better than other embodiments.
可以理解的是,道路坑洼或道路低矮凸起障碍物会导致自动驾驶车辆行驶不稳,继而发生颠簸,这不仅会影响乘车人的舒适度,而且会导致车辆突然失控,引发交通事故。若车辆长期在坑洼和低矮凸起障碍物路面上行驶,会损坏车辆的悬挂系统、轮胎、传感器,进而影响车辆性能和车辆安全性。此外,道路坑洼会破坏路面材料,导致路基的沉降,如果不及时修复,坑洼区域极易扩大和加深,加速道路基础设施的损坏,影响道路的使用寿命和承载能力。为了避免上述问题,车辆行驶过程中,会实时监测前方行驶路面是否存在坑洼区域或凸起障碍物区域,通过提醒驾驶员路面检测情况,并发出警告来减慢汽车的速度,或者是通过智能悬挂系统吸收和抵消由路面坑洼/低矮凸起障碍物引起的冲击和振动。It is understandable that potholes or low raised obstacles on the road can cause the autonomous vehicle to drive unsteadily and then bump, which will not only affect the comfort of the passengers, but also cause the vehicle to suddenly lose control and cause traffic accidents. If the vehicle is driven on potholes and low raised obstacles for a long time, it will damage the vehicle's suspension system, tires, and sensors, which will affect the vehicle's performance and safety. In addition, potholes on the road will damage the road material and cause the settlement of the roadbed. If they are not repaired in time, the potholes will easily expand and deepen, accelerating the damage to the road infrastructure and affecting the service life and carrying capacity of the road. In order to avoid the above problems, during the driving process of the vehicle, it will monitor in real time whether there are potholes or raised obstacle areas on the road ahead, remind the driver of the road detection situation, and issue a warning to slow down the speed of the car, or absorb and offset the impact and vibration caused by potholes/low raised obstacles on the road through the intelligent suspension system.
目前,传统监测道路是否存在坑洼或低矮凸起障碍物的方式是人工目视检查,通过结构工程师和认证检查员定期检查路面是否存在坑洼或低矮凸起障碍物,并报告这类目标区域的位置。这种方式依赖检查员和工程师的个人经验,不仅无法保证检测结果准确度,而且效率低、成本高、还会存在危险。为了解决人工目视检查存在的弊端问题,相关技术通过计算机视觉技术实现高效率和安全性更好的道路障碍物如坑洼区域或低矮障碍物的识别与检测。At present, the traditional way to monitor whether there are potholes or low raised obstacles on the road is manual visual inspection, where structural engineers and certified inspectors regularly check whether there are potholes or low raised obstacles on the road surface and report the location of such target areas. This method relies on the personal experience of inspectors and engineers, which not only cannot guarantee the accuracy of the detection results, but is also inefficient, costly, and dangerous. In order to solve the drawbacks of manual visual inspection, related technologies use computer vision technology to achieve more efficient and safer identification and detection of road obstacles such as potholes or low obstacles.
相关技术利用计算机视觉算法通过如3D(三维)点云建模实现三维道路成像,然后基于分割方法检测路面障碍物区域,以提高道路目标区域的检测的准确性。这种方式通过一些算法,如RANSAC(Random Sample Consensus,随机样本一致算法)提高道路点云建模的鲁棒性,DCNN(Deep Convolution Neural Network,深度卷积神经网络模型)提高图像分类精度,但是,这类方法需要一个小视场,且是建立在一个单帧3D道路点云是一个平面或二次曲面的基础,检测精度及场景适应性还难以达到实用,且难以高精度获取路面凹坑的宽度、深度、体积等几何特性,进而无法精准识别并计算路面障碍物的三维形貌信息。举例来说,一种相关技术通过多传感器融合实现对无人车路面坑洼的检测,该方法通过声波对前方路面扫描,当检测到坑洼时,利用相机抓取路面图像,然后将预处理后的图像进行阈值分割得到二值化图像,通过统计二值化图像中白色像素点的数量来判断坑洼中是否有水。通过摄像头拍摄的图像信息计算坑洼边缘距离两侧道路边缘的宽度,并计算无人车通过坑洼的最小扭矩,最后通过最小扭矩与最大扭矩的关系来判断无人车是否能够安全通过。该方法图像二值化受阈值影响非常大,难以适应各种光照条件,且无法给出坑洼深度等信息,也就无法精准计算坑洼区域和低矮障碍物的三维形貌信息。另一种相关技术利用RGB‐D(RedGreen Blue-Depth,具有深度信息的彩色图像)深度相机实现低矮凸起障碍物、路面坑洼检测方,该方法利用RGB‐D深度相机的原始点云进行预处理得到对应的平面点云,从该平面点云中提取凸起低矮障碍物点云数据,并将提取的凸起低矮障碍物点云投影到平面模型参数平面上,对模型参数平面上的凸起低矮障碍物点云进行降采样、凸包分割,将凸包变换到/map(地图)坐标系下并发布,对投影后的凸起低矮障碍物点云合并到地平面点云以形成新平面点云,将新平面点云变换到base_footprint(机器人本体坐标系)下转成地平面二维栅格地图,将地平面二维栅格地图进行伪坑洞滤波,边缘提取方法对坑洞进行提取。由于RGB-D相机有效工作距离非常有限,通常是在10米以内,且在户外强光下基本失效,这就导致RGB-D相机生成的点云精度因受发生器投影密度影响而较差,不仅使用场景受限,而且实际使用过程有效性有待验证,无法确保坑洼区域和低矮障碍物的三维形貌信息的计算准确度。另一种相关技术基于神经网络和双目立体视觉实现对路面坑槽的检测,该方法使用安装在车辆前方的双目相机获取道路场景的左右视图;使用目标检测神经网络对左视图进行检测。当检测到坑槽存在时,对左右视图立体匹配得到深度图,利用神经网络预测的坑槽位置信息裁切出坑槽以及坑槽四周路面的深度图,最后使用RANSAC和遗传算法相结合的方法拟合坑槽周边的路平面,从而计算出坑槽点云到拟合平面的距离,对距离求和平均得到坑槽深度。通过将坑槽点云投影到拟合路平面上来确定投影点的最小外接矩形,从而确定最小外接矩形的面积得到坑槽的面积。但是,由于视差存在空洞点,这种方法在纵深方向估计误差较大,难以获取高精度的坑洼深度和凸起高度,进而无法精准识别坑洼区域和低矮障碍物,也无法精准确定坑洼区域和低矮障碍物的三维形貌信息。Related technologies use computer vision algorithms to achieve three-dimensional road imaging through 3D point cloud modeling, and then detect road obstacle areas based on segmentation methods to improve the accuracy of road target area detection. This method uses some algorithms, such as RANSAC (Random Sample Consensus) to improve the robustness of road point cloud modeling and DCNN (Deep Convolution Neural Network) to improve image classification accuracy. However, this type of method requires a small field of view and is based on a single-frame 3D road point cloud that is a plane or quadratic surface. The detection accuracy and scene adaptability are still difficult to achieve practicality, and it is difficult to obtain the geometric characteristics of road pits such as width, depth, and volume with high precision, and thus it is impossible to accurately identify and calculate the three-dimensional morphological information of road obstacles. For example, a related technology realizes the detection of potholes on the road surface of unmanned vehicles through multi-sensor fusion. This method uses sound waves to scan the road surface ahead. When a pothole is detected, the camera is used to capture the road image, and then the pre-processed image is threshold segmented to obtain a binary image. The number of white pixels in the binary image is counted to determine whether there is water in the pothole. The image information captured by the camera is used to calculate the width of the pothole edge from the road edge on both sides, and the minimum torque for the unmanned vehicle to pass through the pothole is calculated. Finally, the relationship between the minimum torque and the maximum torque is used to determine whether the unmanned vehicle can pass safely. This method is greatly affected by the threshold value for image binarization, making it difficult to adapt to various lighting conditions, and it is unable to provide information such as the depth of the pothole, so it is impossible to accurately calculate the three-dimensional shape information of the pothole area and low obstacles. Another related technology uses an RGB-D (RedGreenBlue-Depth, a color image with depth information) depth camera to detect low raised obstacles and potholes on the road. This method uses the original point cloud of the RGB-D depth camera for preprocessing to obtain the corresponding plane point cloud, extracts the raised low obstacle point cloud data from the plane point cloud, and projects the extracted raised low obstacle point cloud onto the plane model parameter plane, downsamples the raised low obstacle point cloud on the model parameter plane, performs convex hull segmentation, transforms the convex hull to the /map (map) coordinate system and publishes it, merges the projected raised low obstacle point cloud into the ground plane point cloud to form a new plane point cloud, transforms the new plane point cloud to the base_footprint (robot body coordinate system) into a ground plane two-dimensional grid map, performs pseudo pothole filtering on the ground plane two-dimensional grid map, and extracts potholes using the edge extraction method. Since the effective working distance of RGB-D cameras is very limited, usually within 10 meters, and they are basically ineffective under strong outdoor light, the accuracy of the point cloud generated by the RGB-D camera is poor due to the influence of the projection density of the generator. Not only is the use scenario limited, but the effectiveness of the actual use process needs to be verified, and the calculation accuracy of the three-dimensional morphological information of potholes and low obstacles cannot be ensured. Another related technology is based on neural networks and binocular stereo vision to detect potholes on the road. This method uses a binocular camera installed in front of the vehicle to obtain the left and right views of the road scene; and uses the target detection neural network to detect the left view. When the existence of a pothole is detected, the left and right views are stereo matched to obtain a depth map, and the pothole position information predicted by the neural network is used to cut out the depth map of the pothole and the road surface around the pothole. Finally, a method combining RANSAC and genetic algorithms is used to fit the road plane around the pothole, so as to calculate the distance from the pothole point cloud to the fitting plane, and the distance is summed and averaged to obtain the pothole depth. The minimum circumscribed rectangle of the projection point is determined by projecting the pothole point cloud onto the fitting road plane, and the area of the minimum circumscribed rectangle is determined to obtain the area of the pothole. However, due to the existence of void points in parallax, this method has large estimation errors in the depth direction, making it difficult to obtain high-precision pothole depths and bump heights, and thus unable to accurately identify pothole areas and low obstacles, nor to accurately determine the three-dimensional morphology information of pothole areas and low obstacles.
鉴于此,本发明通过点云数据和图像数据确定障碍物的三维形貌信息,然后随着车辆的不断前行,障碍物所在区域不断接近车辆,随着新的高程信息不断更新空间体素的权重,从而可以不断提升高程信息和几何形态数据的精度,进而有效提升低矮障碍物/坑洼识别与三维形貌信息计算精度。在介绍了本发明的技术方案后,下面详细的说明本发明的各种非限制性实施方式。为了更好的说明本发明,在下文的具体实施方式中给出了众多的具体细节。本领域技术人员应当理解,没有这些具体细节,本发明同样可以实施。在另外一些实例中,对于所属领域技术人员熟知的方法、手段、元件和电路未作详细描述,以便于凸显本发明的主旨。In view of this, the present invention determines the three-dimensional shape information of the obstacle through point cloud data and image data, and then as the vehicle continues to move forward, the area where the obstacle is located continues to approach the vehicle, and the weights of the spatial voxels are continuously updated with the new elevation information, so that the accuracy of the elevation information and geometric data can be continuously improved, thereby effectively improving the accuracy of low obstacle/pothole recognition and three-dimensional shape information calculation. After introducing the technical scheme of the present invention, various non-limiting embodiments of the present invention are described in detail below. In order to better illustrate the present invention, numerous specific details are given in the specific embodiments below. Those skilled in the art should understand that the present invention can also be implemented without these specific details. In other examples, methods, means, components and circuits well known to those skilled in the art are not described in detail in order to highlight the subject matter of the present invention.
首先请参见图1,图1为本实施例提供的一种路面障碍物三维形貌信息确定方法的流程示意图,本实施例可包括以下内容:First, please refer to FIG. 1 , which is a flow chart of a method for determining three-dimensional shape information of a road obstacle provided by this embodiment. This embodiment may include the following contents:
S101:获取栅格化处理的待检测路面区域的点云数据和图像数据在目标视角下的特征。S101: Acquire the features of the rasterized point cloud data and image data of the road area to be detected under the target viewing angle.
在本实施例中,待检测路面区域位于车辆行驶前方,以车头所在位置向前延伸的一段空间区域,其尺寸可根据实际情况来灵活选择,车辆可以为任何一种具有自动驾驶功能的车辆,并不限制于自动驾驶车辆或无人驾驶车辆,车辆也可称为自车(ego vehicle,当前自动驾驶车辆)。在计算机视觉领域,通过一个网络结构预测高程信息来替代单目深度估计,诸如BEVheight、RoadBEV(鸟瞰视图中的道路表面重建),但是这种方法的预测高程的误差可以达到2厘米左右,而该误差导致低矮凸起障碍物及坑洼区域无法实现精确判断,本发明通过将待检测路面区域划分为多个栅格,也即栅格化处理,如图2所示,车辆行驶前方的待检测路面区域在目标视角下划分为多个栅格,由于车辆安装的激光雷达在车辆行驶过程中会发射激光光线,图2中的黑色区域表示激光雷达点云在路面投影点,白色区域表示没有数据,栅格尺寸和栅格数量可根据实际情况,如计算量、障碍物检测精度及车辆安装的激光雷达的配置参数,来灵活选择,这均不影响本发明的实现。其中,障碍物为本发明需要检测在待检测路面区域进行检测的物体,该物体在垂直于路面水平面的方向上具有较小的高度,包括但并不限制于低矮凸起障碍物,如减速带,凸起裂缝,及坑洼区域,如水坑、浅坑、深坑。考虑到路面与车辆坐标系xy平面可能存在一定夹角,如果不考虑该情况,会将该误差引入高程计算,对于高程较小的区域,也即低矮凸起障碍物及坑洼区域这样的障碍物,会造成误判,所以本发明还为每个栅格在垂直于路面水平方向上设置有空间体素,空间体素也即在垂直于路面水平方向上设置多个空间立方体,空间体素可以反应地面的高低起伏情况,其在后续精确计算低矮凸起障碍物及坑洼区域高程信息中使用,以便用于路侧感知、智慧交通、V2X(vehicle to everything,车对外界的信息交换)、智能安防、协同感知等领域中进行避障、环境感知等。In this embodiment, the road surface area to be detected is located in front of the vehicle, and is a space area extending forward from the position of the front of the vehicle. Its size can be flexibly selected according to actual conditions. The vehicle can be any vehicle with an automatic driving function, and is not limited to an automatic driving vehicle or an unmanned vehicle. The vehicle can also be called an ego vehicle (the current automatic driving vehicle). In the field of computer vision, a network structure is used to predict elevation information to replace monocular depth estimation, such as BEVheight and RoadBEV (road surface reconstruction in a bird's-eye view). However, the error in the predicted elevation of this method can reach about 2 cm, and this error makes it impossible to accurately judge low raised obstacles and potholes. The present invention divides the road surface area to be detected into multiple grids, that is, rasterization processing. As shown in Figure 2, the road surface area to be detected in front of the vehicle is divided into multiple grids under the target perspective. Since the laser radar installed on the vehicle will emit laser light during the driving process of the vehicle, the black area in Figure 2 represents the projection point of the laser radar point cloud on the road surface, and the white area represents no data. The grid size and the number of grids can be flexibly selected according to actual conditions, such as the amount of calculation, the obstacle detection accuracy and the configuration parameters of the laser radar installed on the vehicle, which does not affect the implementation of the present invention. Among them, the obstacle is an object that needs to be detected in the road area to be detected in the present invention, and the object has a small height in the direction perpendicular to the horizontal plane of the road surface, including but not limited to low raised obstacles, such as speed bumps, raised cracks, and potholes, such as puddles, shallow pits, and deep pits. Considering that there may be a certain angle between the road surface and the xy plane of the vehicle coordinate system, if this situation is not considered, the error will be introduced into the elevation calculation, and for areas with smaller elevations, that is, obstacles such as low raised obstacles and potholes, misjudgment will be caused. Therefore, the present invention also provides each grid with a spatial voxel in a direction perpendicular to the horizontal direction of the road surface. The spatial voxel is a plurality of spatial cubes arranged in a direction perpendicular to the horizontal direction of the road surface. The spatial voxel can reflect the ups and downs of the ground, which is used in the subsequent accurate calculation of the elevation information of low raised obstacles and potholes, so as to be used in roadside perception, smart transportation, V2X (vehicle to everything, information exchange between vehicles and the outside world), smart security, collaborative perception and other fields for obstacle avoidance, environmental perception, etc.
其中,点云数据为车辆激光雷达采集的点云数据,图像数据为车载相机获取的车辆行驶过程所采集的图像,可将获取的点云数据和图像数据转换至目标视角下,目标视角可根据实际采用的三维形貌信息所采用的方法来确定,示例性的,目标视角为鸟瞰视角,为了便于描述,将转换后所得到特征定义为点云特征和图像特征。为了提高路面障碍物三维形貌信息的确定精度,在获取点云数据和图像数据之前,可通过车辆的内部机制完成对车辆各传感器的标定,传感器包括但并不限制于惯性测量单元、里程计、激光雷达、前向相机,获取传感器之间坐标系变换关系,示例性的,可直接采用车辆出厂前的标定参数。为了进一步提高路面障碍物三维形貌信息生成精度,可通过数据时间同步机制确保点云数据及图像数据时间对齐。Among them, the point cloud data is the point cloud data collected by the vehicle laser radar, and the image data is the image collected by the vehicle camera during the vehicle driving process. The acquired point cloud data and image data can be converted to the target perspective. The target perspective can be determined according to the method used for the actual three-dimensional shape information. For example, the target perspective is a bird's-eye view. For the convenience of description, the features obtained after the conversion are defined as point cloud features and image features. In order to improve the accuracy of determining the three-dimensional shape information of road obstacles, before acquiring the point cloud data and image data, the calibration of each sensor of the vehicle can be completed through the internal mechanism of the vehicle. The sensor includes but is not limited to an inertial measurement unit, an odometer, a laser radar, and a forward camera. The coordinate system transformation relationship between the sensors is obtained. For example, the calibration parameters before the vehicle leaves the factory can be directly used. In order to further improve the accuracy of generating three-dimensional shape information of road obstacles, the time alignment of point cloud data and image data can be ensured through a data time synchronization mechanism.
S102:将点云特征和图像特征进行融合,并通过对融合特征进行目标检测处理,得到待检测路面区域内的障碍物的初始三维形貌信息。S102: Fusing the point cloud features and the image features, and performing target detection processing on the fused features to obtain initial three-dimensional shape information of obstacles in the road area to be detected.
当上个步骤获取到目标视角下的多模态特征之后,为了进一步提升障碍物,如低矮障碍物及坑洼区域识别精度,可融合点云特征和图像特征。基于融合特征,采用任何一种目标检测算法,如R-CNN(Region-based Convolutional Neural Networks,基于区域的卷积神经网络)、Fast R-CNN(一种快速的基于区域的卷积神经网络)、Faster R-CNN(一种更快的基于区域的卷积神经网络)和YOLOv3(You Only Look Once version 3,网络模型名称),对融合特征进行目标检测,目标是指本实施例需要识别的障碍物,通过目标检测算法识别到融合特征中的障碍物,可以进一步的确定障碍物的高程信息、位置、几何形态数据,如朝向、长、宽、区域面积,并将其作为识别到的障碍物的三维形貌信息,由于本步骤所确定的三维形貌信息精度不高,会通过车辆不断接近障碍物所采集的数据对三维形貌信息进行不断更新,所以本步骤将其定义为初始三维形貌信息。After the multimodal features under the target perspective are obtained in the previous step, in order to further improve the recognition accuracy of obstacles, such as low obstacles and potholes, point cloud features and image features can be fused. Based on the fused features, any target detection algorithm, such as R-CNN (Region-based Convolutional Neural Networks), Fast R-CNN (a fast region-based convolutional neural network), Faster R-CNN (a faster region-based convolutional neural network) and YOLOv3 (You Only Look Once version 3, network model name), is used to perform target detection on the fused features. The target refers to the obstacle to be identified in this embodiment. The obstacle in the fused features identified by the target detection algorithm can further determine the elevation information, position, and geometric data of the obstacle, such as direction, length, width, and area, and use it as the three-dimensional shape information of the identified obstacle. Since the three-dimensional shape information determined in this step is not accurate, the three-dimensional shape information will be continuously updated through the data collected by the vehicle continuously approaching the obstacle, so this step defines it as the initial three-dimensional shape information.
S103:根据车辆前行过程中不断获取的各帧点云数据对应的栅格新高程信息,更新当前时刻的空间体素权重,并基于更新后的空间体素权重更新障碍物区域内栅格的高程信息和几何形态数据,以对初始三维形貌信息进行不断更新。S103: According to the new elevation information of the grid corresponding to each frame of point cloud data continuously acquired during the vehicle's forward movement, the spatial voxel weight at the current moment is updated, and the elevation information and geometric data of the grid in the obstacle area are updated based on the updated spatial voxel weight to continuously update the initial three-dimensional shape information.
当S101步骤将待检测路面区域划分为多个具有空间体素的栅格,且每个栅格设置有空间体素,通过平面拟合方法可以确定待检测路面区域与车辆坐标系的水平面的倾斜角度,示例性的,对各栅格的点云数据进行平面拟合,得到对应的三维平面方程,根据三维平面方程得到路面法向量,路面法向量与车辆坐标系水平面的角度即为倾斜角度。根据激光雷达安装位置及车辆结构参数,可以获取激光雷达坐标系原点到地面距离,基于三角几何原理,根据该距离和倾斜角度可以得到车辆坐标系X轴与路面交叉点的x坐标,根据该坐标与栅格的横坐标的差值以及倾斜角度便可以得到相应栅格的高程信息。对于经典的LSS(Lattice Sampling and Segmentation,用于三维点云处理和场景理解的技术),为了估计每个像素的深度信息,通常认为对于每个像素,其对应的沿着光射线上设定栅格,且每个栅格并没有类似于优先级的权重设定。本实施例为了精准计算障碍物的三维形貌信息,本实施例还可进一步为空间体素设置权重,随着车辆不断前行,即不断靠近低矮障碍物/坑洼区域,车载传感器会不断采集点云数据,对于每帧点云数据,均可获得每个栅格的新的高程信息,依据新的高程信息不断更新权重,通过迭代机制,逐步获取更为精确的高程信息,从而通过更新权重实现对高程信息和几何形态数据的不断更新,得到更加精准的三维形貌数据。本实施例基于点云数据引导的目标高程信息计算方法,相比LSS,可以有效提高深度计算精度,此外,基于LSS的深度或高程估计方法对标注数据的精度要求较高,且标注难度较大。而本实施例通过具有一定精度的初始真值基础上,进一步通过回归方法,可以进一步提升深度预测精度,并无需大量标注数据。When step S101 divides the road surface area to be detected into multiple grids with spatial voxels, and each grid is provided with spatial voxels, the inclination angle between the road surface area to be detected and the horizontal plane of the vehicle coordinate system can be determined by the plane fitting method. For example, the point cloud data of each grid is plane fitted to obtain the corresponding three-dimensional plane equation, and the road surface normal vector is obtained according to the three-dimensional plane equation. The angle between the road surface normal vector and the horizontal plane of the vehicle coordinate system is the inclination angle. According to the installation position of the laser radar and the structural parameters of the vehicle, the distance from the origin of the laser radar coordinate system to the ground can be obtained. Based on the principle of triangular geometry, the x-coordinate of the intersection of the vehicle coordinate system X-axis and the road surface can be obtained according to the distance and the inclination angle. , the elevation information of the corresponding grid can be obtained according to the difference between the coordinate and the horizontal coordinate of the grid and the inclination angle. For the classic LSS (Lattice Sampling and Segmentation, a technology used for three-dimensional point cloud processing and scene understanding), in order to estimate the depth information of each pixel, it is generally believed that for each pixel, the corresponding grid is set along the light ray, and each grid does not have a weight setting similar to a priority. In order to accurately calculate the three-dimensional shape information of the obstacle, this embodiment can further set weights for the spatial voxels. As the vehicle continues to move forward, that is, it continues to approach low obstacles/potholes, the on-board sensor will continuously collect point cloud data. For each frame of point cloud data, new elevation information of each grid can be obtained, and the weight is continuously updated according to the new elevation information. Through the iterative mechanism, more accurate elevation information is gradually obtained, so that the elevation information and geometric data are continuously updated by updating the weights, and more accurate three-dimensional shape data is obtained. The target elevation information calculation method based on point cloud data in this embodiment can effectively improve the depth calculation accuracy compared to LSS. In addition, the depth or elevation estimation method based on LSS has high requirements on the accuracy of the labeled data and is difficult to label. However, this embodiment can further improve the depth prediction accuracy by using a regression method based on the initial true value with a certain accuracy, without the need for a large amount of labeled data.
在本实施例提供的技术方案中,通过目标检测方法识别障碍物,并获取障碍物初始的三维形貌数据。在车辆不断前行过程采集多帧传感器数据,各帧数据均能获得每个栅格新的高程信息,随着障碍物不断靠近车辆,获取传感器数据的精度会越来越高,依据新的高程信息不断更新栅格空间体素的权重,便可以通过权重的更新逐步获取更为精确的高程信息和障碍物几何形态数据,进而能够精准确定道路坑洼区域和道路低矮凸起障碍物区域的三维形貌信息,不仅有利于保障自动驾驶车辆的安全平稳运行,有效提升乘车人的乘车体验,还有利于快速维修道路,延长道路及其基础设施的寿命。In the technical solution provided in this embodiment, obstacles are identified through a target detection method, and initial three-dimensional shape data of the obstacles are obtained. Multiple frames of sensor data are collected as the vehicle moves forward, and each frame of data can obtain new elevation information for each grid. As the obstacle approaches the vehicle, the accuracy of obtaining sensor data will become higher and higher. The weights of the grid space voxels are continuously updated based on the new elevation information, and more accurate elevation information and obstacle geometry data can be gradually obtained through the update of the weights, thereby accurately determining the three-dimensional shape information of the potholes and low raised obstacle areas on the road, which is not only conducive to ensuring the safe and stable operation of the autonomous driving vehicle, effectively improving the riding experience of the passengers, but also conducive to rapid road maintenance and extending the life of the road and its infrastructure.
当上述实施例确定障碍物的三维形貌信息之后,为了在诸如路侧感知场景、智慧交通场景、V2X(vehicle to everything,即车对外界的信息交换)场景、智能安防场景、协同感知场景中起到相应的作用,如用于准确评估低矮凸起障碍物/坑洼区域的危害性,如自动规划路线时的避障处理。还可基于车辆坐标系与局部世界坐标系之间的变换关系,将待检测路面区域的障碍物的三维形貌信息转换至局部世界坐标系。After the above embodiments determine the three-dimensional shape information of the obstacle, in order to play a corresponding role in roadside perception scenarios, smart traffic scenarios, V2X (vehicle to everything, i.e., vehicle-to-external information exchange) scenarios, smart security scenarios, and collaborative perception scenarios, such as accurately assessing the harmfulness of low raised obstacles/potholes, such as obstacle avoidance during automatic route planning. The three-dimensional shape information of the obstacle in the road area to be detected can also be converted to the local world coordinate system based on the transformation relationship between the vehicle coordinate system and the local world coordinate system.
在本实施例中,为了数据处理,本发明还对涉及到的坐标系进行相应的定义,如图3所示,车辆包括远场相机、近场相机、激光雷达、里程计、惯性测量单元,为了绘图方便,没有区分远场相机和近场相机,也没有绘制里程计、惯性测量单元。在图3中,为相机坐标系,沿着相机光轴方向。为标准相机坐标系,该坐标系为沿着旋转,其中垂直向下。为激光雷达坐标系,为局部世界坐标系,该坐标系定义在路面上,其中,垂直向上,为水平向前,为水平向左,通常车辆坐标系可采用激光雷达坐标系,局部世界坐标系为车辆行驶路面所在的坐标系。示例性的,车辆坐标系与局部世界坐标系之间的变换关系的获取过程可为:获取当前时刻的点云数据和局部地图点云;通过对点云数据和局部地图点云进行点云匹配,确定车辆坐标系与局部世界坐标系之间的变换关系;其中,局部地图点云为沿着时间轴,对不同时刻的点云数据进行拼接后生成的数据。当然,为了进一步提高三维形貌信息的精度,在确定坐标间变换关系时,点云数据可先进行稠密化处理,并在获取激光雷达点云数据时,通过里程计和惯性测量单元的车辆姿态数据对激光雷达点云数据进行畸变矫正。点云匹配算法可采用任何一种算法,如ICP(Iterative Closest Point,迭代最近点算法)、NDT(Normal DistributionTransform,正态分布变换),这均不影响本发明的实现。In this embodiment, for data processing, the present invention also defines the coordinate system involved accordingly. As shown in FIG3 , the vehicle includes a far-field camera, a near-field camera, a laser radar, an odometer, and an inertial measurement unit. For the convenience of drawing, the far-field camera and the near-field camera are not distinguished, and the odometer and the inertial measurement unit are not drawn. In FIG3 , is the camera coordinate system, Along the camera optical axis. is the standard camera coordinate system, which is Along Rotation, where Vertically downward. is the laser radar coordinate system, is the local world coordinate system, which is defined on the road surface, where Vertically upward, For horizontal forward, To the left horizontally, the vehicle coordinate system can usually adopt the laser radar coordinate system, and the local world coordinate system is the coordinate system of the road where the vehicle is traveling. Exemplarily, the process of obtaining the transformation relationship between the vehicle coordinate system and the local world coordinate system can be: obtaining the point cloud data and the local map point cloud at the current moment; determining the transformation relationship between the vehicle coordinate system and the local world coordinate system by matching the point cloud data and the local map point cloud; wherein the local map point cloud is the data generated by splicing the point cloud data at different moments along the time axis. Of course, in order to further improve the accuracy of the three-dimensional morphology information, when determining the transformation relationship between coordinates, the point cloud data can be first densified, and when obtaining the laser radar point cloud data, the laser radar point cloud data can be distorted and corrected by the vehicle posture data of the odometer and the inertial measurement unit. The point cloud matching algorithm can adopt any algorithm, such as ICP (Iterative Closest Point, iterative closest point algorithm) and NDT (Normal Distribution Transform, normal distribution transformation), which does not affect the implementation of the present invention.
上述实施例对如何将待检测路面区域进行栅格化处理并不做任何限定,基于上述实施例,本发明还提供了待检测路面区域的一种示例性的栅格化处理实现方式,可包括下述内容:The above embodiment does not limit how to perform rasterization processing on the road surface area to be detected. Based on the above embodiment, the present invention also provides an exemplary rasterization processing implementation method of the road surface area to be detected, which may include the following contents:
作为一种示例性的栅格化方式,可获取自动驾驶车辆安装的激光雷达线数;根据激光雷达线数确定栅格数量及栅格尺寸,并根据栅格数据及栅格尺寸确定待检测路面区域的空间尺寸;按照栅格数量及栅格尺寸,对位于行驶方向前方相应空间尺寸内的待检测路面区域进行栅格化处理。As an exemplary rasterization method, the number of laser radar lines installed on the autonomous driving vehicle can be obtained; the number of grids and the grid size are determined according to the number of laser radar lines, and the spatial size of the road surface area to be detected is determined according to the grid data and the grid size; according to the number of grids and the grid size, the road surface area to be detected within the corresponding spatial size in front of the driving direction is rasterized.
作为另一种示例性的栅格化方式,可获取自动驾驶车辆安装的激光雷达的分辨率;根据激光雷达的分辨率确定栅格数量及栅格尺寸,并根据栅格数据及栅格尺寸确定待检测路面区域的空间尺寸;按照栅格数量及栅格尺寸,对位于行驶方向前方相应空间尺寸内的待检测路面区域进行栅格化处理。As another exemplary rasterization method, the resolution of the laser radar installed on the autonomous driving vehicle can be obtained; the number of grids and the grid size are determined according to the resolution of the laser radar, and the spatial size of the road surface area to be detected is determined according to the grid data and the grid size; according to the number of grids and the grid size, the road surface area to be detected within the corresponding spatial size in front of the driving direction is rasterized.
在本实施例中,鸟瞰视图栅格区域的栅格数量,以及各栅格对应在待检测路面的空间大小可以根据激光雷达线数或分辨率进行灵活调整,举例来说,车辆配备的是32线机械式激光雷达,考虑的精度及计算量问题,鸟瞰视图栅格区域栅格数量为2000×2000,每个栅格对应的空间大小为0.01米×0.01米,鸟瞰视图栅格区域对应的空间区域长宽分别为20米和20米,对于20米以外,为远场检测区域。车辆配备的是128线机械式激光雷达,则鸟瞰视图栅格区域对应空间大小可以提升至40米×30米,为了提升检测识别精度,可将每个栅格对应的空间大小设定为0.01米×0.01米。In this embodiment, the number of grids in the bird's-eye view grid area and the size of the space corresponding to each grid on the road to be detected can be flexibly adjusted according to the number of laser radar lines or resolution. For example, the vehicle is equipped with a 32-line mechanical laser radar. Considering the accuracy and calculation amount, the number of grids in the bird's-eye view grid area is 2000×2000, and the space size corresponding to each grid is 0.01 meter×0.01 meter. The length and width of the space area corresponding to the bird's-eye view grid area are 20 meters and 20 meters respectively. For more than 20 meters, it is a far-field detection area. If the vehicle is equipped with a 128-line mechanical laser radar, the size of the space corresponding to the bird's-eye view grid area can be increased to 40 meters×30 meters. In order to improve the detection and recognition accuracy, the size of the space corresponding to each grid can be set to 0.01 meter×0.01 meter.
为了进一步提高路面障碍物的识别精度,基于上述实施例,本发明还提供了栅格的空间体素的参数的确定方法,可包括下述内容:In order to further improve the recognition accuracy of road obstacles, based on the above embodiments, the present invention further provides a method for determining parameters of spatial voxels of a grid, which may include the following contents:
获取待检测路面区域的障碍物最大高程;按照预设精度阈值划分障碍物最大高程,以确定各栅格的空间体素包含的空间立方体总个数;根据障碍物最大高程、空间立方体总个数和预设空间立方体平面尺寸,确定各栅格的空间体素的空间立方体对应的真实物理空间尺寸。Obtain the maximum elevation of obstacles in the road area to be detected; divide the maximum elevation of obstacles according to a preset accuracy threshold to determine the total number of spatial cubes contained in the spatial voxels of each grid; determine the real physical space size corresponding to the spatial cube of the spatial voxels of each grid based on the maximum elevation of obstacles, the total number of spatial cubes and the preset plane size of spatial cubes.
在本实施例中,如图4所示,图4的左边为空间体素示意图,空间体素设置了2t个空间立方体,右边为路面纵剖面,为空间体素所在路面与自动驾驶车辆坐标系也即自车坐标系水平面高程差。当路面水平或自车坐标系与路面之间的俯仰角为0时候,为0。但由于路面自身可能存在倾斜或者当车辆通过低矮凸起障碍物或坑洼区域,自车坐标系与地面存在一定俯仰角时候,为了更为精确获取低矮凸起障碍物或坑洼区域高程,取值如图4右图所示,即取空间体素所在路面与自车坐标系水平面高程差。In this embodiment, as shown in FIG. 4 , the left side of FIG. 4 is a schematic diagram of a space voxel, in which 2t space cubes are set for the space voxel, and the right side is a longitudinal section of the road surface. is the elevation difference between the road surface where the spatial voxel is located and the coordinate system of the autonomous driving vehicle, that is, the horizontal plane of the vehicle coordinate system. When the road surface is horizontal or the pitch angle between the vehicle coordinate system and the road surface is 0, 0. However, since the road surface itself may be inclined or when the vehicle passes through a low raised obstacle or a pothole area, there is a certain pitch angle between the vehicle coordinate system and the ground. In order to more accurately obtain the elevation of the low raised obstacle or pothole area, The values are shown in the right figure of Figure 4, that is, Take the elevation difference between the road surface where the spatial voxel is located and the horizontal plane of the vehicle coordinate system.
其中,空间体素所包含的空间立方体的个数由待识别的障碍物如低矮凸起障碍 物/坑洼区域的最大高程进行确定,示例性的,可根据障碍物最大高程的划分段数的倍数作 为空间立方体总个数,根据障碍物最大高程和空间立方体总个数的比值作为空间立方体对 应的真实物理空间高度;根据真实物理空间高度和栅格尺寸确定各栅格的空间体素的空间 立方体对应的真实物理空间尺寸。举例来说,空间体素可包括个空间立方体,根据 待识别的低矮凸起障碍物/坑洼区域最大高程进行设定,例如待识别的低矮凸起障碍物/坑 洼区域最大高程为0.15米,为了精确识别每个障碍物如低矮凸起障碍物的高程信息,可将 最大高程也即15厘米按照预设精度阈值进行划分,预设精度阈值例如可选取1cm,那么15厘 米按照1cm进行划分,相应的,空间体素所包含的空间立方体总数即为2×15= 30。示例性的,为了便于数据处理,空间立方体的长宽对应空间大小与单位栅格的长宽设置 一致,如可设置为0.01米×0.01米,相应的,空间立方体高度对应真实空间大小为米,因此每个立体框对应真实空间大小为0.01×0.01×0.005,单位米。 The number of space cubes contained in the space voxel is determined by the maximum elevation of the obstacle to be identified, such as a low raised obstacle/pothole area. For example, the total number of space cubes can be determined by the multiple of the number of segments of the maximum elevation of the obstacle, and the real physical space height corresponding to the space cube can be determined by the ratio of the maximum elevation of the obstacle to the total number of space cubes. The real physical space size corresponding to the space cube of the space voxel of each grid is determined according to the real physical space height and the grid size. For example, the space voxel may include A space cube, The maximum elevation of the low raised obstacle/pothole area to be identified is set. For example, the maximum elevation of the low raised obstacle/pothole area to be identified is 0.15 meters. In order to accurately identify the elevation information of each obstacle such as a low raised obstacle, the maximum elevation, that is, 15 centimeters, can be divided according to the preset accuracy threshold. The preset accuracy threshold can be selected as 1 cm, so 15 centimeters is divided into 1 cm, and the corresponding , the total number of space cubes contained in the space voxel is 2×15=30. For example, in order to facilitate data processing, the length and width of the space cube correspond to the same space size as the length and width of the unit grid, such as 0.01 meters × 0.01 meters. Correspondingly, the height of the space cube corresponds to the real space size Meters, so the real space size of each stereo frame is 0.01×0.01×0.005, in meters.
由上可知,本实施例根据车辆的激光雷达参数对栅格区域的参数进行设置,有利于提升道路障碍物的识别精度。通过目标物的最大高程设置空间体素参数,由于空间体素用于计算障碍物高程信息的,可以进一步提高障碍物识别精度。As can be seen from the above, this embodiment sets the parameters of the grid area according to the laser radar parameters of the vehicle, which is conducive to improving the recognition accuracy of road obstacles. The spatial voxel parameters are set by the maximum elevation of the target object. Since the spatial voxels are used to calculate the elevation information of the obstacle, the obstacle recognition accuracy can be further improved.
上述实施例对如何获取栅格高程信息并不做任何限定,基于上述实施例,本发明还给出一种示例性的实现方式,可包括下述内容:The above embodiment does not limit how to obtain grid elevation information. Based on the above embodiment, the present invention also provides an exemplary implementation method, which may include the following contents:
根据待检测路面区域与车辆坐标系的水平面的倾斜数据和各栅格的位置信息,确定鸟瞰视图栅格区域的各栅格的空间体素所在路面与车辆坐标系的水平面高程差。According to the inclination data of the horizontal plane of the road area to be detected and the vehicle coordinate system and the position information of each grid, the elevation difference between the road surface where the spatial voxels of each grid in the bird's-eye view grid area are located and the horizontal plane of the vehicle coordinate system is determined.
示例性的,根据各栅格对应路面的倾斜程度,将待检测路面区域划分为多类路面 区域,同一类路面区域的倾斜程度满足预设相同相似倾斜条件;根据各路面区域与车辆坐 标系的水平面的倾斜数据和各栅格的位置信息,确定鸟瞰视图栅格区域的各栅格的空间体 素所在路面与车辆坐标系的水平面高程差。其中,对各栅格内的点云数据进行平面拟合处 理,得到对应的三维空间平面方程;将各栅格的三维空间平面方程的法向量与车辆坐标系 的水平面的法向量之间的夹角,作为对应栅格的路面区域与车辆坐标系的水平面之间的平 面夹角。如图5所示,根据激光雷达安装位置及车辆结果参数确定激光雷达坐标系原点与路 面之间的坐标系垂直距离H,并根据坐标系垂直距离H和平面夹角,确定车辆坐标系的水 平方向的坐标轴与各栅格对应路面的交叉点的位置信息。路面区域划分过程可包括:根 据各栅格内部的点云数据的三维位置信息,拟合得到相应栅格区域的三维空间平面方程; 基于各栅格的三维空间平面方程,确定各栅格内的平均法向量;根据各栅格的平均法向量 的聚类结果,通过鸟瞰视图栅格区域与待检测路面区域之间的空间位置对应关系对待检测 路面区域进行分割,得到多类路面区域。对各栅格,确定车辆坐标系的水平轴与当前栅格对 应路面的交叉点在水平轴上的位置信息;基于平面夹角,根据各栅格在水平轴的位置信息 与对应交叉点在水平轴的位置信息的位置,计算鸟瞰视图栅格区域的各栅格的空间体素所 在路面与车辆坐标系的水平面高程差。作为一种高效的计算方式,可预先在本地存储高程 计算关系式,通过调用高程计算关系式计算该段路面对应的每个栅格(i、j)对应空间体素 的高程信息,高程计算关系式可表示为: Exemplarily, the road area to be detected is divided into multiple types of road areas according to the inclination of the road surface corresponding to each grid, and the inclination of the same type of road area meets the preset same similar inclination conditions; based on the inclination data of the horizontal plane of each road area and the vehicle coordinate system and the position information of each grid, the elevation difference between the road surface where the spatial voxels of each grid in the bird's-eye view grid area are located and the horizontal plane of the vehicle coordinate system is determined. Among them, the point cloud data in each grid is plane fitting processed to obtain the corresponding three-dimensional space plane equation; the angle between the normal vector of the three-dimensional space plane equation of each grid and the normal vector of the horizontal plane of the vehicle coordinate system is used as the plane angle between the road area of the corresponding grid and the horizontal plane of the vehicle coordinate system. As shown in Figure 5, the coordinate system vertical distance H between the origin of the laser radar coordinate system and the road surface is determined according to the installation position of the laser radar and the vehicle result parameters, and the coordinate system vertical distance H and the plane angle are used. , determine the position information of the intersection between the horizontal coordinate axis of the vehicle coordinate system and the road surface corresponding to each grid . The road area division process may include: fitting the three-dimensional space plane equation of the corresponding grid area according to the three-dimensional position information of the point cloud data inside each grid; determining the average normal vector in each grid based on the three-dimensional space plane equation of each grid; segmenting the road area to be detected according to the spatial position correspondence between the bird's-eye view grid area and the road area to be detected according to the clustering result of the average normal vector of each grid, and obtaining multiple types of road areas. For each grid, determine the position information of the intersection of the horizontal axis of the vehicle coordinate system and the road surface corresponding to the current grid on the horizontal axis; based on the plane angle, calculate the elevation difference between the road surface where the spatial voxels of each grid in the bird's-eye view grid area are located and the horizontal plane of the vehicle coordinate system according to the position information of each grid on the horizontal axis and the position information of the corresponding intersection on the horizontal axis. As an efficient calculation method, the elevation calculation relationship can be stored locally in advance, and the elevation information of the spatial voxels corresponding to each grid (i, j) corresponding to the section of the road surface can be calculated by calling the elevation calculation relationship. , the elevation calculation formula can be expressed as:
; ;
其中,i、j表示鸟瞰视图栅格区域中位于第i行第j列的栅格,如(2,2)表示位于图2中的栅格区域中第二行第二列的栅格。Here, i and j represent the grid located in the i-th row and j-th column in the bird's-eye view grid area. For example, (2, 2) represents the grid located in the second row and second column in the grid area in FIG. 2 .
由上可知,本实施例将路面按照倾斜角度进行划分,并通过简单的方式确定路面倾斜角度,有效提高障碍物的高程信息的确定精度。As can be seen from the above, this embodiment divides the road surface according to the inclination angle, and determines the road surface inclination angle in a simple manner, thereby effectively improving the determination accuracy of the elevation information of the obstacle.
在上述实施例中,对于如何执行步骤S101并不做限定,本实施例中给出获取栅格化处理的待检测路面区域的点云数据和图像数据在目标视角下的特征的一种示例性的实现方式,可包括下述内容:In the above embodiment, there is no limitation on how to perform step S101. In this embodiment, an exemplary implementation method for obtaining the features of the point cloud data and image data of the road surface area to be detected in the target perspective after rasterization is provided, which may include the following contents:
当检测到车辆行驶前方区域存在障碍物,则在车辆前行过程监视障碍物位置;当障碍物进入待检测路面区域时,将位于当前车辆前方的待检测路面区域进行栅格化处理,生成包括多个栅格、且各栅格在垂直于路面水平方向上具有空间体素的鸟瞰视图栅格区域;获取鸟瞰视图栅格区域的点云数据和图像数据,并将点云数据和图像数据转换至鸟瞰视角,得到相应视角的点云特征和图像特征。When an obstacle is detected in the area ahead of the vehicle, the position of the obstacle is monitored during the vehicle's forward movement; when an obstacle enters the road surface area to be detected, the road surface area to be detected in front of the current vehicle is rasterized to generate a bird's-eye view grid area including multiple grids, each grid having spatial voxels in a horizontal direction perpendicular to the road surface; point cloud data and image data of the bird's-eye view grid area are obtained, and the point cloud data and image data are converted to a bird's-eye view perspective to obtain point cloud features and image features of the corresponding perspective.
在本发明中,考虑到现有主流激光雷达水平/垂直角分辨率优于0.06°,激光雷达空间感知能力随着目标距离越近,精度越高。按角分辨率按0.06°计算,在距离雷达50米处,最小空间感知大小为5.2厘米;在距离雷达20米处最小空间感知能力为2.0厘米;在距离雷达10米处最小空间感知能力为1厘米;在距离雷达5米处最小空间感知能力为0.5厘米。车辆在行驶前方若存在障碍物,如低矮凸起障碍物和坑洼区域,随着车辆在行驶过程中会逐步接近车辆,传感器采集的点云数据信息会更加丰富,由于近场校验较为复杂,需要较多计算资源,所以本步骤可先通过远场区域进行初始识别,若远场区域没有发现疑似低矮凸起障碍物/坑洼区域,则不必启动近场校验计算流程,从而可以有效降低不必要的计算负担。基于此,本步骤按照车辆与预先设置的距离阈值之间的距离对车辆行驶前方进行划分,将其划分为车辆行驶前方区域和待检测路面区域,待检测路面区域在沿车辆行驶方向上与车辆的最大距离小于等于车辆行驶前方区域与车辆间的最小距离。其中,距离阈值可根据当前激光雷达线数或分辨率来预先设置,线数或分辨率越低,距离阈值越小。也即车辆行驶前方区域为远场区域,待检测路面区域为近场区域,以距离阈值为20m为例,车辆前20 m后区域定义为车辆行驶前方区域,车辆行驶前方区域的尺寸为:80 m×20 m,车辆前20m内的近场区域为待检测路面区域,待检测路面区域的尺寸为20 m×20 m。In the present invention, considering that the horizontal/vertical angular resolution of the existing mainstream laser radar is better than 0.06°, the laser radar spatial perception capability increases with the closer the target distance. Calculated at an angular resolution of 0.06°, the minimum spatial perception size is 5.2 cm at a distance of 50 meters from the radar; the minimum spatial perception capability is 2.0 cm at a distance of 20 meters from the radar; the minimum spatial perception capability is 1 cm at a distance of 10 meters from the radar; and the minimum spatial perception capability is 0.5 cm at a distance of 5 meters from the radar. If there are obstacles in front of the vehicle, such as low raised obstacles and potholes, as the vehicle gradually approaches the vehicle during driving, the point cloud data information collected by the sensor will be richer. Since the near-field verification is more complicated and requires more computing resources, this step can first perform initial identification through the far-field area. If no suspected low raised obstacles/potholes are found in the far-field area, it is not necessary to start the near-field verification calculation process, thereby effectively reducing unnecessary computing burdens. Based on this, this step divides the area in front of the vehicle according to the distance between the vehicle and the preset distance threshold, and divides it into the area in front of the vehicle and the road area to be detected. The maximum distance between the road area to be detected and the vehicle in the direction of vehicle travel is less than or equal to the minimum distance between the area in front of the vehicle and the vehicle. Among them, the distance threshold can be preset according to the current number of laser radar lines or resolution. The lower the number of lines or resolution, the smaller the distance threshold. That is, the area in front of the vehicle is the far-field area, and the road area to be detected is the near-field area. Taking the distance threshold of 20m as an example, the area 20m in front of the vehicle is defined as the area in front of the vehicle. The size of the area in front of the vehicle is: 80m×20m, and the near-field area within 20m in front of the vehicle is the road area to be detected. The size of the road area to be detected is 20m×20m.
其中,本实施例以目标视角为bev(Bird's Eye View,鸟瞰图)视角,对于激光雷达的点云数据,可利用任何一种点云目标检测算法,如Pointpillars(算法名称)、Second(算法名称)、Centerpoin(算法名称)的特征抽取模块和Neck(模块名称)模块来获取点云在Bev视角下的特征,同样的,网络配置采用图2所示的bev栅格和空间体素,特征抽取模块用于提取图像的基本特征,生成不同尺度的特征图,Neck模块通过融合来自不同层次的特征,帮助网络更好地检测不同尺度的目标,还可提高特征表示能力,特别是对于小目标和复杂场景的检测。此外,Neck模块通常会进行特征图的下采样和上采样操作,以便更好地捕捉不同尺度的信息。对于图像数据,可先利用常用图像特征骨干网络,如resnet50(网络模型名称),抽取图像二维特征,然后基于相机、激光雷达等标定参数,利用LSS算法或交叉注意力方法,将图像二维特征转换为图像在bev的特征,同样的,网络配置采用图2所示的bev栅格和空间体素。Among them, this embodiment takes the target perspective as the bev (Bird's Eye View) perspective. For the point cloud data of the laser radar, any point cloud target detection algorithm, such as Pointpillars (algorithm name), Second (algorithm name), Centerpoin (algorithm name) feature extraction module and Neck (module name) module can be used to obtain the features of the point cloud under the Bev perspective. Similarly, the network configuration uses the bev grid and spatial voxel shown in Figure 2. The feature extraction module is used to extract the basic features of the image and generate feature maps of different scales. The Neck module helps the network to better detect targets of different scales by fusing features from different levels, and can also improve the feature representation capability, especially for the detection of small targets and complex scenes. In addition, the Neck module usually performs downsampling and upsampling operations on the feature map to better capture information of different scales. For image data, we can first use the commonly used image feature backbone network, such as resnet50 (network model name), to extract the two-dimensional features of the image, and then use the LSS algorithm or cross-attention method based on the calibration parameters of the camera, lidar, etc. to convert the two-dimensional features of the image into the features of the image in bev. Similarly, the network configuration uses the bev grid and spatial voxels shown in Figure 2.
由上可知,本实施例在远场仅仅进行疑似低矮障碍物/坑洼区域识别,并获取每个疑似障碍物在全局坐标系(也即世界坐标系)下坐标,同时设定ID信息用于后续连续判读。然后,当低矮障碍物/坑洼区域逐渐进入近场区域时,识别每个低矮障碍物/坑洼区域,并获得其三维形貌数据。当在远场发现疑似低矮障碍物/坑洼区域,则开始近场校验工作。不仅有效降低不必要的计算负担,加快三维形貌数据确定效率。而且通过多重校验,还能提高障碍物识别精度。As can be seen from the above, this embodiment only identifies suspected low obstacles/potholes in the far field, obtains the coordinates of each suspected obstacle in the global coordinate system (that is, the world coordinate system), and sets the ID information for subsequent continuous interpretation. Then, when the low obstacle/pothole area gradually enters the near field area, each low obstacle/pothole area is identified and its three-dimensional shape data is obtained. When a suspected low obstacle/pothole area is found in the far field, the near field verification work begins. It not only effectively reduces unnecessary computing burden and speeds up the efficiency of determining three-dimensional shape data. Moreover, through multiple verifications, the obstacle recognition accuracy can also be improved.
上述实施例对如何得到初始三维形貌信息并不做任何限定,基于上述实施例,本发明还给出了通过对融合特征得到障碍物三维形貌信息的一种示例性的实现方式,可包括下述内容:The above embodiment does not limit how to obtain the initial three-dimensional shape information. Based on the above embodiment, the present invention also provides an exemplary implementation method for obtaining the three-dimensional shape information of the obstacle by fusing features, which may include the following contents:
预先构建并训练三维目标检测网络模型;将融合特征输入至三维目标检测网络模型,得到障碍物的位置信息、类别信息和实例分割结果;基于实例分割结果和障碍物区域内点云数据,通过三维曲面拟合结果确定障碍物的几何形态描述数据;将障碍物的位置信息、类别信息和几何形态描述数据,作为障碍物的初始三维形貌信息。Pre-build and train a 3D target detection network model; input the fused features into the 3D target detection network model to obtain the location information, category information and instance segmentation results of the obstacle; based on the instance segmentation results and the point cloud data in the obstacle area, determine the geometric description data of the obstacle through the 3D surface fitting results; use the location information, category information and geometric description data of the obstacle as the initial 3D shape information of the obstacle.
在本实施例中,三维目标检测网络模型可基于任何一种深度学习的三维目标检测算法训练得到。其可包括检测任务头、分类任务头和实例分割任务头,检测任务头用于识别障碍物,并输出障碍物位置信息;分类任务头用于识别障碍物类别信息;实例分割任务头用于对障碍物在目标平面上进行分割,得到实例分割结果。利用解码器-编码器或解码器机制或者直接卷积进行卷积,将目标视角下的点云特征与图像特征进行融合。对于融合特征,先通过骨干网络和neck模块抽取特征,然后利用检测任务头获取低矮障碍物/坑洼区域位置信息;利用分类任务头获取障碍物类别信息,如低矮障碍物、坑、裂缝、路面凸起等。最后利用实例分割任务头获取障碍物如低矮障碍物、坑洼区域在水平面上的分割结果,基于分割结果和其内部点云数据,通过三维曲面拟合方法,便可以得到障碍物的几何形态描述数据,如图8所示,几何形态描述数据可为长、宽、与车道线夹角的角度,图8为俯视图,为车辆坐标系x坐标轴,为车辆坐标系y坐标轴。In this embodiment, the three-dimensional target detection network model can be trained based on any deep learning three-dimensional target detection algorithm. It may include a detection task head, a classification task head and an instance segmentation task head. The detection task head is used to identify obstacles and output obstacle location information; the classification task head is used to identify obstacle category information; the instance segmentation task head is used to segment obstacles on the target plane to obtain instance segmentation results. The point cloud features and image features under the target perspective are fused by convolution using a decoder-encoder or decoder mechanism or direct convolution. For fused features, features are first extracted through the backbone network and the neck module, and then the detection task head is used to obtain the location information of low obstacles/pothole areas; the classification task head is used to obtain obstacle category information, such as low obstacles, pits, cracks, road bumps, etc. Finally, the instance segmentation task head is used to obtain the segmentation results of obstacles such as low obstacles and pothole areas on the horizontal plane. Based on the segmentation results and its internal point cloud data, the geometric morphology description data of the obstacle can be obtained through a three-dimensional surface fitting method. As shown in Figure 8, the geometric morphology description data can be long ,Width , the angle between the lane line and the , Figure 8 is a top view, is the x-axis of the vehicle coordinate system, is the y-axis of the vehicle coordinate system.
由上可知,本实施例借助三维目标检测网络模型优异的学习能力来计算障碍物的三维形貌数据,有利用提高障碍物三维形貌数据的确定精度。As can be seen from the above, this embodiment uses the excellent learning ability of the three-dimensional target detection network model to calculate the three-dimensional shape data of the obstacle, which is useful for improving the determination accuracy of the three-dimensional shape data of the obstacle.
在计算机视觉领域,为了解决单目深度估计存在的问题,相关技术通过一个网络结构预测高程信息来替代单目深度估计,诸如BEVheight(基于路边场景设计的具有高度信息的视觉三维对象检测框架)、RoadBEV(鸟瞰视图中的道路表面重建),但是这种方法的预测高程的误差可以达到2厘米左右,而该误差导致低矮凸起障碍物及坑洼区域无法实现精确判断,究其原因,标注数据的精度,特别是高程信息标注精度,难以满足,像素级的高程信息的精确标注非常费时,且效果难以把控。因此,本实施例提供了一种通过稠密点云引导的高程信息精确获取方法,可包括下述内容:In the field of computer vision, in order to solve the problems existing in monocular depth estimation, related technologies use a network structure to predict elevation information to replace monocular depth estimation, such as BEVheight (a visual three-dimensional object detection framework with height information designed based on roadside scenes) and RoadBEV (reconstruction of road surface in bird's-eye view). However, the error in predicted elevation of this method can reach about 2 cm, and this error makes it impossible to accurately judge low raised obstacles and potholes. The reason is that the accuracy of the annotation data, especially the accuracy of elevation information annotation, is difficult to meet. Accurate annotation of pixel-level elevation information is very time-consuming and the effect is difficult to control. Therefore, this embodiment provides a method for accurately obtaining elevation information guided by dense point clouds, which may include the following:
获取栅格化的待检测路面区域的各栅格的空间体素及对应的权重初始值;获取当前帧点云数据对应的栅格的高程信息,以作为新高程信息,并利用新高程信息更新相应位置的空间体素的权重值。The spatial voxels of each grid in the rasterized road surface area to be detected and the corresponding initial weight values are obtained; the elevation information of the grid corresponding to the current frame point cloud data is obtained as the new elevation information, and the weight values of the spatial voxels at the corresponding positions are updated using the new elevation information.
在本实施例中,如图6所示,待检测路面区域存在多种区域,也即栅格区域类型,其中,一个区域内不同栅格(i,j)的权重分布因其高程的不同而不同,高程是该区域内每个栅格(i,j)内的高程信息,每个栅格对应的真实路面的高低可能不同,所以每个栅格的权重分布也会不同。栅格区域类型包括但并不限制于上一帧识别的低矮障碍物/坑洼区域在当前时刻的栅格区域内的位置、当前帧识别的低矮障碍物/坑洼区域在当前时刻的栅格区域内位置以及远场识别的低矮障碍物/坑洼区域在当前时刻的栅格区域内的位置。为了进一步提高高程信息的计算精度,本实施例根据不同栅格区域类型确定不同的权重,示例性的,可先获取栅格化的待检测路面区域中各栅格区域类型;当当前栅格属于上一帧识别的第一类障碍物区域,若当前栅格不属于雷达点云投影区域,则基于第一类障碍物区域的各栅格的空间体素坐标位置、高程信息和当前栅格的坐标位置,确定当前栅格的空间体素的权重,若当前栅格属于雷达点投影区域,则基于第一类障碍物区域的各栅格的空间体素坐标位置、高程信息、雷达点投影区域的高程信息及当前栅格的坐标位置,确定当前栅格的空间体素的权重;当当前栅格属于当前帧识别的第二类障碍物区域,若当前栅格不属于雷达点云投影区域,则基于第二类障碍物区域的各栅格的空间体素坐标位置确定当前栅格的空间体素的权重,若当前栅格属于雷达点投影区域,则基于第二类障碍物区域的各栅格的空间体素坐标位置及雷达点投影区域的高程信息,确定当前栅格的空间体素的权重。In this embodiment, as shown in FIG6 , there are multiple regions in the road area to be detected, namely, grid region types, wherein the weight distribution of different grids (i, j) in one region is different due to their elevation. The height varies with the is the elevation information of each grid (i, j) in the area. The height of the real road surface corresponding to each grid may be different, so the weight distribution of each grid will also be different. The grid area type includes but is not limited to the position of the low obstacle/pothole area identified in the previous frame in the grid area at the current moment, the position of the low obstacle/pothole area identified in the current frame in the grid area at the current moment, and the position of the low obstacle/pothole area identified in the far field in the grid area at the current moment. In order to further improve the calculation accuracy of the elevation information, this embodiment determines different weights according to different grid area types. For example, each grid area type in the rasterized road surface area to be detected can be obtained first; when the current grid belongs to the first type of obstacle area identified in the previous frame, if the current grid does not belong to the radar point cloud projection area, then based on the spatial voxel coordinate position, elevation information and coordinate position of each grid in the first type of obstacle area, the weight of the spatial voxel of the current grid is determined; if the current grid belongs to the radar point projection area, then based on the spatial voxel coordinate position of each grid in the first type of obstacle area, the elevation information and the coordinate position of the current grid, the weight of the spatial voxel of the current grid is determined. If the current grid belongs to the radar point projection area, then based on the spatial voxel coordinate position of each grid in the first type of obstacle area The weight of the spatial voxel of the current grid is determined based on the voxel coordinate position, elevation information, elevation information of the radar point projection area and the coordinate position of the current grid; when the current grid belongs to the second type of obstacle area identified by the current frame, if the current grid does not belong to the radar point cloud projection area, the weight of the spatial voxel of the current grid is determined based on the spatial voxel coordinate position of each grid in the second type of obstacle area; if the current grid belongs to the radar point projection area, the weight of the spatial voxel of the current grid is determined based on the spatial voxel coordinate position of each grid in the second type of obstacle area and the elevation information of the radar point projection area.
举例来说,如图7所示,空间体素的坐标范围为-30~30,针对不同的栅格区域,空间体素的权重计算如下:For example, as shown in Figure 7, the coordinate range of the spatial voxel is -30~30. For different grid areas, the weight of the spatial voxel is calculated as follows:
对于上一帧识别的低矮障碍物/坑洼区域在当前时刻的鸟瞰图栅格区域的位置,其对应空间体素分为两类,第一类栅格区域是对应的鸟瞰图栅格上没有雷达点云投影,第二类栅格区域为对应的鸟瞰图栅格上有雷达点云投影。对于时刻,计算所得其内部包含的所有栅格的高程信息,雷达点云投影的高程信息为,第一类栅格区域的权重计算可通过下述关系式(1)实现,第二类栅格区域的权重计算可通过下述关系式(2)实现:For the position of the low obstacle/pothole area identified in the previous frame in the bird's-eye view grid area at the current moment, the corresponding spatial voxels are divided into two categories: the first category of grid areas is the corresponding bird's-eye view grid without radar point cloud projection, and the second category of grid areas is the corresponding bird's-eye view grid with radar point cloud projection. At this moment, the elevation information of all the grids contained in it is calculated , the elevation information of the radar point cloud projection is , the weight calculation of the first type of grid area can be realized by the following relationship (1), and the weight calculation of the second type of grid area can be realized by the following relationship (2):
;(1) ; (1)
。(2) . (2)
对于当前帧识别的低矮障碍物/坑洼区域在当前时刻的鸟瞰图栅格区域的位置,其对应空间体素同样分为两类。第三类栅格区域为其对应的鸟瞰图栅格上没有雷达点云投影,第四类栅格区域为其对应的鸟瞰图栅格上有雷达点云投影,该空间体素对应的鸟瞰图栅格中也有雷达点云投影,相应的,雷达点云投影的高程信息为,第三类栅格区域的权重计算可通过下述关系式(3)实现,第四类栅格区域的权重计算可通过下述关系式(4)实现:For the position of the low obstacle/pothole area identified in the current frame in the bird's-eye view grid area at the current moment, the corresponding spatial voxels are also divided into two categories. The third type of grid area has no radar point cloud projection on its corresponding bird's-eye view grid, and the fourth type of grid area has a radar point cloud projection on its corresponding bird's-eye view grid. The spatial voxel also has a radar point cloud projection in the bird's-eye view grid. Correspondingly, the elevation information of the radar point cloud projection is , the weight calculation of the third type of grid area can be realized by the following relationship (3), and the weight calculation of the fourth type of grid area can be realized by the following relationship (4):
;(3) ; (3)
。(4) . (4)
对于远场识别的低矮障碍物/坑洼区域在当前时刻的鸟瞰图栅格区域的位置,其对应空间体素同样分为两类。第五类栅格区域为其对应的鸟瞰图栅格上没有雷达点云投影,第六类栅格区域为其对应的鸟瞰图栅格上有雷达点云投影,该空间体素对应的鸟瞰图栅格中也有雷达点云投影,相应的,雷达点云投影的高程信息为,第五类栅格区域的权重计算可通过下述关系式(5)实现,第六类栅格区域的权重计算可通过下述关系式(6)实现:For the position of the low obstacle/pothole area identified in the far field in the bird's-eye view grid area at the current moment, the corresponding spatial voxels are also divided into two categories. The fifth type of grid area has no radar point cloud projection on its corresponding bird's-eye view grid, and the sixth type of grid area has a radar point cloud projection on its corresponding bird's-eye view grid. The spatial voxel also has a radar point cloud projection in the bird's-eye view grid. Correspondingly, the elevation information of the radar point cloud projection is , the weight calculation of the fifth type of grid area can be realized by the following relationship (5), and the weight calculation of the sixth type of grid area can be realized by the following relationship (6):
;(5) ; (5)
。(6) . (6)
示例性的,作为一种简单的权重设置方式,可基于一维高斯函数关系式,根据预设高斯曲线控制参数值、由不同栅格区域类型确定的中心位置坐标、栅格空间体素位置信息作为变量,生成权重计算关系式,以通过调用权重计算关系式确定各栅格空间体素的权重值。举例来说,预先设置高斯曲线控制参数可为1.0,可为1.0,空间体素的权重相应的可采用一维高斯曲线表示,为曲线中心位置,g(x)表示高斯曲线,在本实施例中,x表示空间体素的坐标值。For example, as a simple weight setting method, a one-dimensional Gaussian function relationship can be used to generate a weight calculation relationship based on the preset Gaussian curve control parameter value, the center position coordinates determined by different grid area types, and the grid space voxel position information as variables, so as to determine the weight value of each grid space voxel by calling the weight calculation relationship. Can be 1.0, Can be 1.0, the weight of the spatial voxel can be correspondingly a one-dimensional Gaussian curve express, is the center position of the curve, g(x) represents the Gaussian curve, and in this embodiment, x represents the coordinate value of the spatial voxel.
当确定了不同栅格区域的权重计算方式,通过高程分类网络得到障碍物区域内栅格的高程信息。其中,高程分类网络为预先训练高程分类网络,用于将深度估计问题转换为分类问题,且高程分类网络基于障碍物类型、高程真值编码信息、障碍物区域掩码信息、障碍物区域栅格的点云特征及图像特征、障碍物区域栅格的空间体素权重确定损失函数。When the weight calculation method of different grid areas is determined, the elevation information of the grid in the obstacle area is obtained through the elevation classification network. Among them, the elevation classification network is a pre-trained elevation classification network, which is used to convert the depth estimation problem into a classification problem, and the elevation classification network determines the loss function based on the obstacle type, elevation true value encoding information, obstacle area mask information, point cloud features and image features of the obstacle area grid, and spatial voxel weights of the obstacle area grid.
在本实施例中,基于LSS方法,将深度估计问题转换为分类问题,通过高程分类网络可以得到待检测路面区域的障碍物内部对应的每个鸟瞰图栅格的高程信息。为了进一步提高高程信息的精度,本实施例还提供了相应的损失函数。作为一种高效的实现方式,可预先存储损失函数,该损失函数基于障碍物类型、高程真值编码信息、障碍物区域掩码信息、障碍物区域栅格的点云特征及图像特征、障碍物区域栅格的空间体素权重来确定,高程真值编码信息采用独热编码进行处理,其可采用如下数学形式来表示:In this embodiment, based on the LSS method, the depth estimation problem is converted into a classification problem, and the elevation information of each bird's-eye view grid corresponding to the obstacle inside the road area to be detected can be obtained through the elevation classification network. In order to further improve the accuracy of the elevation information, this embodiment also provides a corresponding loss function. As an efficient implementation method, the loss function can be pre-stored. The loss function is determined based on the obstacle type, elevation true value coding information, obstacle area mask information, point cloud features and image features of the obstacle area grid, and the spatial voxel weight of the obstacle area grid. The elevation true value coding information is processed using one-hot coding, which can be expressed in the following mathematical form:
; ;
其中,,为识别的障碍物的种类数,为one-hot(独热编码)的GT mask(表示高程真值编码信息),其为利用预先标注的高程真值生成的掩码,为障碍物的掩码信息,为第个鸟瞰图栅格获得点云特征和图像特征,为权重。损失函数基于深度学习领域的公知常识,由于底数不影响模型训练过程,所以通常忽略不写。in, , is the number of types of obstacles identified, is a one-hot GT mask (representing elevation truth encoding information), which is a mask generated using pre-labeled elevation truth values. is the mask information of the obstacle, For the A bird's-eye view grid is used to obtain point cloud features and image features. The loss function is based on common knowledge in the field of deep learning. Since the base does not affect the model training process, it is usually ignored.
当将待检测路面区域栅格化之后,举例来说,若栅格数量为2000×2000,且体素设定为60个高程估计,如果逐点计算,计算量巨大。而通常障碍物只会占据部分栅格,对于非低矮障碍物及坑洼区域,并不需要计算其高程信息。为了降低计算量,本发明还可先确定障碍物所包含的栅格,在对栅格进行高程信息的计算,可包括下述内容:After the road area to be detected is rasterized, for example, if the number of grids is 2000×2000 and the voxel is set to 60 elevation estimates, if the calculation is performed point by point, the amount of calculation is huge. Usually, obstacles only occupy part of the grid, and for non-low obstacles and potholes, it is not necessary to calculate their elevation information. In order to reduce the amount of calculation, the present invention can also first determine the grids contained in the obstacle, and the calculation of the elevation information of the grid can include the following:
根据上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息、当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,确定当前帧障碍物区域包含的栅格;获取当前帧障碍物区域包含栅格的高程信息,并对当前帧障碍物区域包含栅格的空间体素权重进行更新。According to the position information of the obstacle identified in the previous frame at the current moment of the bird's-eye view grid area, and the position information of the obstacle identified in the current frame at the current moment of the bird's-eye view grid area, the grid included in the obstacle area of the current frame is determined; the elevation information of the grid included in the obstacle area of the current frame is obtained, and the spatial voxel weights of the grid included in the obstacle area of the current frame are updated.
本实施例仅仅保留上一帧识别的低矮障碍物/坑洼区域、当前帧识别的低矮障碍物/坑洼区域以及远场识别的进入近场区域的疑似低矮障碍物/坑洼区域。为了进一步有效降低计算量,本实施例还基于障碍物占用栅格生成掩码信息,示例性的,当前帧障碍物区域包含的栅格的确定方法为:根据上一帧和当前帧对应的障碍物识别结果,确定对应障碍物区域内栅格;获取车辆行驶前方区域检测到的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息;车辆行驶前方区域在沿车辆行驶方向上与车辆间的最小距离大于等于待检测路面区域与车辆的最大距离;根据车辆在相邻帧时间段内在局部世界坐标系的位置变化,更新上一帧障碍物区域内栅格的位置信息,以作为上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息;若当前点位置为上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或前方行驶区域识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,则为当前点设置第一预设值,以作为掩码值;若当前点位置不为上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或前方行驶区域识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,则为当前点设置第二预设值,以作为掩码值;根据待检测路面区域的各点的掩码值,生成障碍物区域掩码信息,以用于标识当前帧障碍物区域包含的栅格。This embodiment only retains the low obstacles/potholes identified in the previous frame, the low obstacles/potholes identified in the current frame, and the suspected low obstacles/potholes that enter the near field area identified in the far field. In order to further effectively reduce the amount of calculation, this embodiment also generates mask information based on the grid occupied by the obstacle. Exemplarily, the method for determining the grid contained in the obstacle area of the current frame is: according to the obstacle recognition results corresponding to the previous frame and the current frame, determine the grid in the corresponding obstacle area; obtain the position information of the bird's-eye view grid area of the obstacle detected in the area ahead of the vehicle at the current moment; the minimum distance between the area ahead of the vehicle and the vehicle along the vehicle's driving direction is greater than or equal to the maximum distance between the road surface area to be detected and the vehicle; according to the position change of the vehicle in the local world coordinate system in the adjacent frame time period, update the position information of the grid in the obstacle area of the previous frame as the position information of the bird's-eye view grid area of the obstacle identified in the previous frame at the current moment; if the current point position is the obstacle identified in the previous frame If the position of the current point is the position information of the obstacle in the bird's-eye view grid area at the current moment, or the position information of the obstacle identified in the current frame in the bird's-eye view grid area at the current moment, or the position information of the obstacle identified in the front driving area in the bird's-eye view grid area at the current moment, a first preset value is set for the current point as a mask value; if the current point position is not the position information of the obstacle identified in the previous frame in the bird's-eye view grid area at the current moment, or the position information of the obstacle identified in the current frame in the bird's-eye view grid area at the current moment, or the position information of the obstacle identified in the front driving area in the bird's-eye view grid area at the current moment, a second preset value is set for the current point as a mask value; according to the mask value of each point in the road surface area to be detected, the obstacle area mask information is generated to identify the grid contained in the obstacle area of the current frame.
在本实施例中,掩码信息生成是基于雷达投影点、上一帧识别的低矮障碍物/坑洼区域在当前时刻的栅格区域的位置、当前识别的低矮障碍物/坑洼区域在当前时刻的栅格区域的位置、远场识别的低矮障碍物/坑洼区域在当前时刻的栅格区域的位置自动生成,不需要人工设定。如图6所示,图中栅格中第一类区域为激光雷达点云在路面投影点,即激光雷达线束与地面交点。若同一栅格内有多个激光雷达点云投影点,则z坐标取其平均值;图中第二类区域为上一帧识别低矮障碍物/坑洼区域,随着车辆不断前行,通过后续计算,对已识别区域的高程信息进行逐步更新;由于自车前行,上一帧识别的低矮障碍物/坑洼区域相对于车辆的距离发生变化,需要根据车辆在局部世界坐标系中的位置变化,更新每个上一帧识别低矮障碍物/坑洼区域在栅格区域的相应的位置。举例来说,障碍物在上一时刻也即t-1时刻所包含的栅格位置为。对于当前时刻t,车辆向前直行了0.2米,若一个栅格对应的空间距离为0.01米,那么的坐标需要按照更新后的x=x-车行进距离*栅格对应空间尺寸进行更新,更新后的栅格可表示为,对应自车坐标系x轴,对应自车坐标系y轴。图中第三类区域为远场检测到的疑似障碍物,随着车辆不断前行,逐渐进入栅格区域,需要对其是否为真实的障碍物进行校验,若不是真正障碍物,则这类区域后续不做三维形貌信息的计算,若其为低矮障碍物或坑洼区域,则进一步计算其精确的三维形貌数据。图中第四类区域为当前帧的识别的障碍物区域。In this embodiment, the mask information is generated automatically based on the radar projection point, the position of the low obstacle/pothole area identified in the previous frame at the grid area at the current moment, the position of the low obstacle/pothole area identified in the current moment, and the position of the low obstacle/pothole area identified in the far field at the grid area at the current moment, without manual setting. As shown in Figure 6, the first type of area in the grid in the figure is the projection point of the laser radar point cloud on the road surface, that is, the intersection of the laser radar beam and the ground. If there are multiple laser radar point cloud projection points in the same grid, the z coordinate is taken as the average value; the second type of area in the figure is the low obstacle/pothole area identified in the previous frame. As the vehicle continues to move forward, the elevation information of the identified area is gradually updated through subsequent calculations; due to the forward movement of the vehicle, the distance of the low obstacle/pothole area identified in the previous frame relative to the vehicle changes, and it is necessary to update the corresponding position of each low obstacle/pothole area identified in the previous frame in the grid area according to the change in the position of the vehicle in the local world coordinate system. For example, the grid contained by the obstacle at the last moment, that is, at time t-1 Location is At the current time t, the vehicle has traveled 0.2 meters straight forward. If the spatial distance corresponding to a grid is 0.01 meters, then The coordinates need to be updated according to the updated x=x-vehicle travel distance*grid corresponding space size. The updated grid can be expressed as , Corresponding to the x-axis of the vehicle coordinate system, Corresponding to the y-axis of the vehicle coordinate system. The third type of area in the figure is a suspected obstacle detected in the far field. As the vehicle continues to move forward and gradually enters the grid area, it is necessary to verify whether it is a real obstacle. If it is not a real obstacle, the three-dimensional shape information of this type of area will not be calculated later. If it is a low obstacle or a pothole area, its accurate three-dimensional shape data will be further calculated. The fourth type of area in the figure is the obstacle area identified in the current frame.
示例性的,作为一种更为高效的方式,可调用预先存储的掩码关系式,生成相应的掩码信息,掩码关系式可表示为:For example, as a more efficient method, a pre-stored mask relationship can be called to generate corresponding mask information. , the mask relation can be expressed as:
; ;
其中,i、j表示一个栅格,集合D包括:上一帧识别的障碍物在当前时刻的栅格区域的位置;当前识别的障碍物在当前时刻的栅格区域的位置;远场识别的障碍物在当前时刻的栅格区域的位置。Among them, i and j represent a grid, and the set D includes: the position of the obstacle recognized in the previous frame in the grid area at the current moment; the position of the obstacle recognized currently in the grid area at the current moment; the position of the obstacle recognized in the far field in the grid area at the current moment.
由上可知,本实施例通过利用上一帧结果、多模态特征、远场识别结果生成障碍物掩码信息,不仅能够有效降低计算量,还有利于提升障碍物的三维形貌数据精度。It can be seen from the above that this embodiment generates obstacle mask information by utilizing the previous frame results, multimodal features, and far-field recognition results, which can not only effectively reduce the amount of calculation, but also help improve the accuracy of the three-dimensional shape data of the obstacle.
为了使本领域技术人员更加清楚本发明的技术方案,本发明还提供了一个示例性的实施方式,如图9所示,可包括下述内容:In order to make the technical solution of the present invention more clear to those skilled in the art, the present invention also provides an exemplary implementation, as shown in FIG9 , which may include the following contents:
本实施例可预先构建三维形貌数据生成模型,用户向三维形貌数据生成模型输入下述数据:近场稠密点云或消除畸变后的单帧点云数据、前向相机图像、局部世界坐标系及局部世界点云、上一帧的识别结果、远场检测结果、传感器标定参数。其中,传感器标定参数包括相机与激光雷达、激光雷达与惯性测量单元、惯性测量单元与里程计的标定参数。随着车辆不断前行,通过沿着时间轴,对不同时刻的点云数据进行拼接,即为局部世界点云,通过建立该局部世界点云地图和稠密点云数据得到对应的点云bev特征。其中,对于机械式激光雷达点云数据,在一帧数据获取周期内,由于车辆运动使得每帧点云数据存在运动畸变。畸变校准方法可以采用任何一种方法,比如通过获取自车运动学数据,如假设匀速直线运动,估算100毫秒内车辆运动距离、方向等,基于此,结合每个点的时间戳,对点云xyz坐标进行校准。This embodiment can pre-construct a three-dimensional shape data generation model, and the user inputs the following data into the three-dimensional shape data generation model: near-field dense point cloud or single-frame point cloud data after distortion is eliminated, forward camera image, local world coordinate system and local world point cloud, recognition result of the previous frame, far-field detection result, and sensor calibration parameters. Among them, the sensor calibration parameters include calibration parameters of camera and lidar, lidar and inertial measurement unit, inertial measurement unit and odometer. As the vehicle continues to move forward, the point cloud data at different times are spliced along the time axis to form a local world point cloud. The corresponding point cloud bev features are obtained by establishing the local world point cloud map and dense point cloud data. Among them, for mechanical lidar point cloud data, within a frame of data acquisition cycle, each frame of point cloud data has motion distortion due to vehicle movement. The distortion calibration method can adopt any method, such as obtaining the kinematic data of the vehicle, estimating the vehicle's movement distance and direction within 100 milliseconds, assuming uniform linear motion, and based on this, calibrate the point cloud xyz coordinates in combination with the timestamp of each point.
三维形貌数据生成模型将点云数据转换为目标视角如bev下的点云特征,也可定义为点云bev特征,将图像数据结合传感器标定参数转换为目标视角如bev下的图像特征,也可定义为图像bev特征,然后将点云bev特征和图像bev特征通过编码器进行融合,得到bev下的融合特征,再通过解码器得到障碍物的定位识别结果、二维识别结果和障碍物类别结果。根据二维识别结果和每个空间体素的新的高程信息得到三维识别结果。根据上一帧识别结果、远场检测结果和当前帧识别结果共同生成障碍物掩码信息,通过障碍物掩码信息、三维识别结果和点云bev特征得到待检测路面区域的障碍物识别结果,最终三维形貌数据生成模型输出为:车辆前方障碍物的三维形貌信息、定位信息、类别信息。The 3D shape data generation model converts point cloud data into point cloud features under the target perspective such as bev, which can also be defined as point cloud bev features. The image data is converted into image features under the target perspective such as bev in combination with the sensor calibration parameters, which can also be defined as image bev features. Then the point cloud bev features and the image bev features are fused through the encoder to obtain the fusion features under bev, and then the obstacle positioning recognition results, two-dimensional recognition results and obstacle category results are obtained through the decoder. The three-dimensional recognition result is obtained based on the two-dimensional recognition result and the new elevation information of each spatial voxel. The obstacle mask information is jointly generated based on the recognition result of the previous frame, the far-field detection result and the current frame recognition result. The obstacle recognition result of the road area to be detected is obtained through the obstacle mask information, the three-dimensional recognition result and the point cloud bev feature. The final output of the 3D shape data generation model is: the three-dimensional shape information, positioning information and category information of the obstacle in front of the vehicle.
由上可知,本实施例通过三维形貌数据生成模型的迭代计算可以不断提升高程信息、实例分割信息的精度,实现对待检测路面区域内障碍物的三维形貌信息的精度的不断提高。It can be seen from the above that this embodiment can continuously improve the accuracy of elevation information and instance segmentation information through iterative calculation of the three-dimensional shape data generation model, thereby continuously improving the accuracy of the three-dimensional shape information of obstacles in the road area to be detected.
需要说明的是,本发明中各步骤之间没有严格的先后执行顺序,只要符合逻辑上的顺序,则这些步骤可以同时执行,也可按照某种预设顺序执行,图1和图9只是一种示意方式,并不代表只能是这样的执行顺序。It should be noted that there is no strict order of execution between the steps in the present invention. As long as they comply with the logical order, these steps can be executed simultaneously or in a preset order. Figures 1 and 9 are only a schematic diagram and do not mean that this is the only execution order.
本发明还针对路面障碍物三维形貌信息确定方法提供了相应的装置,进一步使得方法更具有实用性。其中,装置可从功能模块的角度和硬件的角度分别说明。下面对本发明提供的路面障碍物三维形貌信息确定装置进行介绍,该装置用以实现本发明提供的路面障碍物三维形貌信息确定方法,在本实施例中,路面障碍物三维形貌信息确定装置可以包括或被分割成一个或多个程序模块,该一个或多个程序模块被存储在存储介质中,并由一个或多个处理器所执行,已完成实施例一公开的路面障碍物三维形貌信息确定方法。本实施例所称的程序模块是指能够完成特定功能的一系列计算机程序指令段,比程序本身更适合于描述路面障碍物三维形貌信息确定装置在存储介质中的执行过程。以下描述将具体介绍本实施例各程序模块的功能,下文描述的路面障碍物三维形貌信息确定装置与上文描述的路面障碍物三维形貌信息确定方法可相互对应参照。The present invention also provides a corresponding device for the method for determining the three-dimensional shape information of road obstacles, which further makes the method more practical. Among them, the device can be described from the perspective of functional modules and hardware. The following is an introduction to the device for determining the three-dimensional shape information of road obstacles provided by the present invention. The device is used to implement the method for determining the three-dimensional shape information of road obstacles provided by the present invention. In this embodiment, the device for determining the three-dimensional shape information of road obstacles can include or be divided into one or more program modules, which are stored in a storage medium and executed by one or more processors to complete the method for determining the three-dimensional shape information of road obstacles disclosed in the first embodiment. The program module referred to in this embodiment refers to a series of computer program instruction segments that can complete specific functions, which is more suitable for describing the execution process of the device for determining the three-dimensional shape information of road obstacles in the storage medium than the program itself. The following description will specifically introduce the functions of each program module of this embodiment. The device for determining the three-dimensional shape information of road obstacles described below and the method for determining the three-dimensional shape information of road obstacles described above can be referred to each other.
基于功能模块的角度,参见图10,图10为本实施例提供的路面障碍物三维形貌信息确定装置在一种具体实施方式下的结构图,该装置可包括:From the perspective of functional modules, see FIG. 10 , which is a structural diagram of a device for determining three-dimensional shape information of road obstacles provided in this embodiment in a specific implementation manner. The device may include:
多模态特征获取模块101,用于获取栅格化处理的待检测路面区域的点云数据和图像数据在目标视角下的特征;其中,栅格化的待检测路面区域包括多个栅格,各栅格在垂直于路面水平方向上设置空间体素。The multimodal feature acquisition module 101 is used to acquire the features of the point cloud data and image data of the rasterized road area to be detected under the target perspective; wherein the rasterized road area to be detected includes multiple grids, and each grid is provided with spatial voxels in a horizontal direction perpendicular to the road surface.
初始三维形貌信息确定模块102,用于将点云特征和图像特征进行融合,并通过对融合特征进行目标检测处理,得到待检测路面区域内的障碍物的初始三维形貌信息。The initial three-dimensional shape information determination module 102 is used to fuse the point cloud features and the image features, and perform target detection processing on the fused features to obtain the initial three-dimensional shape information of the obstacles in the road area to be detected.
精度更新模块103,用于根据车辆前行过程中不断获取的各帧点云数据对应的栅格新高程信息,更新当前时刻的空间体素权重,并基于更新后的空间体素权重更新障碍物区域内栅格的高程信息和几何形态数据,以对初始三维形貌信息进行不断更新。The precision update module 103 is used to update the spatial voxel weight at the current moment according to the new elevation information of the grid corresponding to each frame of point cloud data continuously acquired during the vehicle's forward movement, and to update the elevation information and geometric data of the grid in the obstacle area based on the updated spatial voxel weight, so as to continuously update the initial three-dimensional shape information.
示例性的,在本实施例的一些实施方式中,上述多模态特征获取模块101还可用于:当检测到车辆行驶前方区域存在障碍物,则在车辆前行过程监视障碍物位置;当障碍物进入待检测路面区域时,将位于当前车辆前方的待检测路面区域进行栅格化处理,生成包括多个栅格、且各栅格在垂直于路面水平方向上具有空间体素的鸟瞰视图栅格区域;获取鸟瞰视图栅格区域的点云数据和图像数据,并将点云数据和图像数据转换至鸟瞰视角,得到相应视角的点云特征和图像特征。其中,待检测路面区域在沿车辆行驶方向上与车辆的最大距离小于等于车辆行驶前方区域与车辆间的最小距离。Exemplarily, in some implementations of this embodiment, the multimodal feature acquisition module 101 can also be used to: when an obstacle is detected in the area ahead of the vehicle, monitor the position of the obstacle during the vehicle's forward movement; when an obstacle enters the road surface area to be detected, rasterize the road surface area to be detected in front of the current vehicle to generate a bird's-eye view grid area including multiple grids, each grid having spatial voxels in a horizontal direction perpendicular to the road surface; obtain point cloud data and image data of the bird's-eye view grid area, and convert the point cloud data and image data to a bird's-eye view perspective to obtain point cloud features and image features of the corresponding perspective. Wherein, the maximum distance between the road surface area to be detected and the vehicle along the vehicle's driving direction is less than or equal to the minimum distance between the area ahead of the vehicle and the vehicle.
示例性的,在本实施例的另一些实施方式中,上述初始三维形貌信息确定模块102还可用于:预先构建并训练三维目标检测网络模型;三维目标检测网络模型包括检测任务头、分类任务头和实例分割任务头,检测任务头用于识别障碍物,并输出障碍物位置信息;分类任务头用于识别障碍物类别信息;实例分割任务头用于对障碍物在目标平面上进行分割,得到实例分割结果;将融合特征输入至三维目标检测网络模型,得到障碍物的位置信息、类别信息和实例分割结果;基于实例分割结果和障碍物区域内点云数据,通过三维曲面拟合结果确定障碍物的几何形态描述数据;将障碍物的位置信息、类别信息和几何形态描述数据,作为障碍物的初始三维形貌信息。Exemplarily, in some other implementations of the present embodiment, the above-mentioned initial three-dimensional shape information determination module 102 can also be used to: pre-build and train a three-dimensional target detection network model; the three-dimensional target detection network model includes a detection task head, a classification task head and an instance segmentation task head, the detection task head is used to identify obstacles and output obstacle position information; the classification task head is used to identify obstacle category information; the instance segmentation task head is used to segment obstacles on the target plane to obtain instance segmentation results; the fusion features are input into the three-dimensional target detection network model to obtain the obstacle position information, category information and instance segmentation results; based on the instance segmentation results and the point cloud data in the obstacle area, the geometric morphology description data of the obstacle is determined through the three-dimensional surface fitting results; the obstacle position information, category information and geometric morphology description data are used as the initial three-dimensional shape information of the obstacle.
示例性的,在本实施例的另一些实施方式中,上述装置还可包括数据转换模块,该模块可用于:基于车辆坐标系与局部世界坐标系之间的变换关系,将待检测路面区域的障碍物的三维形貌信息转换至局部世界坐标系;其中,局部世界坐标系为车辆行驶路面所在的坐标系。Exemplarily, in some other implementations of this embodiment, the above-mentioned device may also include a data conversion module, which can be used to: based on the transformation relationship between the vehicle coordinate system and the local world coordinate system, convert the three-dimensional shape information of obstacles in the road area to be detected into the local world coordinate system; wherein the local world coordinate system is the coordinate system of the road surface on which the vehicle is traveling.
示例性的,在本实施例的另一些实施方式中,上述装置还可包括变换关系确定模块,该模块可用于:获取当前时刻的点云数据和局部地图点云;通过对点云数据和局部地图点云进行点云匹配,确定车辆坐标系与局部世界坐标系之间的变换关系;其中,局部地图点云为沿着时间轴,对不同时刻的点云数据进行拼接后生成的数据。Exemplarily, in some other implementations of the present embodiment, the above-mentioned device may also include a transformation relationship determination module, which can be used to: obtain point cloud data and local map point cloud at the current moment; determine the transformation relationship between the vehicle coordinate system and the local world coordinate system by performing point cloud matching on the point cloud data and the local map point cloud; wherein the local map point cloud is data generated by splicing point cloud data at different moments along the time axis.
示例性的,在本实施例的另一些实施方式中,上述精度更新模块103还可用于:获取栅格化的待检测路面区域的各栅格的空间体素及对应的权重初始值;获取当前帧点云数据对应的栅格的高程信息,以作为新高程信息,并利用新高程信息更新相应位置的空间体素的权重值。Exemplarily, in some other implementations of the present embodiment, the above-mentioned precision update module 103 can also be used to: obtain the spatial voxels of each grid in the rasterized road surface area to be inspected and the corresponding initial weight values; obtain the elevation information of the grid corresponding to the current frame point cloud data as new elevation information, and use the new elevation information to update the weight values of the spatial voxels at the corresponding positions.
示例性的,在本实施例的另一些实施方式中,上述精度更新模块103还可用于:获取栅格化的待检测路面区域中各栅格区域类型;当当前栅格属于上一帧识别的第一类障碍物区域,若当前栅格不属于雷达点云投影区域,则基于第一类障碍物区域的各栅格的空间体素坐标位置、高程信息和当前栅格的坐标位置,确定当前栅格的空间体素的权重,若当前栅格属于雷达点投影区域,则基于第一类障碍物区域的各栅格的空间体素坐标位置、高程信息、雷达点投影区域的高程信息及当前栅格的坐标位置,确定当前栅格的空间体素的权重;当当前栅格属于当前帧识别的第二类障碍物区域,若当前栅格不属于雷达点云投影区域,则基于第二类障碍物区域的各栅格的空间体素坐标位置确定当前栅格的空间体素的权重,若当前栅格属于雷达点投影区域,则基于第二类障碍物区域的各栅格的空间体素坐标位置及雷达点投影区域的高程信息,确定当前栅格的空间体素的权重。Exemplarily, in some other implementations of this embodiment, the precision updating module 103 may also be used to: obtain the type of each grid area in the gridded road surface area to be detected; when the current grid belongs to the first type of obstacle area identified in the previous frame, if the current grid does not belong to the radar point cloud projection area, then based on the spatial voxel coordinate position of each grid in the first type of obstacle area, the elevation information and the coordinate position of the current grid, determine the weight of the spatial voxel of the current grid; if the current grid belongs to the radar point projection area, then based on the spatial voxel coordinate position of each grid in the first type of obstacle area, the elevation information and the coordinate position of the current grid, determine the weight of the spatial voxel of the current grid; if the current grid belongs to the radar point projection area, then based on the spatial voxel coordinate position of each grid in the first type of obstacle area, determine the weight of the spatial voxel of the current grid; The weight of the spatial voxel of the current grid is determined based on the location, elevation information, elevation information of the radar point projection area and the coordinate position of the current grid; when the current grid belongs to the second type of obstacle area identified by the current frame, if the current grid does not belong to the radar point cloud projection area, the weight of the spatial voxel of the current grid is determined based on the spatial voxel coordinate position of each grid in the second type of obstacle area; if the current grid belongs to the radar point projection area, the weight of the spatial voxel of the current grid is determined based on the spatial voxel coordinate position of each grid in the second type of obstacle area and the elevation information of the radar point projection area.
示例性的,在本实施例的另一些实施方式中,上述精度更新模块103还可进一步用于:基于一维高斯函数关系式,根据预设高斯曲线控制参数值、由不同栅格区域类型确定的中心位置坐标、栅格空间体素位置信息作为变量,生成权重计算关系式,以通过调用权重计算关系式确定各栅格空间体素的权重值。Exemplarily, in some other implementations of the present embodiment, the above-mentioned precision update module 103 can be further used to: based on a one-dimensional Gaussian function relationship, generate a weight calculation relationship according to preset Gaussian curve control parameter values, center position coordinates determined by different grid area types, and grid space voxel position information as variables, so as to determine the weight value of each grid space voxel by calling the weight calculation relationship.
示例性的,在本实施例的另一些实施方式中,上述精度更新模块103还可用于:预先训练高程分类网络,高程分类网络用于将深度估计问题转换为分类问题,且高程分类网络基于障碍物类型、高程真值编码信息、障碍物区域掩码信息、障碍物区域栅格的点云特征及图像特征、障碍物区域栅格的空间体素权重确定损失函数;通过高程分类网络得到障碍物区域内栅格的高程信息。Exemplarily, in some other implementations of the present embodiment, the above-mentioned precision update module 103 can also be used for: pre-training an elevation classification network, the elevation classification network is used to convert the depth estimation problem into a classification problem, and the elevation classification network is based on the obstacle type, elevation true value coding information, obstacle area mask information, point cloud features and image features of the obstacle area grid, and spatial voxel weights of the obstacle area grid to determine the loss function; the elevation information of the grid in the obstacle area is obtained through the elevation classification network.
示例性的,在本实施例的另一些实施方式中,上述精度更新模块103还可用于:根据上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息、当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,确定当前帧障碍物区域包含的栅格;获取当前帧障碍物区域包含栅格的高程信息,并对当前帧障碍物区域包含栅格的空间体素权重进行更新。Exemplarily, in some other implementations of this embodiment, the above-mentioned precision update module 103 can also be used to: determine the grid included in the obstacle area of the current frame according to the position information of the obstacle identified in the previous frame at the current moment in the bird's-eye view grid area, and the position information of the obstacle identified in the current frame at the current moment in the bird's-eye view grid area; obtain the elevation information of the grid included in the obstacle area of the current frame, and update the spatial voxel weights of the grid included in the obstacle area of the current frame.
作为上述实施例的一种示例性的实施方式,上述精度更新模块103还可进一步用于:根据上一帧和当前帧对应的障碍物识别结果,确定对应障碍物区域内栅格;获取车辆行驶前方区域检测到的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息;车辆行驶前方区域在沿车辆行驶方向上与车辆间的最小距离大于等于待检测路面区域与车辆的最大距离;根据车辆在相邻帧时间段内在局部世界坐标系的位置变化,更新上一帧障碍物区域内栅格的位置信息,以作为上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息;若当前点位置为上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或前方行驶区域识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,则为当前点设置第一预设值,以作为掩码值;若当前点位置不为上一帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或当前帧识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息或前方行驶区域识别的障碍物在当前时刻的鸟瞰视图栅格区域的位置信息,则为当前点设置第二预设值,以作为掩码值;根据待检测路面区域的各点的掩码值,生成障碍物区域掩码信息,以用于标识当前帧障碍物区域包含的栅格。As an exemplary implementation of the above embodiment, the above precision update module 103 can also be further used to: determine the grid in the corresponding obstacle area according to the obstacle recognition results corresponding to the previous frame and the current frame; obtain the position information of the bird's-eye view grid area of the obstacle detected in the area ahead of the vehicle at the current moment; the minimum distance between the area ahead of the vehicle and the vehicle along the vehicle's driving direction is greater than or equal to the maximum distance between the road surface area to be detected and the vehicle; update the position information of the grid in the obstacle area of the previous frame according to the position change of the vehicle in the local world coordinate system in the adjacent frame time period, as the position information of the bird's-eye view grid area of the obstacle identified in the previous frame at the current moment; if the current point position is the bird's-eye view grid area of the obstacle identified in the previous frame at the current moment If the position information of the grid area or the position information of the bird's-eye view grid area of the obstacle identified in the current frame at the current moment or the position information of the bird's-eye view grid area of the obstacle identified in the front driving area at the current moment is present, a first preset value is set for the current point as a mask value; if the current point position is not the position information of the bird's-eye view grid area of the obstacle identified in the previous frame at the current moment or the position information of the bird's-eye view grid area of the obstacle identified in the current frame at the current moment or the position information of the bird's-eye view grid area of the obstacle identified in the front driving area at the current moment, a second preset value is set for the current point as a mask value; according to the mask value of each point in the road surface area to be detected, the obstacle area mask information is generated to identify the grid contained in the obstacle area of the current frame.
上文中提到的路面障碍物三维形貌信息确定装置是从功能模块的角度描述,进一步的,本发明还提供一种电子设备,是从硬件角度描述。图11为本发明实施例提供的电子设备在一种实施方式下的结构示意图。如图11所示,该电子设备包括存储器110,用于存储计算机程序;处理器111,用于执行计算机程序时实现如上述任一实施例提到的路面障碍物三维形貌信息确定方法的步骤。The device for determining the three-dimensional shape information of road obstacles mentioned above is described from the perspective of functional modules. Furthermore, the present invention also provides an electronic device, which is described from the perspective of hardware. Figure 11 is a schematic diagram of the structure of an electronic device provided by an embodiment of the present invention under one implementation. As shown in Figure 11, the electronic device includes a memory 110 for storing a computer program; a processor 111 for implementing the steps of the method for determining the three-dimensional shape information of road obstacles as mentioned in any of the above embodiments when executing the computer program.
其中,处理器111可以包括一个或多个处理核心,比如4核心处理器、8核心处理器,处理器111还可为控制器、微控制器、微处理器或其他数据处理芯片等。处理器111可以采用DSP(Digital Signal Processing,数字信号处理)、FPGA(Field-Programmable GateArray,现场可编程门阵列)、PLA(Programmable Logic Array,可编程逻辑阵列)中的至少一种硬件形式来实现。处理器111也可以包括主处理器和协处理器,主处理器是用于对在唤醒状态下的数据进行处理的处理器,也称CPU(Central Processing Unit,中央处理器);协处理器是用于对在待机状态下的数据进行处理的低功耗处理器。在一些实施例中,处理器111可以集成有GPU(Graphics Processing Unit,图形处理器),GPU用于负责显示屏所需要显示的内容的渲染和绘制。一些实施例中,处理器111还可以包括AI(ArtificialIntelligence,人工智能)处理器,该AI处理器用于处理有关机器学习的计算操作。Among them, the processor 111 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and the processor 111 may also be a controller, a microcontroller, a microprocessor or other data processing chip. The processor 111 may be implemented in at least one hardware form of DSP (Digital Signal Processing), FPGA (Field-Programmable Gate Array), and PLA (Programmable Logic Array). The processor 111 may also include a main processor and a coprocessor. The main processor is a processor for processing data in the awake state, also known as a CPU (Central Processing Unit); the coprocessor is a low-power processor for processing data in the standby state. In some embodiments, the processor 111 may be integrated with a GPU (Graphics Processing Unit), which is responsible for rendering and drawing the content to be displayed on the display screen. In some embodiments, the processor 111 may also include an AI (Artificial Intelligence) processor, which is used to process computing operations related to machine learning.
存储器110可以包括一个或多个计算机非易失性存储介质,该计算机非易失性存储介质可以是非暂态的。存储器110还可包括高速随机存取存储器以及非易失性存储器,比如一个或多个磁盘存储设备、闪存存储设备。存储器110在一些实施例中可以是电子设备的内部存储单元,例如服务器的硬盘。存储器110在另一些实施例中也可以是电子设备的外部存储设备,例如服务器上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,存储器110还可以既包括电子设备的内部存储单元也包括外部存储设备。存储器110不仅可以用于存储安装于电子设备的应用软件及各类数据,例如:执行路面障碍物三维形貌信息确定方法过程中的程序的代码等,还可以用于暂时地存储已经输出或者将要输出的数据。本实施例中,存储器110至少用于存储以下计算机程序1101,其中,该计算机程序被处理器111加载并执行之后,能够实现前述任一实施例公开的路面障碍物三维形貌信息确定方法的相关步骤。另外,存储器110所存储的资源还可以包括操作系统1102和数据1103等,存储方式可以是短暂存储或者永久存储。其中,操作系统1102可以包括Windows、Unix、Linux等。数据1103可以包括但不限于路面障碍物三维形貌信息确定结果对应的数据等。The memory 110 may include one or more computer non-volatile storage media, which may be non-transitory. The memory 110 may also include high-speed random access memory and non-volatile memory, such as one or more disk storage devices and flash memory storage devices. In some embodiments, the memory 110 may be an internal storage unit of an electronic device, such as a hard disk of a server. In other embodiments, the memory 110 may also be an external storage device of an electronic device, such as a plug-in hard disk equipped on a server, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, a flash card (Flash Card), etc. Further, the memory 110 may also include both an internal storage unit of an electronic device and an external storage device. The memory 110 can not only be used to store application software and various types of data installed in the electronic device, such as: the code of the program in the process of executing the method for determining the three-dimensional shape information of road obstacles, but also can be used to temporarily store data that has been output or is to be output. In this embodiment, the memory 110 is used to store at least the following computer program 1101, wherein after the computer program is loaded and executed by the processor 111, the relevant steps of the method for determining the three-dimensional shape information of the road obstacle disclosed in any of the above embodiments can be implemented. In addition, the resources stored in the memory 110 may also include an operating system 1102 and data 1103, etc., and the storage method may be temporary storage or permanent storage. Among them, the operating system 1102 may include Windows, Unix, Linux, etc. The data 1103 may include, but is not limited to, data corresponding to the result of determining the three-dimensional shape information of the road obstacle, etc.
在一些实施例中,上述电子设备还可包括有显示屏112、输入输出接口113、通信接口114或者称为网络接口、电源115以及通信总线116。其中,显示屏112、输入输出接口113比如键盘(Keyboard)属于用户接口,示例性的用户接口还可以包括标准的有线接口、无线接口等。可选地,在一些实施例中,显示器可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(Organic Light-Emitting Diode,有机发光二极管)触摸器等。显示器也可以适当的称为显示屏或显示单元,用于显示在电子设备中处理的信息以及用于显示可视化的用户界面。通信接口114示例性的可以包括有线接口和/或无线接口,如WI-FI接口、蓝牙接口等,通常用于在电子设备与其他电子设备之间建立通信连接。通信总线116可以是外设部件互连标准(peripheral component interconnect,简称PCI)总线或扩展工业标准结构(extended industry standard architecture,简称EISA)总线等。该总线可以分为地址总线、数据总线、控制总线等。为便于表示,图11中仅用一条粗线表示,但并不表示仅有一根总线或一种类型的总线。In some embodiments, the electronic device may further include a display screen 112, an input/output interface 113, a communication interface 114 or a network interface, a power supply 115 and a communication bus 116. Among them, the display screen 112 and the input/output interface 113, such as a keyboard, belong to a user interface, and an exemplary user interface may also include a standard wired interface, a wireless interface, etc. Optionally, in some embodiments, the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, and an OLED (Organic Light-Emitting Diode) touch device, etc. The display may also be appropriately referred to as a display screen or a display unit, which is used to display information processed in the electronic device and to display a visual user interface. The communication interface 114 may exemplarily include a wired interface and/or a wireless interface, such as a WI-FI interface, a Bluetooth interface, etc., which is generally used to establish a communication connection between an electronic device and other electronic devices. The communication bus 116 may be a peripheral component interconnect (PCI) bus or an extended industry standard architecture (EISA) bus, etc. The bus may be divided into an address bus, a data bus, a control bus, etc. For ease of representation, FIG11 only uses one thick line, but this does not mean that there is only one bus or one type of bus.
本领域技术人员可以理解,图11中示出的结构并不构成对该电子设备的限定,可以包括比图示更多或更少的组件,例如还可包括实现各类功能的传感器117。Those skilled in the art will appreciate that the structure shown in FIG. 11 does not limit the electronic device and may include more or fewer components than shown in the figure, for example, may also include a sensor 117 for implementing various functions.
可以理解的是,如果上述实施例中路面障碍物三维形貌信息确定方法以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个非易失性存储介质中。基于这样的理解,本发明的技术方案本质上或者说对相关技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,执行本发明各个实施例方法的全部或部分步骤。而前述的存储介质包括并不限制于:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random AccessMemory,RAM)、电可擦除可编程ROM、寄存器、硬盘、多媒体卡、卡型存储器(例如SD或DX存储器等)、磁性存储器、可移动磁盘、CD-ROM、磁碟或者光盘等各种可以存储程序代码的介质。基于此,本发明还提供了一种非易失性存储介质,其上存储有计算机程序,该计算机程序被处理器执行时如上任意一实施例所记载的路面障碍物三维形貌信息确定方法的步骤。It is understandable that if the method for determining the three-dimensional shape information of road obstacles in the above embodiment is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a non-volatile storage medium. Based on such an understanding, the technical solution of the present invention is essentially or the part that contributes to the relevant technology or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium to execute all or part of the steps of the methods of various embodiments of the present invention. The aforementioned storage medium includes but is not limited to: U disk, mobile hard disk, read-only memory (ROM), random access memory (RAM), electrically erasable programmable ROM, register, hard disk, multimedia card, card-type memory (such as SD or DX memory, etc.), magnetic memory, removable disk, CD-ROM, magnetic disk or optical disk, etc. Various media that can store program codes. Based on this, the present invention also provides a non-volatile storage medium on which a computer program is stored, and when the computer program is executed by a processor, the steps of the method for determining the three-dimensional shape information of road obstacles recorded in any of the above embodiments are executed.
可以理解的是,如果上述实施例中的路面障碍物三维形貌信息确定方法以软件功能单元的形式实现并作为独立的产品销售或使用时,该计算机软件产品也可不需要存储在一个实体的存储介质中,例如可通过有线网络或无线网络直接传输至计算机等具有信息处理能力的装置执行本发明各个实施例方法的全部或部分步骤。基于这样的理解,本发明的技术方案本质上或者说对相关技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,基于此,本发明还提供了一种计算机程序产品,存储有计算机程序,该计算机程序被处理器执行时如上任意一实施例所记载的路面障碍物三维形貌信息确定方法的步骤。It is understandable that if the method for determining the three-dimensional shape information of road obstacles in the above-mentioned embodiments is implemented in the form of a software functional unit and sold or used as an independent product, the computer software product does not need to be stored in a physical storage medium. For example, it can be directly transmitted to a computer or other device with information processing capabilities through a wired network or a wireless network to perform all or part of the steps of the methods of various embodiments of the present invention. Based on this understanding, the technical solution of the present invention is essentially or the part that contributes to the relevant technology or all or part of the technical solution can be embodied in the form of a software product. Based on this, the present invention also provides a computer program product, which stores a computer program. When the computer program is executed by a processor, the steps of the method for determining the three-dimensional shape information of road obstacles recorded in any of the above embodiments are performed.
最后,本发明还提供了一种无人驾驶车辆,完全复用自身的传感器,如相机、激光雷达、里程计、惯性测量单元,不需要额外加装其他传感器,在现有的自动驾驶感知系统中内置实现如前任一个实施例所记载的路面障碍物三维形貌信息确定方法的计算机程序,该自动驾驶感知系统可用于执行存储器中存储的计算机程序时实现如前任一个实施例所记载的路面障碍物三维形貌信息确定方法的步骤,通过点云稠密化方法增加低矮凸起障碍物/坑洼区域识别准确性,基于多源传感器数据,针对近场数据特性,通过检测识别与校验,可以精准识别低矮障碍物/坑洼区域相对于无人驾驶车辆的三维位置、三维形貌数据,如包括高度、深度、大小,这样无人驾驶车辆的自动驾驶场景下便能够尽早准确识别低矮凸起障碍物,如减速带和裂缝,深坑、浅坑或水坑这类坑洼路面,如图12所示,并将低矮障碍物/坑洼信息融入自动驾驶规划控制,提升车辆的避障能力,有效提升自驾系统安全性和舒适度。Finally, the present invention also provides an unmanned vehicle, which fully reuses its own sensors, such as cameras, laser radars, odometers, and inertial measurement units, without the need to install other sensors. A computer program for implementing the method for determining the three-dimensional shape information of road obstacles as recorded in any of the previous embodiments is built into the existing automatic driving perception system. The automatic driving perception system can be used to implement the steps of the method for determining the three-dimensional shape information of road obstacles as recorded in any of the previous embodiments when executing the computer program stored in the memory. The recognition accuracy of low raised obstacles/pothole areas is increased by a point cloud densification method. Based on multi-source sensor data, according to the characteristics of near-field data, through detection, identification and verification, the three-dimensional position and three-dimensional shape data of low obstacles/pothole areas relative to the unmanned vehicle can be accurately identified, such as height, depth, and size. In this way, the unmanned vehicle can accurately identify low raised obstacles such as speed bumps and cracks, deep pits, shallow pits or puddles as early as possible in the automatic driving scene, as shown in Figure 12, and the low obstacle/pothole information is integrated into the automatic driving planning and control to improve the obstacle avoidance capability of the vehicle, and effectively improve the safety and comfort of the self-driving system.
本说明书中各个实施例采用递进的方式描述,每个实施例重点说明的都是与其它实施例的不同之处,各个实施例之间相同或相似部分互相参见即可。对于实施例公开的硬件包括电子设备、非易失性存储介质、计算机程序产品及无人驾驶车辆而言,由于其与实施例公开的方法相对应,所以描述的比较简单,相关之处参见方法部分说明即可。In this specification, each embodiment is described in a progressive manner, and each embodiment focuses on the differences from other embodiments. The same or similar parts between the embodiments can be referred to each other. For the hardware disclosed in the embodiments, including electronic devices, non-volatile storage media, computer program products and unmanned vehicles, since they correspond to the methods disclosed in the embodiments, the description is relatively simple, and the relevant parts can be referred to the method part description.
专业人员还可以进一步意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、计算机软件或者二者的结合来实现,为了清楚地说明硬件和软件的可互换性,在上述说明中已经按照功能一般性地描述了各示例的组成及步骤。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本发明的范围。Professionals may further appreciate that the units and algorithm steps of each example described in conjunction with the embodiments disclosed herein can be implemented in electronic hardware, computer software, or a combination of the two. In order to clearly illustrate the interchangeability of hardware and software, the composition and steps of each example have been generally described in the above description according to function. Whether these functions are performed in hardware or software depends on the specific application and design constraints of the technical solution. Professionals and technicians may use different methods to implement the described functions for each specific application, but such implementation should not be considered to be beyond the scope of the present invention.
以上对本发明所提供的一种路面障碍物三维形貌信息确定方法、电子设备、非易失性存储介质、计算机程序产品及无人驾驶车辆进行了详细介绍。本文中应用了具体个例对本发明的原理及实施方式进行了阐述,以上实施例的说明只是用于帮助理解本发明的方法及其核心思想。应当指出,基于本发明中的实施例,对于本技术领域的普通技术人员来说,在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。在不脱离本发明原理的前提下,还可以对本发明进行若干改进和修饰,这些改进和修饰也落入本发明的保护范围内。The above is a detailed introduction to a method for determining three-dimensional morphological information of road obstacles, an electronic device, a non-volatile storage medium, a computer program product and an unmanned vehicle provided by the present invention. Specific examples are used herein to illustrate the principles and implementation methods of the present invention. The description of the above embodiments is only used to help understand the method of the present invention and its core idea. It should be pointed out that based on the embodiments of the present invention, for ordinary technicians in this technical field, all other embodiments obtained without creative work are within the scope of protection of the present invention. Without departing from the principles of the present invention, the present invention may also be improved and modified in a number of ways, and these improvements and modifications also fall within the scope of protection of the present invention.
Claims (15)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202411379463.3A CN118898838B (en) | 2024-09-30 | 2024-09-30 | Method, device, medium and vehicle for determining three-dimensional shape information of road obstacles |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202411379463.3A CN118898838B (en) | 2024-09-30 | 2024-09-30 | Method, device, medium and vehicle for determining three-dimensional shape information of road obstacles |
Publications (2)
Publication Number | Publication Date |
---|---|
CN118898838A true CN118898838A (en) | 2024-11-05 |
CN118898838B CN118898838B (en) | 2025-01-24 |
Family
ID=93270020
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202411379463.3A Active CN118898838B (en) | 2024-09-30 | 2024-09-30 | Method, device, medium and vehicle for determining three-dimensional shape information of road obstacles |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN118898838B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN119090950A (en) * | 2024-11-07 | 2024-12-06 | 安徽擎天智能科技有限公司 | Obstacle position detection method, device, control equipment and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210041887A1 (en) * | 2019-08-06 | 2021-02-11 | Boston Dynamics, Inc. | Constrained Mobility Mapping |
CN116524472A (en) * | 2023-06-30 | 2023-08-01 | 广汽埃安新能源汽车股份有限公司 | Obstacle detection method, device, storage medium and equipment |
CN117095382A (en) * | 2023-09-05 | 2023-11-21 | 北京星网船电科技有限公司 | Obstacle detection method, device, equipment and medium based on camera and radar |
CN117302263A (en) * | 2023-10-10 | 2023-12-29 | 中国重汽集团济南动力有限公司 | Method, device, equipment and storage medium for detecting drivable area |
-
2024
- 2024-09-30 CN CN202411379463.3A patent/CN118898838B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210041887A1 (en) * | 2019-08-06 | 2021-02-11 | Boston Dynamics, Inc. | Constrained Mobility Mapping |
CN116524472A (en) * | 2023-06-30 | 2023-08-01 | 广汽埃安新能源汽车股份有限公司 | Obstacle detection method, device, storage medium and equipment |
CN117095382A (en) * | 2023-09-05 | 2023-11-21 | 北京星网船电科技有限公司 | Obstacle detection method, device, equipment and medium based on camera and radar |
CN117302263A (en) * | 2023-10-10 | 2023-12-29 | 中国重汽集团济南动力有限公司 | Method, device, equipment and storage medium for detecting drivable area |
Non-Patent Citations (1)
Title |
---|
李博杨: "基于三维激光雷达的道路环境感知", 中国优秀硕士论文电子期刊网, no. 2020, 31 January 2020 (2020-01-31) * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN119090950A (en) * | 2024-11-07 | 2024-12-06 | 安徽擎天智能科技有限公司 | Obstacle position detection method, device, control equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN118898838B (en) | 2025-01-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111874006B (en) | Route planning processing method and device | |
CN109470254B (en) | Map lane line generation method, device, system and storage medium | |
WO2018068653A1 (en) | Point cloud data processing method and apparatus, and storage medium | |
CN108955702A (en) | Based on the lane of three-dimensional laser and GPS inertial navigation system grade map creation system | |
CN118898838B (en) | Method, device, medium and vehicle for determining three-dimensional shape information of road obstacles | |
CN118898689B (en) | Cost map generation method, equipment, medium, product and unmanned vehicle | |
CN116863687B (en) | Quasi-all-weather traffic safety passing guarantee system based on vehicle-road cooperation | |
CN118898825B (en) | Road environment state perception method, equipment, medium, program product and vehicle | |
CN116299315A (en) | Method and device for detecting road surface obstacle in real time by using laser radar | |
CN118894117B (en) | Unmanned vehicle, driving route generation method, device, medium and product | |
CN117471463A (en) | Obstacle detection method based on 4D radar and image recognition fusion | |
CN117671644A (en) | Signboard detection method and device and vehicle | |
KR102824305B1 (en) | Method and System for change detection and automatic updating of road marking in HD map through IPM image and HD map fitting | |
CN117765230A (en) | Image processing method, device and mobile tool | |
Huo et al. | A Review of Key Technologies for Environment Sensing in Driverless Vehicles. | |
CN119516503A (en) | Road inspection method and system based on data fusion and edge computing | |
CN118941626A (en) | Roadside visual positioning method, device, equipment and medium for traffic participants | |
US11555928B2 (en) | Three-dimensional object detection with ground removal intelligence | |
CN118015132A (en) | A method, device and storage medium for processing vehicle driving data | |
CN116935344A (en) | Road boundary polygonal contour construction method and system | |
US12298400B2 (en) | False positive object removal with surfel maps | |
CN118898826B (en) | Road target detection method, equipment, medium, product and unmanned vehicle | |
CN117795566A (en) | Perception of three-dimensional objects in sensor data | |
CN114004951B (en) | Road sign extraction method and system combining point cloud intensity and geometric structure | |
CN119169064B (en) | Vehicle bottom transparent registration method, system, medium and program product |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |