CN117011656A - Panoramic camera and laser radar fusion method for obstacle avoidance of unmanned boarding bridge - Google Patents
Panoramic camera and laser radar fusion method for obstacle avoidance of unmanned boarding bridge Download PDFInfo
- Publication number
- CN117011656A CN117011656A CN202310808228.2A CN202310808228A CN117011656A CN 117011656 A CN117011656 A CN 117011656A CN 202310808228 A CN202310808228 A CN 202310808228A CN 117011656 A CN117011656 A CN 117011656A
- Authority
- CN
- China
- Prior art keywords
- module
- panoramic camera
- image
- data
- obstacle avoidance
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/54—Extraction of image or video features relating to texture
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/761—Proximity, similarity or dissimilarity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/762—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using clustering, e.g. of similar faces in social networks
- G06V10/763—Non-hierarchical techniques, e.g. based on statistics of modelling distributions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
- G06V10/765—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects using rules for classification or partitioning the feature space
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/766—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using regression, e.g. by projecting features on hyperplanes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/7715—Feature extraction, e.g. by transforming the feature space, e.g. multi-dimensional scaling [MDS]; Mappings, e.g. subspace methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- General Physics & Mathematics (AREA)
- Software Systems (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Image Processing (AREA)
Abstract
Description
技术领域Technical field
本发明涉及无人对接廊桥的避障感知技术领域,具体涉及一种面向无人登机桥避障的全景相机和激光雷达融合方法。The invention relates to the technical field of obstacle avoidance sensing for unmanned docking bridges, and specifically relates to a panoramic camera and lidar fusion method for obstacle avoidance on unmanned boarding bridges.
背景技术Background technique
无人对接廊桥是一种能够自动与飞机对接的智能设备,它可以实现无人值守、高效安全、节能环保等优点。为了实现无人对接廊桥的避障和人员等目标感知功能,需要对周围环境进行实时、准确、全面的感知,以便及时调整运动轨迹和速度,避免碰撞和伤害。The unmanned docking bridge is an intelligent device that can automatically dock with aircraft. It can achieve the advantages of unattended, efficient, safe, energy-saving and environmentally friendly. In order to realize the target sensing function of the unmanned docking bridge such as obstacle avoidance and personnel, it is necessary to have real-time, accurate and comprehensive perception of the surrounding environment in order to adjust the movement trajectory and speed in a timely manner to avoid collisions and injuries.
目前,常用的环境感知传感器有摄像机、激光雷达、超声波、毫米波等。其中,摄像机可以提供丰富的图像信息,但是深度信息不准确或缺失;激光雷达可以提供精确的距离信息,但是分辨率低或视场角小;超声波和毫米波可以提供近距离的避障信息,但是受干扰大或成本高。因此,单一传感器难以满足无人对接廊桥的环境感知需求。Currently, commonly used environment sensing sensors include cameras, laser radar, ultrasonic waves, millimeter waves, etc. Among them, cameras can provide rich image information, but depth information is inaccurate or missing; lidar can provide accurate distance information, but has low resolution or small field of view; ultrasonic waves and millimeter waves can provide short-range obstacle avoidance information. But the interference is large or the cost is high. Therefore, it is difficult for a single sensor to meet the environmental sensing needs of unmanned docking bridges.
随着高速通信和人工智能技术的快速发展,人类对真实世界场景的感知不再局限于使用小视场(FoV)和低维场景检测设备。全景成像应运而生,成为下一代创新智能仪器用于环境感知和测量。With the rapid development of high-speed communication and artificial intelligence technology, human perception of real-world scenes is no longer limited to the use of small field of view (FoV) and low-dimensional scene detection devices. Panoramic imaging emerged as the next generation of innovative smart instruments for environmental sensing and measurement.
现有技术(Review on Panoramic Imaging and Its Applications in SceneUnderstanding,Shaohua Gao Kailun Yang Hao Shi Kaiwei Wang and Jian Bai1)介绍了全景相机的原理,阐述了全景相机在全景语义图像分割、全景深度估计、全景视觉定位等方向的重要价值和制约因素,展望了全景成像仪器未来的潜力和研究方向,也指出虽然全景成像仪器满足了大视场摄影成像的需求,但还需要具备高分辨率、无盲区、微型化和多维智能感知能力,需要与人工智能方法结合,才能实现更深入的对360°真实环境的理解和全面感知。Existing Technology (Review on Panoramic Imaging and Its Applications in SceneUnderstanding, Shaohua Gao Kailun Yang Hao Shi Kaiwei Wang and Jian Bai1) introduces the principle of panoramic cameras and explains the functions of panoramic cameras in panoramic semantic image segmentation, panoramic depth estimation, and panoramic visual positioning. The important value and restrictive factors in such directions, looking forward to the future potential and research direction of panoramic imaging instruments, also pointed out that although panoramic imaging instruments meet the needs of large-field photography and imaging, they also need to have high resolution, no blind spots, and miniaturization. And multi-dimensional intelligent perception capabilities need to be combined with artificial intelligence methods to achieve a deeper understanding and comprehensive perception of the 360° real environment.
现有技术(Rethinking Supervised Depth Estimation for 360°PanoramicImagery Lu He,Bing Jian,Yangming Wen,Haichao Zhu,Kelin Liu,Weiwei Feng,ShanLiu Tencent America,Palo Alto,USA)提出了一个从单个360°全景图像中估算深度的方法,认为这是一个复杂的任务。因为固有的尺度歧义问题,从RGB全景图像估算深度图变得更为棘手。为了解决这个问题,他们基于预估的相机高度提出了一个规范化深度数据的方法,以缓解真实深度图中的尺度不一致问题。除此之外,他们还设计了一个多头平面引导的深度网络,旨在为深度估计提供更多的几何约束。实验结果明确表示,相对深度估测比绝对深度估测更准确。这个模型在Matterport3D和Stanford2D3D数据集上都表现出了不错的性能。其限制在于:所提出的方法需要估计摄像机高度,在某些情况下,这可能不可用或不准确。所提出的方法依赖于场景中平面结构的假设,这在某些情况下可能不成立。由于其计算复杂性,所提出的方法可能不适合实时应用。The existing technology (Rethinking Supervised Depth Estimation for 360° PanoramicImagery Lu He, Bing Jian, Yangming Wen, Haichao Zhu, Kelin Liu, Weiwei Feng, ShanLiu Tencent America, Palo Alto, USA) proposes an estimation from a single 360° panoramic image Deep approach considers this a complex task. Estimating depth maps from RGB panoramic images becomes more difficult because of the inherent scale ambiguity problem. To solve this problem, they proposed a method to normalize depth data based on the estimated camera height to alleviate the scale inconsistency problem in real depth maps. In addition to this, they also designed a multi-head planar guided depth network aiming to provide more geometric constraints for depth estimation. Experimental results clearly show that relative depth estimation is more accurate than absolute depth estimation. This model shows good performance on both Matterport3D and Stanford2D3D datasets. Its limitation is: the proposed method requires an estimate of the camera height, which may not be available or accurate in some cases. The proposed method relies on the assumption of planar structure in the scene, which may not hold in some cases. Due to its computational complexity, the proposed method may not be suitable for real-time applications.
上述全景视觉传感器进行环境感知的方法。通过设计合理的全景相机方案,实现对360度视场的覆盖并估计深度。该方法具有视场广、成本低、易于安装等优点,但是也存在以下不足:The above panoramic vision sensor is a method for environmental perception. Through a properly designed panoramic camera solution, coverage of the 360-degree field of view and depth estimation can be achieved. This method has the advantages of wide field of view, low cost, and easy installation, but it also has the following shortcomings:
a、全景相机深度估计精度低,容易受到光照、纹理等因素的影响;a. The depth estimation accuracy of panoramic cameras is low and is easily affected by factors such as lighting and texture;
b、全景相机深度估计计算量大,难以实现实时性;b. Panoramic camera depth estimation requires a large amount of calculation and is difficult to achieve real-time performance;
c、全景相机深度估计结果不稳定,容易出现漏检、误检等现象。c. The depth estimation results of panoramic cameras are unstable and prone to missed detections, false detections, etc.
因此,如何提高全景相机深度估计的精度、速度和稳定性,是当前无人对接廊桥环境感知领域亟待解决的技术问题。Therefore, how to improve the accuracy, speed and stability of panoramic camera depth estimation is an urgent technical issue to be solved in the field of environmental perception for unmanned docking bridges.
发明内容Contents of the invention
为解决现有技术的缺陷和不足问题;本发明的目的在于提供一种面向无人登机桥避障的全景相机和激光雷达融合方法;能够有效地利用全景相机和激光传感器的数据特点,实现对环境图像和距离数据的匹配、校准、分割、识别和定位,从而提高全景相机深度估计的精度、速度和稳定性。In order to solve the defects and shortcomings of the existing technology; the purpose of the present invention is to provide a panoramic camera and lidar fusion method for unmanned boarding bridge obstacle avoidance; it can effectively utilize the data characteristics of the panoramic camera and the laser sensor to achieve Match, calibrate, segment, identify and position environmental images and distance data to improve the accuracy, speed and stability of panoramic camera depth estimation.
本发明至少通过如下技术方案之一实现。The present invention is realized through at least one of the following technical solutions.
一种面向无人登机桥避障的全景相机和激光雷达融合方法,包括以下步骤:A panoramic camera and lidar fusion method for unmanned boarding bridge obstacle avoidance, including the following steps:
(S1)、数据预处理模块对全景相机和激光传感器的数据进行预处理,包括去噪、滤波、校正;(S1). The data preprocessing module preprocesses the data from panoramic cameras and laser sensors, including denoising, filtering, and correction;
(S2)、特征提取模块对全景相机的环境图像进行特征提取,利用卷积神经网络,提取图像的边缘、纹理、颜色特征,并将特征表示为高维向量;(S2), the feature extraction module extracts features from the environment image of the panoramic camera, uses a convolutional neural network to extract the edge, texture, and color features of the image, and represents the features as high-dimensional vectors;
(S3)、对激光传感器的距离数据进行特征提取,利用点云处理方法,提取距离数据的形状、大小、方向特征,并将特征表示为高维向量;(S3), perform feature extraction on the distance data of the laser sensor, use point cloud processing methods to extract the shape, size, and direction features of the distance data, and represent the features as high-dimensional vectors;
(S4)、特征匹配模块对全景相机和激光传感器的特征向量进行匹配,利用相似度度量方法,计算不同传感器之间的特征向量;(S4). The feature matching module matches the feature vectors of the panoramic camera and the laser sensor, and uses the similarity measurement method to calculate the feature vectors between different sensors;
(S5)、数据融合模块对匹配的特征向量进行融合,利用加权平均方法,将不同传感器的特征向量融合为一个统一的特征向量,反映环境的图像和距离信息;(S5). The data fusion module fuses the matching feature vectors and uses the weighted average method to fuse the feature vectors of different sensors into a unified feature vector that reflects the image and distance information of the environment;
(S6)、图像分割模块对融合的特征向量进行分割,利用聚类方法,将融合的特征向量分割为不同的区域,每个区域对应一个目标或背景;(S6). The image segmentation module segments the fused feature vectors and uses the clustering method to divide the fused feature vectors into different areas, each area corresponding to a target or background;
(S7)、目标识别模块对分割的区域进行识别,利用分类方法,根据区域的特征向量,判断区域的类别;(S7). The target recognition module identifies the segmented areas, uses the classification method, and determines the category of the area based on the feature vector of the area;
(S8)、目标定位模块对识别的目标进行定位,利用回归方法,根据区域的特征向量,计算目标的参数。(S8). The target positioning module locates the identified target and uses the regression method to calculate the parameters of the target based on the feature vector of the region.
进一步地,步骤(S1)中对鱼眼相机采集到的扭曲图像进行畸变校正为:Further, in step (S1), the distortion correction of the distorted image collected by the fisheye camera is as follows:
x'=x(1+k1r2+k2r4+k3r6)x'=x(1+k1r 2 +k2r 4 +k3r 6 )
y'=y(1+k1r2+k2r4+k3r6)y'=y(1+k1r 2 +k2r 4 +k3r 6 )
其中,x和y是扭曲图像上的坐标,x'和y'是校正后图像上的坐标,r是扭曲图像上点到中心点的距离,k1、k2和k3是畸变系数。Among them, x and y are the coordinates on the distorted image, x' and y' are the coordinates on the corrected image, r is the distance from the point on the distorted image to the center point, k1, k2 and k3 are the distortion coefficients.
进一步地,步骤(S2)中对RGB-D图像进行特征提取为:Further, the feature extraction of the RGB-D image in step (S2) is:
f=W*x+bf=W*x+b
其中,f是特征向量,W是卷积核,x是RGB-D图像,b是偏置项。Among them, f is the feature vector, W is the convolution kernel, x is the RGB-D image, and b is the bias term.
进一步地,步骤(S3)中使用以下公式对点云数据进行特征提取:Further, in step (S3), the following formula is used to extract features from the point cloud data:
f=g(h(p))f=g(h(p))
其中,f是特征向量,g是非线性激活函数,h是点云处理函数,p是点云数据。Among them, f is the feature vector, g is the nonlinear activation function, h is the point cloud processing function, and p is the point cloud data.
进一步地,步骤(S4)中使用以下公式计算相似度:Further, in step (S4), the following formula is used to calculate the similarity:
s=cos(f1,f2)=f1*f2/(|f1|*|f2|)s=cos(f1,f2)=f1*f2/(|f1|*|f2|)
其中,s是相似度,cos是余弦函数,f1和f2是不同传感器的特征向量。Among them, s is the similarity, cos is the cosine function, f1 and f2 are the feature vectors of different sensors.
进一步地,步骤(S5)中,使用以下公式进行数据融合:Further, in step (S5), the following formula is used for data fusion:
F=a*f1+(1-a)*f2F=a*f1+(1-a)*f2
其中,F是融合后的特征向量,a是权重系数,f1和f2是不同传感器的特征向量。Among them, F is the fused feature vector, a is the weight coefficient, f1 and f2 are the feature vectors of different sensors.
进一步地,步骤(S6)中,使用以下公式进行图像分割:Further, in step (S6), the following formula is used for image segmentation:
L=argmin(S(F))L=argmin(S(F))
其中,L是分割后的标签,S是图像分割函数,F是融合后的特征向量。Among them, L is the segmented label, S is the image segmentation function, and F is the fused feature vector.
实现所述的一种面向无人登机桥避障的全景相机和激光雷达融合方法的系统,全景相机模块、激光传感器模块、数据预处理模块、特征提取模块、特征匹配模块、数据融合模块、图像分割模块、目标识别模块和目标定位模块;所述全景相机模块和激光传感器模块均与数据预处理模块连接,数据预处理模块与特征提取模块、特征匹配模块、数据融合模块、图像分割模块、目标识别模块和目标定位模块之间相互连接。A system that implements the panoramic camera and lidar fusion method for unmanned boarding bridge obstacle avoidance, including a panoramic camera module, a laser sensor module, a data preprocessing module, a feature extraction module, a feature matching module, and a data fusion module. Image segmentation module, target recognition module and target positioning module; the panoramic camera module and laser sensor module are connected with the data preprocessing module, and the data preprocessing module is connected with the feature extraction module, feature matching module, data fusion module, image segmentation module, The target identification module and the target positioning module are connected to each other.
进一步地,所述全景相机模块为双全景相机,所述双全景相机包括六个120度鱼眼相机,用于采集环境图像数据,其中,每个鱼眼相机采集120度的视场角,三个鱼眼相机拼接成一个360度的全景相机,两个全景相机构成一个双目全景相机,实现对360度视场的覆盖并估计深度。Further, the panoramic camera module is a dual panoramic camera. The dual panoramic camera includes six 120-degree fisheye cameras for collecting environmental image data. Each fisheye camera collects a field of view of 120 degrees. Two fisheye cameras are spliced into a 360-degree panoramic camera, and two panoramic cameras form a binocular panoramic camera to achieve coverage of the 360-degree field of view and estimate depth.
进一步地,所述激光传感器模块包括旋转式多线激光雷达。Further, the laser sensor module includes a rotating multi-line lidar.
与现有的技术相比,本发明的有益效果为:Compared with existing technology, the beneficial effects of the present invention are:
1、能够有效地利用全景相机和激光传感器的数据特点,实现对环境图像和距离数据的匹配、校准、分割、识别和定位;1. Able to effectively utilize the data characteristics of panoramic cameras and laser sensors to achieve matching, calibration, segmentation, identification and positioning of environmental images and distance data;
2、能够有效地提高全景相机深度估计的精度、速度和稳定性,提高无人对接廊桥避障和人员等目标感知功能的性能;2. It can effectively improve the accuracy, speed and stability of panoramic camera depth estimation, and improve the performance of unmanned docking bridge obstacle avoidance and target sensing functions such as people;
3、能够有效地降低数据处理和传输的计算量和带宽需求,适用于嵌入式设备和无线网络环境。3. It can effectively reduce the calculation amount and bandwidth requirements of data processing and transmission, and is suitable for embedded devices and wireless network environments.
附图说明Description of the drawings
为了更清楚地说明本发明实施例或现有技术中的技术方案,本发明由下述的具体实施及附图作以详细描述。In order to more clearly illustrate the embodiments of the present invention or technical solutions in the prior art, the present invention is described in detail with the following specific implementations and drawings.
图1为本发明实施例的一种面向无人登机桥避障的全景相机和激光雷达融合系统的结构图;Figure 1 is a structural diagram of a panoramic camera and lidar fusion system for unmanned boarding bridge obstacle avoidance according to an embodiment of the present invention;
图2为本发明实施例的全景相机模块示意图;Figure 2 is a schematic diagram of a panoramic camera module according to an embodiment of the present invention;
图3为本发明实施例的一种面向无人登机桥避障的全景相机和激光雷达融合系统的步骤流程图;Figure 3 is a step flow chart of a panoramic camera and lidar fusion system for unmanned boarding bridge obstacle avoidance according to an embodiment of the present invention;
附图标记说明:硬件系统1、软件系统2;Explanation of reference signs: hardware system 1, software system 2;
全景相机模块11、激光传感器模块12;Panoramic camera module 11, laser sensor module 12;
数据预处理模块21、特征提取模块22、特征匹配模块23、数据融合模块24、图像分割模块25、目标识别模块26、目标定位模块27。Data preprocessing module 21, feature extraction module 22, feature matching module 23, data fusion module 24, image segmentation module 25, target recognition module 26, target positioning module 27.
具体实施方式Detailed ways
为使本发明的目的、技术方案和优点更加清楚明了,下面通过附图中示出的具体实施例来描述本发明。但是应该理解,这些描述只是示例性的,而并非要限制本发明的范围。此外,在以下说明中,省略了对公知结构和技术的描述,以避免不必要地混淆本发明的概念。In order to make the purpose, technical solutions and advantages of the present invention clearer, the present invention is described below through the specific embodiments shown in the drawings. However, it should be understood that these descriptions are exemplary only and are not intended to limit the scope of the invention. Furthermore, in the following description, descriptions of well-known structures and techniques are omitted to avoid unnecessarily confusing the concepts of the present invention.
在此,还需要说明的是,为了避免因不必要的细节而模糊了本发明,在附图中仅仅示出了与根据本发明的方案密切相关的结构和/或处理步骤,而省略了与本发明关系不大的其他细节。Here, it should also be noted that, in order to avoid obscuring the present invention with unnecessary details, only the structures and/or processing steps closely related to the solution according to the present invention are shown in the drawings, and the details related to them are omitted. Other details are less relevant to the invention.
如图1-图3所示,本一种面向无人登机桥避障的全景相机和激光雷达融合系统包含硬件系统1和软件系统2;As shown in Figures 1-3, this panoramic camera and lidar fusion system for unmanned boarding bridge obstacle avoidance includes hardware system 1 and software system 2;
其中硬件系统1包含全景相机模块11和激光传感器模块12;The hardware system 1 includes a panoramic camera module 11 and a laser sensor module 12;
所述软件系统2包含数据预处理模块21、特征提取模块22、特征匹配模块23、数据融合模块24、图像分割模块25、目标识别模块26和目标定位模块27。The software system 2 includes a data preprocessing module 21 , a feature extraction module 22 , a feature matching module 23 , a data fusion module 24 , an image segmentation module 25 , a target recognition module 26 and a target positioning module 27 .
所述全景相机模块11和激光传感器模块21均与数据预处理模块21连接,数据预处理模块21与特征提取模块22、特征匹配模块23、数据融合模块24、图像分割模块25、目标识别模块26和目标定位模块27之间相互连接。The panoramic camera module 11 and the laser sensor module 21 are both connected to the data preprocessing module 21, and the data preprocessing module 21 is connected to the feature extraction module 22, feature matching module 23, data fusion module 24, image segmentation module 25, and target recognition module 26 and the target positioning module 27 are connected to each other.
所述全景相机模块11为双全景相机,双全景相机可以安装在无人对接廊桥的前后两端,分别对应前后两个观察圆。The panoramic camera module 11 is a dual panoramic camera. The dual panoramic cameras can be installed at the front and rear ends of the unmanned docking bridge, corresponding to the front and rear observation circles respectively.
作为一种优选的实施例,所述双全景相机由六个120度鱼眼相机组成,用于采集环境图像数据。其中,每个鱼眼相机可以采集120度的视场角,三个鱼眼相机可以拼接成一个360度的全景相机,两个全景相机可以构成一个双目全景相机,实现对360度视场的覆盖并估计深度。As a preferred embodiment, the dual panoramic cameras are composed of six 120-degree fisheye cameras and are used to collect environmental image data. Among them, each fisheye camera can capture a 120-degree field of view, three fisheye cameras can be spliced into a 360-degree panoramic camera, and two panoramic cameras can form a binocular panoramic camera to achieve a 360-degree field of view. Cover and estimate depth.
所述激光传感器模12块由一个旋转式多线激光雷达组成,用于采集环境距离数据。该激光雷达可以在水平方向上旋转,并在垂直方向上发射多条激光束,形成一个扫描平面。该扫描平面可以与双全景相机的观察圆平行,并与之有一定的距离。该激光雷达可以采集到环境中物体表面的稀疏点云数据。The laser sensor module 12 consists of a rotating multi-line laser radar and is used to collect environmental distance data. The lidar can rotate in the horizontal direction and emit multiple laser beams in the vertical direction to form a scanning plane. This scanning plane can be parallel to, and at a certain distance from, the observation circle of the dual panoramic camera. This lidar can collect sparse point cloud data on the surface of objects in the environment.
所述数据预处理模块21用于对全景相机和激光传感器的数据进行预处理,包括去噪、滤波、校正等操作,提高数据的质量和一致性。具体地说,该模块可以对鱼眼相机采集到的扭曲图像进行畸变校正,使之恢复为正常视角;对多个鱼眼相机采集到的图像进行标定和拼接,使之形成一个完整的全景图像;对双全景相机采集到的前后两幅全景图像进行立体匹配和深度估计,使之形成一个完整的RGB-D图像;对激光雷达采集到的点云数据进行去噪、滤波、变换等操作,使之与RGB-D图像在同一坐标系下,并与之对齐。The data preprocessing module 21 is used to preprocess data from panoramic cameras and laser sensors, including denoising, filtering, correction and other operations, to improve the quality and consistency of the data. Specifically, this module can perform distortion correction on distorted images collected by fisheye cameras to restore them to normal viewing angles; calibrate and splice images collected by multiple fisheye cameras to form a complete panoramic image ; Perform stereo matching and depth estimation on the front and rear panoramic images collected by the dual panoramic cameras to form a complete RGB-D image; perform denoising, filtering, transformation and other operations on the point cloud data collected by the lidar, Make it in the same coordinate system as the RGB-D image and align it with it.
所述特征提取模块22用于对全景相机和激光传感器的数据进行特征提取,并将特征表示为高维向量。The feature extraction module 22 is used to extract features from the data of the panoramic camera and the laser sensor, and represent the features as high-dimensional vectors.
作为一种优选的实施例,该模块可以利用卷积神经网络(CNN)或其他深度学习方法,对RGB-D图像进行特征提取,提取出图像中物体表面的边缘、纹理、颜色等特征,并将特征表示为高维向量;该模块还可以利用点云处理方法,对点云数据进行特征提取,提取出点云中物体表面的形状、大小、方向等特征,并将特征表示为高维向量。As a preferred embodiment, this module can use convolutional neural network (CNN) or other deep learning methods to perform feature extraction on RGB-D images, extract the edges, texture, color and other features of the object surface in the image, and Represent features as high-dimensional vectors; this module can also use point cloud processing methods to extract features from point cloud data, extract the shape, size, direction and other features of the object surface in the point cloud, and represent the features as high-dimensional vectors .
所述特征匹配模块23用于对全景相机和激光传感器的特征向量进行匹配,并计算相似度。具体地说,该模块可以利用相似度度量方法,如余弦相似度、欧氏距离等,计算RGB-D图像中每个像素点的特征向量与点云数据中最近邻点的特征向量之间的相似度,并将相似度作为匹配的依据。该模块还可以利用阈值或其他筛选方法,剔除相似度低于一定值的匹配对,提高匹配的准确性。The feature matching module 23 is used to match the feature vectors of the panoramic camera and the laser sensor and calculate the similarity. Specifically, this module can use similarity measurement methods, such as cosine similarity, Euclidean distance, etc., to calculate the difference between the feature vector of each pixel in the RGB-D image and the feature vector of the nearest neighbor point in the point cloud data. similarity, and use similarity as the basis for matching. This module can also use thresholds or other screening methods to eliminate matching pairs whose similarity is lower than a certain value to improve the accuracy of matching.
所述数据融合模块24用于对匹配的特征向量进行融合,并将融合后的特征向量表示为一个统一的特征向量。具体地说,该模块可以利用加权平均或其他数据融合方法,将RGB-D图像中每个像素点的特征向量与其匹配的点云数据中最近邻点的特征向量进行加权平均,得到一个融合后的特征向量。该模块还可以利用归一化或其他标准化方法,使得融合后的特征向量具有一定的范围和分布,便于后续处理。The data fusion module 24 is used to fuse matching feature vectors and represent the fused feature vectors as a unified feature vector. Specifically, this module can use weighted average or other data fusion methods to perform a weighted average of the feature vector of each pixel in the RGB-D image and the feature vector of its nearest neighbor point in the matching point cloud data to obtain a fused eigenvector. This module can also use normalization or other standardization methods to make the fused feature vectors have a certain range and distribution to facilitate subsequent processing.
所述图像分割模块25用于对融合后的特征向量进行分割,并将分割后的区域表示为不同的标签。具体地说,该模块可以利用聚类或其他图像分割方法,将融合后的特征向量分割为不同的区域,每个区域对应一个目标或背景。该模块还可以利用边缘检测或其他边界提取方法,提取出每个区域的边界,并将边界表示为一组连续或离散的点。The image segmentation module 25 is used to segment the fused feature vectors and represent the segmented areas as different labels. Specifically, this module can use clustering or other image segmentation methods to segment the fused feature vector into different regions, each region corresponding to a target or background. The module can also use edge detection or other boundary extraction methods to extract the boundaries of each region and represent the boundaries as a set of continuous or discrete points.
所述目标识别模块26用于对分割后的区域进行识别,并判断区域的类别。具体地说,该模块可以利用分类或其他目标识别方法,根据每个区域的特征向量,判断区域属于哪一类目标,如人、车、障碍物等。该模块还可以利用标签或其他标记方法,给每个区域赋予一个唯一的标识符,并在图像上显示出来。The target recognition module 26 is used to identify the segmented areas and determine the category of the areas. Specifically, this module can use classification or other target recognition methods to determine which type of target the area belongs to based on the feature vector of each area, such as people, cars, obstacles, etc. The module can also use tags or other marking methods to give each area a unique identifier and display it on the image.
所述目标定位模块27用于对识别后的目标进行定位,并计算目标的位置、姿态、速度等参数。具体地说,该模块可以利用回归或其他目标定位方法,根据每个目标区域的特征向量和边界点,计算目标在空间中的位置、姿态、速度等参数。该模块还可以利用箭头或其他指示方法,在图像上显示出目标的运动方向和速度大小The target positioning module 27 is used to position the identified target and calculate the position, attitude, speed and other parameters of the target. Specifically, this module can use regression or other target positioning methods to calculate the position, attitude, speed and other parameters of the target in space based on the feature vectors and boundary points of each target area. This module can also use arrows or other indication methods to display the target's movement direction and speed on the image.
如图3所示,一种面向无人登机桥避障的全景相机和激光雷达融合方法,其步骤包括:As shown in Figure 3, a panoramic camera and lidar fusion method for unmanned boarding bridge obstacle avoidance includes the following steps:
S1、对全景相机和激光传感器的数据进行预处理,包括去噪、滤波、校正,提高数据的质量和一致性;S1. Preprocess data from panoramic cameras and laser sensors, including denoising, filtering, and correction to improve data quality and consistency;
其中,对鱼眼相机采集到的扭曲图像进行畸变校正时,使用以下公式:Among them, when correcting the distortion of the distorted image collected by the fisheye camera, the following formula is used:
x'=x(1+k1r2+k2r4+k3r6)x'=x(1+k1r 2 +k2r 4 +k3r 6 )
y'=y(1+k1r2+k2r4+k3r6)y'=y(1+k1r 2 +k2r 4 +k3r 6 )
上述公式中,x和y是扭曲图像上的坐标,x'和y'是校正后图像上的坐标,r是扭曲图像上点到中心点的距离,k1、k2和k3是畸变系数。In the above formula, x and y are the coordinates on the distorted image, x' and y' are the coordinates on the corrected image, r is the distance from the point on the distorted image to the center point, k1, k2 and k3 are the distortion coefficients.
S2、对全景相机的环境图像进行特征提取,利用卷积神经网络(CNN)或其他深度学习方法,提取图像的边缘、纹理、颜色等特征,并将特征表示为高维向量;S2. Extract features from the environment image of the panoramic camera, use convolutional neural network (CNN) or other deep learning methods to extract features such as edges, texture, and color of the image, and represent the features as high-dimensional vectors;
其中对RGB-D图像进行特征提取时,使用以下公式:When extracting features from RGB-D images, the following formula is used:
f=W*x+bf=W*x+b
上述公式中,f是特征向量,W是卷积核,x是RGB-D图像,b是偏置项。In the above formula, f is the feature vector, W is the convolution kernel, x is the RGB-D image, and b is the bias term.
S3、对激光传感器的距离数据进行特征提取,利用点云处理方法,提取距离数据的形状、大小、方向特征,并将特征表示为高维向量;S3. Extract features from the distance data of the laser sensor, use point cloud processing methods to extract the shape, size, and direction features of the distance data, and represent the features as high-dimensional vectors;
其中对点云数据进行特征提取时,使用以下公式:When extracting features from point cloud data, the following formula is used:
f=g(h(p))f=g(h(p))
上述公式中,f是特征向量,g是非线性激活函数,h是点云处理函数,p是点云数据。In the above formula, f is the feature vector, g is the nonlinear activation function, h is the point cloud processing function, and p is the point cloud data.
S4、对全景相机和激光传感器的特征向量进行匹配,利用相似度度量方法,计算不同传感器之间的特征向量;S4. Match the feature vectors of the panoramic camera and the laser sensor, and use the similarity measurement method to calculate the feature vectors between different sensors;
其中计算相似度时,使用以下公式:When calculating similarity, the following formula is used:
s=cos(f1,f2)=f1*f2/(|f1|*|f2|)s=cos(f1,f2)=f1*f2/(|f1|*|f2|)
上述公式中,s是相似度,cos是余弦函数,f1和f2是不同传感器的特征向量。In the above formula, s is the similarity, cos is the cosine function, f1 and f2 are the feature vectors of different sensors.
S5、对匹配的特征向量进行融合,利用加权平均或其他数据融合方法,将不同传感器的特征向量融合为一个统一的特征向量,反映环境的图像和距离信息;S5. Fusion of matching feature vectors, using weighted average or other data fusion methods, to fuse the feature vectors of different sensors into a unified feature vector that reflects the image and distance information of the environment;
其中,进行数据融合时,使用以下公式:Among them, when performing data fusion, the following formula is used:
F=a*f1+(1-a)*f2F=a*f1+(1-a)*f2
上述公式中,F是融合后的特征向量,a是权重系数,f1和f2是不同传感器的特征向量。In the above formula, F is the fused feature vector, a is the weight coefficient, f1 and f2 are the feature vectors of different sensors.
S6、对融合的特征向量进行分割,利用聚类或其他图像分割方法,将融合的特征向量分割为不同的区域,每个区域对应一个目标或背景;S6. Segment the fused feature vector and use clustering or other image segmentation methods to divide the fused feature vector into different areas, each area corresponding to a target or background;
其中,进行图像分割时,使用以下公式:Among them, when performing image segmentation, the following formula is used:
L=argmin(S(F))L=argmin(S(F))
上述公式中,L是分割后的标签,S是图像分割函数,F是融合后的特征向量。In the above formula, L is the segmented label, S is the image segmentation function, and F is the fused feature vector.
S7、对分割的区域进行识别,利用分类或其他目标识别方法,根据区域的特征向量,判断区域的类别,如人、车、障碍物等;S7. Identify the segmented areas, use classification or other target recognition methods, and determine the category of the area based on the feature vector of the area, such as people, vehicles, obstacles, etc.;
其中,进行目标识别时,使用以下公式:Among them, when performing target recognition, the following formula is used:
C=softmax(W*f+b)C=softmax(W*f+b)
上述公式中,C是目标类别,softmax是归一化指数函数,W是分类权重,f是区域的特征向量,b是分类偏置项。In the above formula, C is the target category, softmax is the normalized exponential function, W is the classification weight, f is the feature vector of the region, and b is the classification bias term.
S8、对识别的目标进行定位,利用回归或其他目标定位方法,根据区域的特征向量,计算目标的位置、姿态、速度等参数;S8. Position the identified target, use regression or other target positioning methods, and calculate the target's position, attitude, speed and other parameters based on the feature vector of the area;
其中,进行目标定位时,使用以下公式:Among them, when performing target positioning, the following formula is used:
P=W*f+bP=W*f+b
上述公式中,P是目标位置、姿态、速度等参数,W是回归权重,f是区域的特征向量,b是回归偏置项。In the above formula, P is the target position, attitude, speed and other parameters, W is the regression weight, f is the feature vector of the area, and b is the regression bias term.
对于本领域技术人员而言,显然本发明不限于上述示范性实施例的细节,而且在不背离本发明的精神或基本特征的情况下,能够以其他的具体形式实现本发明。因此,无论从哪一点来看,均应将实施例看作是示范性的,而且是非限制性的,本发明的范围由所附权利要求而不是上述说明限定,因此旨在将落在权利要求的等同要件的含义和范围内的所有变化囊括在本发明内。It is obvious to those skilled in the art that the present invention is not limited to the details of the above-described exemplary embodiments, and that the present invention can be implemented in other specific forms without departing from the spirit or essential characteristics of the present invention. Therefore, the embodiments should be regarded as illustrative and non-restrictive from any point of view, and the scope of the present invention is defined by the appended claims rather than the above description, and it is therefore intended that all claims falling within the claims All changes within the meaning and scope of equivalent elements are included in the present invention.
此外,应当理解,虽然本说明书按照实施方式加以描述,但并非每个实施方式仅包含一个独立的技术方案,说明书的这种叙述方式仅仅是为清楚起见,本领域技术人员应当将说明书作为一个整体,各实施例中的技术方案也可以经适当组合,形成本领域技术人员可以理解的其他实施方式。In addition, it should be understood that although this specification is described in terms of implementations, not each implementation only contains an independent technical solution. This description of the specification is only for the sake of clarity, and those skilled in the art should take the specification as a whole. , the technical solutions in each embodiment can also be appropriately combined to form other implementations that can be understood by those skilled in the art.
Claims (10)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310808228.2A CN117011656A (en) | 2023-07-03 | 2023-07-03 | Panoramic camera and laser radar fusion method for obstacle avoidance of unmanned boarding bridge |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310808228.2A CN117011656A (en) | 2023-07-03 | 2023-07-03 | Panoramic camera and laser radar fusion method for obstacle avoidance of unmanned boarding bridge |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117011656A true CN117011656A (en) | 2023-11-07 |
Family
ID=88573626
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310808228.2A Pending CN117011656A (en) | 2023-07-03 | 2023-07-03 | Panoramic camera and laser radar fusion method for obstacle avoidance of unmanned boarding bridge |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117011656A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118519449A (en) * | 2024-04-15 | 2024-08-20 | 哈尔滨工业大学(威海) | Multi-sensor fusion-based aircraft cabin door pose locking system and method |
CN119066313A (en) * | 2024-08-23 | 2024-12-03 | 内蒙古自治区民航机场集团有限责任公司 | An operation scheduling method for centralized control of unmanned boarding bridge |
-
2023
- 2023-07-03 CN CN202310808228.2A patent/CN117011656A/en active Pending
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118519449A (en) * | 2024-04-15 | 2024-08-20 | 哈尔滨工业大学(威海) | Multi-sensor fusion-based aircraft cabin door pose locking system and method |
CN118519449B (en) * | 2024-04-15 | 2024-12-27 | 哈尔滨工业大学(威海) | Aircraft door posture locking system and method based on multi-sensor fusion |
CN119066313A (en) * | 2024-08-23 | 2024-12-03 | 内蒙古自治区民航机场集团有限责任公司 | An operation scheduling method for centralized control of unmanned boarding bridge |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112396650B (en) | Target ranging system and method based on fusion of image and laser radar | |
CN115439424B (en) | Intelligent detection method for aerial video images of unmanned aerial vehicle | |
CN110221603B (en) | Remote obstacle detection method based on laser radar multi-frame point cloud fusion | |
CN106681353B (en) | Obstacle avoidance method and system for UAV based on binocular vision and optical flow fusion | |
WO2020135446A1 (en) | Target positioning method and device and unmanned aerial vehicle | |
CN109949361A (en) | An Attitude Estimation Method for Rotor UAV Based on Monocular Vision Positioning | |
CN113568435B (en) | Unmanned aerial vehicle autonomous flight situation perception trend based analysis method and system | |
CN117011656A (en) | Panoramic camera and laser radar fusion method for obstacle avoidance of unmanned boarding bridge | |
CN111968128A (en) | Unmanned aerial vehicle visual attitude and position resolving method based on image markers | |
CN106774296A (en) | A kind of disorder detection method based on laser radar and ccd video camera information fusion | |
CN113506318A (en) | A 3D object perception method in vehicle edge scene | |
CN111815717A (en) | A Multi-sensor Fusion External Reference Joint Semi-autonomous Calibration Method | |
EP4213128A1 (en) | Obstacle detection device, obstacle detection system, and obstacle detection method | |
CN115376109B (en) | Obstacle detection method, obstacle detection device, and storage medium | |
CN110083177A (en) | A kind of quadrotor and control method of view-based access control model landing | |
CN116978009A (en) | Dynamic object filtering method based on 4D millimeter wave radar | |
CN114120283A (en) | Method for distinguishing unknown obstacles in road scene three-dimensional semantic segmentation | |
CN109764864B (en) | A method and system for indoor UAV pose acquisition based on color recognition | |
CN112733678A (en) | Ranging method, ranging device, computer equipment and storage medium | |
CN110660113A (en) | Method and device for establishing characteristic map, acquisition equipment and storage medium | |
CN114998436A (en) | Object labeling method and device, electronic equipment and storage medium | |
CN110197104B (en) | Distance measurement method and device based on vehicle | |
CN114812558A (en) | Monocular vision unmanned aerial vehicle autonomous positioning method combined with laser ranging | |
Li et al. | Edgecalib: Multi-frame weighted edge features for automatic targetless lidar-camera calibration | |
CN117152210B (en) | Image dynamic tracking method and related device based on dynamic observation field angle |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |