CN116989772B - An air-ground multi-modal multi-agent collaborative positioning and mapping method - Google Patents
An air-ground multi-modal multi-agent collaborative positioning and mapping method Download PDFInfo
- Publication number
- CN116989772B CN116989772B CN202311248105.4A CN202311248105A CN116989772B CN 116989772 B CN116989772 B CN 116989772B CN 202311248105 A CN202311248105 A CN 202311248105A CN 116989772 B CN116989772 B CN 116989772B
- Authority
- CN
- China
- Prior art keywords
- perspective
- map
- imu
- ground
- measurement data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/38—Electronic maps specially adapted for navigation; Updating thereof
- G01C21/3804—Creation or updating of map data
- G01C21/3833—Creation or updating of map data characterised by the source of data
- G01C21/3841—Data obtained from two or more sources, e.g. probe vehicles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/005—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 with correlation of navigation data from several sources, e.g. map or contour matching
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/10—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration
- G01C21/12—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning
- G01C21/16—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation
- G01C21/165—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/10—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration
- G01C21/12—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning
- G01C21/16—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation
- G01C21/165—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments
- G01C21/1656—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments with passive imaging devices, e.g. cameras
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/38—Electronic maps specially adapted for navigation; Updating thereof
- G01C21/3804—Creation or updating of map data
- G01C21/3807—Creation or updating of map data characterised by the type of data
- G01C21/3826—Terrain data
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Automation & Control Theory (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
Abstract
本发明提出一种空地多模态多智能体协同定位与建图方法,包括,获取智能体的测量数据,智能体包括无人机和无人车,无人车设置有视觉标志物;通过无人车的测量数据进行局部视角局部建图,通过无人机的测量数据进行全局视角局部建图;通过无人机对视觉标志物进行检测,当检测成功时获取无人机和无人车之间的相对位姿,利用相对位姿的转换关系进行空地视角地图融合与优化;基于回环不断检测智能体是否经过重叠区域,当检测到重叠区域时,通过匹配的关键帧建立智能体的两幅地图的关联,进行轨迹校准和相近视角地图融合与优化;根据空地视角融合与优化后的地图和相近视角融合与优化后的地图得到位姿轨迹和全局一致的地图。
The invention proposes a multi-modal multi-agent collaborative positioning and mapping method in the air and ground, which includes obtaining measurement data of the intelligent agents. The intelligent agents include drones and unmanned vehicles, and the unmanned vehicles are equipped with visual markers; The measurement data of people and vehicles are used for local mapping from a local perspective, and the measurement data of drones are used for local mapping from a global perspective; visual markers are detected through drones, and when the detection is successful, the relationship between drones and unmanned vehicles is obtained. The relative pose between the objects is used to fuse and optimize the air-ground perspective map using the conversion relationship of the relative pose; based on loop closure, it is continuously detected whether the agent passes through the overlapping area. When the overlapping area is detected, two images of the agent are established through matching key frames. Map correlation, trajectory calibration and similar perspective map fusion and optimization are performed; based on the map fused and optimized from the air and ground perspective and the map fused and optimized from the similar perspective, a map with consistent pose trajectory and global consistency is obtained.
Description
技术领域Technical field
本发明属于无人系统的定位建图领域。The invention belongs to the field of positioning and mapping of unmanned systems.
背景技术Background technique
随着智能机器人领域技术的不断突破,智能机器人在现代社会被越来越多地应用于解决复杂问题,而自主导航能力被认为是智能机器人实现自主任务的基础。为了实现机器人自主规划导航,如何使智能机器人协同进行定位与建图(simultaneouslocalizationand mapping, SLAM)成为目前的研究热门课题,特别是在军事, 灾难环境搜救等领域拥有巨大的潜力。With the continuous technological breakthroughs in the field of intelligent robots, intelligent robots are increasingly used to solve complex problems in modern society, and autonomous navigation capabilities are considered to be the basis for intelligent robots to achieve autonomous tasks. In order to realize robot autonomous planning and navigation, how to make intelligent robots coordinate positioning and mapping (SLAM) has become a hot research topic at present, especially in the fields of military, disaster environment search and rescue, etc. It has huge potential.
目前, 多机器人之间的互相位姿校正和信息感知成为一个难点。在进行SLAM过程中, 机器人无法预先获取整个环境的信息,地图的构建主要靠机器人的探索。传统的单机器人在大范围场景下进行SLAM存在许多缺陷,例如传感器作用范围小、观测角度局限、计算复杂性高、存储能力弱等。只依靠局部传感器信息,机器人定位误差会越来越大,最终导致建图与定位偏移。而对于一些特殊任务,例如军事、灾难搜救等场合,需要尽可能短的时间完成任务。对于上述问题,通过分散于环境中的多个机器人通信和相互协调工作, 整个系统可以获得更强的环境探测能力和准确的定位能力。At present, mutual pose correction and information perception among multiple robots have become a difficulty. During the SLAM process, the robot cannot obtain information about the entire environment in advance, and the construction of the map mainly relies on the robot's exploration. Traditional single robots have many shortcomings in performing SLAM in large-scale scenarios, such as small sensor range, limited observation angle, high computational complexity, and weak storage capacity. Relying only on local sensor information, the robot's positioning error will become larger and larger, eventually leading to mapping and positioning deviations. For some special tasks, such as military, disaster search and rescue, etc., the task needs to be completed in the shortest possible time. For the above problems, through the communication and mutual coordination of multiple robots scattered in the environment, the entire system can obtain stronger environmental detection capabilities and accurate positioning capabilities.
发明内容Contents of the invention
本发明旨在至少在一定程度上解决相关技术中的技术问题之一。The present invention aims to solve one of the technical problems in the related art, at least to a certain extent.
为此,本发明的目的在于提出一种空地多模态多智能体协同定位与建图方法,用于构建出全局一致的环境地图模型。To this end, the purpose of the present invention is to propose an open-ground multi-modal multi-agent collaborative positioning and mapping method for constructing a globally consistent environmental map model.
为达上述目的,本发明第一方面实施例提出了一种空地多模态多智能体协同定位与建图方法,包括:In order to achieve the above purpose, the first embodiment of the present invention proposes an air-ground multi-modal multi-agent collaborative positioning and mapping method, including:
获取智能体的测量数据,所述智能体包括无人机和无人车,所述无人车设置有视觉标志物;Obtaining measurement data of an intelligent agent, the intelligent agent includes a drone and an unmanned vehicle, and the unmanned vehicle is provided with a visual marker;
通过所述无人车的测量数据进行局部视角局部建图,通过所述无人机的测量数据进行全局视角局部建图;Local mapping from a local perspective is performed using the measurement data of the unmanned vehicle, and local mapping from a global perspective is performed using the measurement data of the UAV;
通过所述无人机对所述视觉标志物进行检测,当检测成功时获取所述无人机和所述无人车之间的相对位姿,利用所述相对位姿的转换关系进行空地视角地图融合与优化;The visual marker is detected by the drone, and when the detection is successful, the relative pose between the drone and the unmanned vehicle is obtained, and the conversion relationship between the relative pose is used to perform an air-ground perspective. Map fusion and optimization;
基于回环不断检测所述智能体是否经过重叠区域,当检测到所述重叠区域时,通过匹配的关键帧建立所述智能体的两幅地图的关联,进行轨迹校准和相近视角地图融合与优化;Continuously detect whether the agent passes through the overlapping area based on loop closure. When the overlapping area is detected, the association between the two maps of the agent is established through matching key frames, and trajectory calibration and similar perspective map fusion and optimization are performed;
根据所述空地视角融合与优化后的地图和所述相近视角融合与优化后的地图得到位姿轨迹和全局一致的地图。According to the map after the fusion and optimization of the air and ground perspectives and the map after the fusion and optimization of the similar perspective, a map with consistent pose trajectory and global consistency is obtained.
另外,根据本发明上述实施例的一种空地多模态多智能体协同定位与建图方法还可以具有以下附加的技术特征:In addition, an air-ground multi-modal multi-agent collaborative positioning and mapping method according to the above embodiments of the present invention may also have the following additional technical features:
进一步地,在本发明的一个实施例中,所述获取智能体的测量数据,包括:Further, in one embodiment of the present invention, obtaining the measurement data of the intelligent agent includes:
对所述测量数据进行预处理,包括视觉检测和光流跟踪、惯性测量单元IMU预积分;其中,Preprocess the measurement data, including visual detection and optical flow tracking, and inertial measurement unit IMU pre-integration; where,
所述视觉检测和光流跟踪包括:应用最小二乘法求得光流的速度矢量,如下式,The visual detection and optical flow tracking include: applying the least squares method to obtain the velocity vector of the optical flow, as follows:
, ,
其中,、/>表示图像中像素点亮度在/>、/>方向上的图像梯度,/>表示在/>方向上的时间梯度,/>、/>为光流沿/>、/>轴的速度矢量;in, ,/> Indicates that the brightness of the pixels in the image is/> ,/> Image gradient in direction, /> Shown in/> Time gradient in direction,/> ,/> is the optical flow edge/> ,/> axis velocity vector;
所述惯性测量单元IMU预积分包括:The inertial measurement unit IMU pre-integration includes:
, ,
其中,为IMU坐标系,/>为原点初始化时IMU所在的坐标系即世界坐标系,/>和/>是 IMU所测得的加速度和角速度,/>为/>时刻从IMU坐标系到世界坐标系的旋转,/>为四元数右乘;in, is the IMU coordinate system,/> The coordinate system where the IMU is located when initializing the origin is the world coordinate system, /> and/> are the acceleration and angular velocity measured by the IMU,/> for/> Time rotation from IMU coordinate system to world coordinate system,/> It is the right multiplication of the quaternion;
将第帧到第/>帧之间的所有IMU数据进行积分,即可得到第/>帧的位置/>、速度/>和旋转/>,/>作为视觉估计的初始值,旋转为四元数形式。General Frame to/> By integrating all the IMU data between frames, we can get the Frame position/> , speed/> and rotate/> ,/> As an initial value for visual estimation, the rotation is in quaternion form.
进一步地,在本发明的一个实施例中,所述通过所述无人车的测量数据进行局部视角局部建图,通过所述无人机的测量数据进行全局视角局部建图,包括:Further, in one embodiment of the present invention, the local perspective mapping is performed using the measurement data of the unmanned vehicle, and the global perspective local mapping is performed using the measurement data of the unmanned aerial vehicle, including:
用SFM求解滑动窗口内所有帧的位姿与所有路标点的三维位置,与IMU 预积分值进行对齐,得到角速度偏置、重力方向、尺度因子和每一帧所对应的速度;Use SFM to solve the poses of all frames and the three-dimensional positions of all landmark points in the sliding window, and align them with the IMU pre-integrated values to obtain the angular velocity offset, gravity direction, scale factor and velocity corresponding to each frame;
使用滑动窗口法优化窗口内的状态变量,在时刻窗口中的优化的状态向量/>如下式,Use the sliding window method to optimize the state variables within the window, in Optimized state vector in time window/> As follows,
, ,
其中,和/>为相机位姿的旋转和平移部分,/>为相机在世界坐标系下的速度,/>和/>分别为IMU的加速度偏置和角速度偏置;系统的状态量的优化目标函数,如下式,in, and/> is the rotation and translation part of the camera pose,/> is the speed of the camera in the world coordinate system,/> and/> are the acceleration bias and angular velocity bias of the IMU respectively; the optimization objective function of the system's state quantity is as follows:
, ,
其中,为最大估计后验值,/>为滑窗初值残差,/>为IMU观测残差,/>为相机观测残差。in, is the maximum estimated posterior value,/> is the sliding window initial value residual,/> is the IMU observation residual,/> is the camera observation residual.
进一步地,在本发明的一个实施例中,所述基于回环不断检测所述智能体是否经过重叠区域,包括:Further, in one embodiment of the present invention, continuously detecting whether the agent passes through an overlapping area based on loop closure includes:
将所有视觉特征进行聚类,一类特征是一个单词,所有特征是一个字典;All visual features are clustered, one type of feature is a word, and all features are a dictionary;
用单个向量描述一个图像:Describe an image with a single vector:
计算两张图像A和B之间的相似度:Calculate the similarity between two images A and B :
, ,
其中,是描述图像A的向量,/>是描述图像B的向量,/>是两张图像A和B之间的相似度,/>表示描述图像A的向量/>的第/>个分量,/>表示描述图像B的向量的第/>个分量;in, is a vector describing image A, /> is a vector describing image B, /> is the similarity between two images A and B, /> Represents a vector describing image A /> of/> components,/> represents a vector describing image B of/> weight;
若所述相似度超过阈值,则认为出现回环。If the similarity exceeds the threshold, a loop is considered to have occurred.
进一步地,在本发明的一个实施例中,所述将所有视觉特征进行聚类,包括:Further, in one embodiment of the present invention, clustering all visual features includes:
在根节点,用K-means算法把所有样本聚成类,得到第一层;At the root node, use the K-means algorithm to cluster all samples into Class, get the first layer;
对所述第一层的每个节点,把属于该节点的样本再聚成类,得到下一层;依此类推,最后得到叶子层,所述叶子层即为单词。For each node in the first layer, the samples belonging to the node are clustered into Class, the next layer is obtained; and so on, finally the leaf layer is obtained, and the leaf layer is the word.
进一步地,在本发明的一个实施例中,所述用单个向量描述一个图像,包括:Further, in one embodiment of the present invention, describing an image with a single vector includes:
定义为单词/>包含的特征数量,/>为所有单词包含的特征数量,/>为单词在图像A中出现的次数,/>为单词/>在所有图像中一共出现的次数,则definition for word/> The number of features included,/> is the number of features contained in all words,/> for words The number of times it appears in image A,/> for word/> The total number of occurrences in all images, then
单词在图像A的权重/>为:/>,word weights in image A/> for:/> ,
通过词袋,用单个向量描述一个图像A:via bag-of-words, using a single vector Describe an image A:
, ,
其中,是图像A在字典中所具有的单词,/>是所对应的权重,/>是描述图像A的向量。in, is the word that image A has in the dictionary,/> yes The corresponding weight,/> is a vector describing image A.
为达上述目的,本发明第二方面实施例提出了一种空地多模态多智能体协同定位与建图装置,包括以下模块:In order to achieve the above purpose, the second embodiment of the present invention proposes an air-ground multi-modal multi-agent collaborative positioning and mapping device, which includes the following modules:
获取模块,用于获取智能体的测量数据,所述智能体包括无人机和无人车,所述无人车设置有视觉标志物;An acquisition module, used to acquire measurement data of an intelligent agent. The intelligent agent includes a drone and an unmanned vehicle, and the unmanned vehicle is provided with a visual marker;
建图模块,用于通过所述无人车的测量数据进行局部视角局部建图,通过所述无人机的测量数据进行全局视角局部建图;A mapping module, configured to perform local mapping from a local perspective using the measurement data of the unmanned vehicle, and perform local mapping from a global perspective using the measurement data of the drone;
空地视角融合模块,用于通过所述无人机对所述视觉标志物进行检测,当检测成功时获取所述无人机和所述无人车之间的相对位姿,利用所述相对位姿的转换关系进行空地视角地图融合与优化;An air-ground perspective fusion module is used to detect the visual marker through the drone, and when the detection is successful, obtain the relative pose between the drone and the unmanned vehicle, using the relative pose Perform air-ground perspective map fusion and optimization based on the posture conversion relationship;
相近视角融合模块,用于基于回环不断检测所述智能体是否经过重叠区域,当检测到所述重叠区域时,通过匹配的关键帧建立所述智能体的两幅地图的关联,进行轨迹校准和相近视角地图融合与优化;A similar perspective fusion module is used to continuously detect whether the agent passes through an overlapping area based on loop closure. When the overlapping area is detected, the association between the two maps of the agent is established through matching key frames, and trajectory calibration and Similar perspective map fusion and optimization;
输出模块,用于根据所述空地视角融合与优化后的地图和所述相近视角融合与优化后的地图得到位姿轨迹和全局一致的地图。An output module is used to obtain a map with pose trajectory and global consistency based on the map after fusion and optimization of the air-ground perspective and the map after fusion and optimization of the similar perspective.
为达上述目的,本发明第三方面实施例提出了一种计算机设备,其特征在于,包括存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时,实现如上所述的一种空地多模态多智能体协同定位与建图方法。In order to achieve the above object, a third embodiment of the present invention provides a computer device, which is characterized in that it includes a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the computer program, it implements an air-ground multi-modal multi-agent collaborative positioning and mapping method as described above.
为达上述目的,本发明第四方面实施例提出了一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如上所述的一种空地多模态多智能体协同定位与建图方法。In order to achieve the above object, a fourth embodiment of the present invention provides a computer-readable storage medium on which a computer program is stored, which is characterized in that when the computer program is executed by a processor, the above-mentioned open space is implemented Multi-modal multi-agent collaborative localization and mapping method.
本发明实施例提出的空地多模态多智能体协同定位与建图方法,适用于多种应用场景中,并且显著提升单一机器人系统的感知能力与工作效率。尤其在搜索和救援领域中,空中无人机(UnmannedAerial Vehicle,UAV)以其空中视角的优势探测未知环境地形,引导地面机器人 (Unmanned ground vehicle,UGV)驶入目标区域实施精准的救援任务,以相较人力高效而低代价的方式实现搜索与营救。其中空中无人机拥有地面全局视野,结合地面机器人局部环境感知能力,能构建出全局一致的环境地图模型,此地图模型将为地面机器人提供全局的导航必备信息。The open-ground multi-modal multi-agent collaborative positioning and mapping method proposed by the embodiment of the present invention is suitable for a variety of application scenarios, and significantly improves the perception ability and work efficiency of a single robot system. Especially in the field of search and rescue, aerial drones (Unmanned Aerial Vehicle, UAV) use their aerial perspective to detect unknown environmental terrain, and guide ground robots (Unmanned ground vehicle, UGV) to drive into the target area to carry out precise rescue missions. Search and rescue can be achieved in an efficient and low-cost way compared to manpower. Among them, aerial drones have a global field of view on the ground. Combined with the local environment perception capabilities of ground robots, they can build a globally consistent environmental map model. This map model will provide global navigation information for ground robots.
附图说明Description of drawings
本发明上述的和/或附加的方面和优点从下面结合附图对实施例的描述中将变得明显和容易理解,其中:The above and/or additional aspects and advantages of the present invention will become apparent and readily understood from the following description of the embodiments in conjunction with the accompanying drawings, in which:
图1为本发明实施例所提供的一种空地多模态多智能体协同定位与建图方法的流程示意图。Figure 1 is a schematic flowchart of an air-ground multi-modal multi-agent collaborative positioning and mapping method provided by an embodiment of the present invention.
图2为本发明实施例所提供的一种空地多模态多智能体协同定位与建图系统示意图。Figure 2 is a schematic diagram of an air-ground multi-modal multi-agent collaborative positioning and mapping system provided by an embodiment of the present invention.
图3为本发明实施例所提供的一种空地多模态多智能体协同定位与建图装置的示意图。Figure 3 is a schematic diagram of an air-ground multi-modal multi-agent collaborative positioning and mapping device provided by an embodiment of the present invention.
具体实施方式Detailed ways
下面详细描述本发明的实施例,所述实施例的示例在附图中示出,其中自始至终相同或类似的标号表示相同或类似的元件或具有相同或类似功能的元件。下面通过参考附图描述的实施例是示例性的,旨在用于解释本发明,而不能理解为对本发明的限制。Embodiments of the present invention are described in detail below, examples of which are illustrated in the accompanying drawings, wherein the same or similar reference numerals throughout represent the same or similar elements or elements with the same or similar functions. The embodiments described below with reference to the drawings are exemplary and are intended to explain the present invention and are not to be construed as limiting the present invention.
下面参考附图描述本发明实施例的空地多模态多智能体协同定位与建图方法。The following describes the air-ground multi-modal multi-agent collaborative positioning and mapping method according to the embodiment of the present invention with reference to the accompanying drawings.
图1为本发明实施例所提供的一种空地多模态多智能体协同定位与建图方法的流程示意图。Figure 1 is a schematic flowchart of an air-ground multi-modal multi-agent collaborative positioning and mapping method provided by an embodiment of the present invention.
如图1所示,该空地多模态多智能体协同定位与建图方法包括以下步骤:As shown in Figure 1, the air-ground multi-modal multi-agent collaborative positioning and mapping method includes the following steps:
S101:获取智能体的测量数据,智能体包括无人机和无人车,无人车设置有视觉标志物;S101: Obtain the measurement data of the intelligent agent. The intelligent agent includes a drone and an unmanned vehicle. The unmanned vehicle is equipped with visual markers;
进一步地,在本发明的一个实施例中,获取智能体的测量数据,包括:Further, in one embodiment of the present invention, obtaining the measurement data of the agent includes:
对测量数据进行预处理,包括视觉检测和光流跟踪、惯性测量单元IMU预积分;其中,Preprocess the measurement data, including visual detection and optical flow tracking, and inertial measurement unit IMU pre-integration; where,
视觉检测和光流跟踪包括:应用最小二乘法求得光流的速度矢量,如下式,Visual detection and optical flow tracking include: applying the least squares method to obtain the velocity vector of the optical flow, as follows:
, ,
其中,、/>表示图像中像素点亮度在/>、/>方向上的图像梯度,/>表示在/>方向上的时间梯度,/>、/>为光流沿/>、/>轴的速度矢量;in, ,/> Indicates that the brightness of the pixels in the image is/> ,/> Image gradient in direction, /> Shown in/> Time gradient in direction,/> ,/> is the optical flow edge/> ,/> axis velocity vector;
惯性测量单元IMU预积分包括:Inertial measurement unit IMU pre-integration includes:
, ,
其中,为IMU坐标系,/>为原点初始化时IMU所在的坐标系即世界坐标系,/>和/>是 IMU所测得的加速度和角速度,/>为/>时刻从IMU坐标系到世界坐标系的旋转,/>为四元数右乘;in, is the IMU coordinate system,/> The coordinate system where the IMU is located when initializing the origin is the world coordinate system, /> and/> are the acceleration and angular velocity measured by the IMU,/> for/> Time rotation from IMU coordinate system to world coordinate system,/> It is the right multiplication of the quaternion;
将第帧到第/>帧之间的所有IMU数据进行积分,即可得到第/>帧的位置/>、速度/>和旋转/>,/>作为视觉估计的初始值,旋转为四元数形式。General Frame to/> By integrating all the IMU data between frames, we can get the Frame position/> , speed/> and rotate/> ,/> As an initial value for visual estimation, the rotation is in quaternion form.
具体地,选取FAST特征进行特征提取和光流追踪。在每张新的图像中,现有的特征点被KLT 算法跟踪并检测新的特征点。为保证特征点的均匀分布,将图像划分为若干个大小完全相同的子区域,每个子区域最多提取10个FAST角点,保持每张图像的角点数量在一定范围内。室外场景相邻两帧图像之间位移较大且各像素的亮度值可能发生突变,对特征点的跟踪造成不良影响,因此需要对特征点进行离群值剔除后再将其投影到单位球面上。离群值剔除使用RANSAC算法,以实现在户外动态场景下较为鲁棒的光流追踪。Specifically, FAST features are selected for feature extraction and optical flow tracking. In each new image, existing feature points are tracked and new feature points are detected by the KLT algorithm. In order to ensure the uniform distribution of feature points, the image is divided into several sub-regions of the same size, and up to 10 FAST corner points are extracted from each sub-region to keep the number of corner points in each image within a certain range. In outdoor scenes, the displacement between two adjacent frames of images is large and the brightness value of each pixel may suddenly change, which will have a negative impact on the tracking of feature points. Therefore, it is necessary to remove outliers from the feature points and then project them onto the unit sphere. . Outlier elimination uses the RANSAC algorithm to achieve more robust optical flow tracking in outdoor dynamic scenes.
在视觉检测和跟踪的同时,进行IMU预积分。IMU响应快,不受成像质量影响,可估计绝对尺度的特性,对室外表面无结构的物体的视觉定位进行补充。如果在相机位姿估算时将IMU所有采样时刻所对应的全部位姿插入帧间进行优化,会降低程序运行效率,所以需进行IMU预积分处理,将高频率输出的加速度和角速度测量值转化为单个观测值,该测量值将在非线性迭代重新进行线性化,形成帧间状态量的约束因子。Perform IMU pre-integration while visual detection and tracking. The IMU responds quickly and is not affected by imaging quality. It can estimate the characteristics of absolute scale and supplement the visual positioning of objects with no structure on the outdoor surface. If all poses corresponding to all sampling moments of the IMU are inserted between frames for optimization during camera pose estimation, the program running efficiency will be reduced. Therefore, IMU pre-integration processing is required to convert the high-frequency output acceleration and angular velocity measurement values into A single observation value that will be relinearized in nonlinear iterations to form a constraint factor for the inter-frame state quantity.
S102:通过无人车的测量数据进行局部视角局部建图,通过无人机的测量数据进行全局视角局部建图;S102: Use the measurement data of unmanned vehicles to perform local mapping from a local perspective, and use the measurement data from UAVs to perform local mapping from a global perspective;
初始化模块恢复单目相机的尺度需对视觉信息和 IMU信息进行松耦合。首先,用SFM求解滑动窗口内所有帧的位姿与所有路标点的三维位置,再将其与之前求得的IMU 预积分值进行对齐,从而解出角速度偏置、重力方向、尺度因子和每一帧所对应的速度。 随着系统的运行,状态变量的数目越来越多,使用滑动窗口法优化窗口内的状态变量。The initialization module needs to loosely couple the visual information and IMU information to restore the scale of the monocular camera. First, use SFM to solve the poses of all frames and the three-dimensional positions of all landmark points in the sliding window, and then align them with the previously obtained IMU pre-integrated values to solve for the angular velocity offset, gravity direction, scale factor and each The speed corresponding to one frame. As the system runs, the number of state variables increases, and the sliding window method is used to optimize the state variables within the window.
进一步地,在本发明的一个实施例中,通过无人车的测量数据进行局部视角局部建图,通过无人机的测量数据进行全局视角局部建图,包括:Further, in one embodiment of the present invention, local perspective mapping is performed using the measurement data of the unmanned vehicle, and local perspective mapping is performed using the measurement data of the unmanned aerial vehicle, including:
用SFM求解滑动窗口内所有帧的位姿与所有路标点的三维位置,与IMU 预积分值进行对齐,得到角速度偏置、重力方向、尺度因子和每一帧所对应的速度;Use SFM to solve the poses of all frames and the three-dimensional positions of all landmark points in the sliding window, and align them with the IMU pre-integrated values to obtain the angular velocity offset, gravity direction, scale factor and velocity corresponding to each frame;
使用滑动窗口法优化窗口内的状态变量,在时刻窗口中的优化的状态向量/>如下式,Use the sliding window method to optimize the state variables within the window, in Optimized state vector in time window/> As follows,
, ,
其中,和/>为相机位姿的旋转和平移部分,/>为相机在世界坐标系下的速度,/>和/>分别为IMU的加速度偏置和角速度偏置;系统的状态量的优化目标函数,如下式,in, and/> is the rotation and translation part of the camera pose,/> is the speed of the camera in the world coordinate system,/> and/> are the acceleration bias and angular velocity bias of the IMU respectively; the optimization objective function of the system's state quantity is as follows:
, ,
其中,为最大估计后验值,/>为滑窗初值残差,/>为IMU观测残差,/>为相机观测残差。in, is the maximum estimated posterior value,/> is the sliding window initial value residual,/> is the IMU observation residual,/> is the camera observation residual.
S103:通过无人机对视觉标志物进行检测,当检测成功时获取无人机和无人车之间的相对位姿,利用相对位姿的转换关系进行空地视角地图融合与优化;S103: Detect the visual markers through the drone. When the detection is successful, obtain the relative pose between the drone and the unmanned vehicle, and use the conversion relationship of the relative pose to perform air-ground perspective map fusion and optimization;
地面机器人的上方安装了一个视觉标志物,空中机器人一旦观测到标志物将会对它进行检测,一旦检测成功,将会把空地端智能体之间的相对位姿和对应的一组关键帧发送给后端。A visual marker is installed above the ground robot. Once the aerial robot observes the marker, it will detect it. Once the detection is successful, the relative pose between the air and ground agents and a corresponding set of key frames will be sent. to the backend.
在空中端检测到地面机器人搭载的视觉标识时,通过视觉标识获取视觉标识在机载相机坐标系下的转换关系,设从视觉标识到地面机器人相机的转换关系为/>,则得到当前时刻无人机相机坐标系和无人车相机坐标系之间的转换关系为/>,即为当前时刻下空地关键帧之间的位姿转换关系。同时已知当前时刻空地端产生的关键帧和/>以及它们与各自参考关键帧/>和/>之间的位姿转换关系,则可以得到地图间的位姿转换矩阵/>。When the air terminal detects the visual mark carried by the ground robot, the transformation relationship of the visual mark in the coordinate system of the airborne camera is obtained through the visual mark. , assuming the conversion relationship from visual signs to ground robot cameras is/> , then the conversion relationship between the UAV camera coordinate system and the UAV camera coordinate system at the current moment is/> , which is the pose transformation relationship between open and ground key frames at the current moment. At the same time, the key frames generated by the air and ground terminals at the current moment are known. and/> and their respective reference keyframes/> and/> The pose transformation relationship between the maps can be obtained by the pose transformation matrix/> .
S104:基于回环不断检测智能体是否经过重叠区域,当检测到重叠区域时,通过匹配的关键帧建立智能体的两幅地图的关联,进行轨迹校准和相近视角地图融合与优化;S104: Continuously detect whether the agent passes through the overlapping area based on loop closure. When the overlapping area is detected, the association between the two maps of the agent is established through matching key frames, and trajectory calibration and similar perspective map fusion and optimization are performed;
回环检测的方法基于词袋模型(Bag-of-Words),应用DBoW2库正逆序索引图像数据库。词袋模型是通过计算统计的词袋向量与当前帧的相似度判断是否产生回环。The loop detection method is based on the Bag-of-Words model and uses the DBoW2 library to index the image database in forward and reverse order. The bag-of-word model determines whether a loop is generated by calculating the similarity between the statistical bag-of-word vector and the current frame.
首先,将所有视觉特征进行聚类,一类特征,即局部相邻特征点的集合,是一个“单词”,这样所有特征就是一个“字典”。如果想把很多图像中的一共个特征点归为/>类,将“字典”做成/>叉树,进行存储和查询。通过判断一张图像中有“字典”中的哪些“单词”,就可以用一个向量描述一张图像。First, all visual features are clustered. One type of feature, that is, a collection of local adjacent feature points, is a "word", so all features are a "dictionary". If you want to have many images in total feature points are classified as/> Class, make "dictionary"into/> Fork tree for storage and query. By determining which "words" in the "dictionary" are present in an image, a vector can be used to describe an image.
进一步地,在本发明的一个实施例中,基于回环不断检测智能体是否经过重叠区域,包括:Further, in one embodiment of the present invention, continuously detecting whether the agent passes through the overlapping area based on loop closure includes:
将所有视觉特征进行聚类,一类特征是一个单词,所有特征是一个字典;All visual features are clustered, one type of feature is a word, and all features are a dictionary;
用单个向量描述一个图像:Describe an image with a single vector:
计算两张图像A和B之间的相似度:Calculate the similarity between two images A and B :
, ,
其中,是描述图像A的向量,/>是描述图像B的向量,/>是两张图像A和B之间的相似度,/>表示描述图像A的向量/>的第/>个分量,/>表示描述图像B的向量的第/>个分量;in, is a vector describing image A, /> is a vector describing image B, /> is the similarity between two images A and B, /> Represents a vector describing image A /> of/> components,/> represents a vector describing image B of/> weight;
若相似度超过阈值,则认为出现回环。If the similarity exceeds the threshold, a loop is considered to have occurred.
进一步地,在本发明的一个实施例中,将所有视觉特征进行聚类,包括:Further, in one embodiment of the present invention, all visual features are clustered, including:
在根节点,用K-means算法把所有样本聚成类,得到第一层;At the root node, use the K-means algorithm to cluster all samples into Class, get the first layer;
对第一层的每个节点,把属于该节点的样本再聚成类,得到下一层;依此类推,最后得到叶子层,叶子层即为单词。For each node in the first layer, the samples belonging to the node are clustered into Class, get the next layer; and so on, finally get the leaf layer, which is the word.
进一步地,在本发明的一个实施例中,用单个向量描述一个图像,包括:Further, in one embodiment of the present invention, a single vector is used to describe an image, including:
定义为单词/>包含的特征数量,/>为所有单词包含的特征数量,/>为单词在图像A中出现的次数,/>为单词/>在所有图像中一共出现的次数,则definition for word/> The number of features included,/> is the number of features contained in all words,/> for words The number of times it appears in image A,/> for word/> The total number of occurrences in all images, then
单词在图像A的权重/>为:/>,word weights in image A/> for:/> ,
通过词袋,用单个向量描述一个图像A:via bag-of-words, using a single vector Describe an image A:
, ,
其中,是图像A在字典中所具有的单词,/>是所对应的权重,/>是描述图像A的向量。in, is the word that image A has in the dictionary,/> yes The corresponding weight,/> is a vector describing image A.
S105:根据空地视角融合与优化后的地图和相近视角融合与优化后的地图得到位姿轨迹和全局一致的地图。S105: Obtain a map with pose trajectory and global consistency based on the map fused and optimized from the air-ground perspective and the map fused and optimized from the similar perspective.
地图融合与优化模块分为相近视角与空地视角两部分,空地视角只有在后端接收到视觉标志检测成功的信息时才会执行,它会利用相对位姿关系直接转换匹配的空地端地图进行转换,融合和优化;The map fusion and optimization module is divided into two parts: the similar perspective and the open-ground perspective. The open-ground perspective will only be executed when the back-end receives information that the visual mark detection is successful. It will use the relative pose relationship to directly convert the matching open-ground end map. , fusion and optimization;
相近视角则一直在进行,新入库的关键帧通过基于回环不断检测子端是否经过了重叠区域,重叠区域包括子端自己的回环区域和子端之间的重叠区域。一旦检测到两子端经过的区域有重叠,则通过匹配的关键帧建立两幅地图的关联,求得坐标转换关系,进一步进行地图融合,地图融合后这两个地图被合并,这些映射将从地图堆栈中删除,而由它们的融合产生的一个新的全局地图将被添加到地图堆栈中,并与两个子端直接关联。Similar viewing angles are always in progress. The newly stored key frames continuously detect whether the sub-end passes through the overlapping area based on the loop. The overlapping area includes the sub-end's own loop area and the overlapping area between the sub-ends. Once it is detected that the areas passed by the two terminals overlap, the association of the two maps is established through the matching key frames, the coordinate conversion relationship is obtained, and the maps are further fused. After the map fusion, the two maps are merged, and these maps will be are removed from the map stack, and a new global map resulting from their fusion will be added to the map stack and directly associated with both subends.
在后端优化后会将优化后的位姿发送给对应的子端,子端会更新位姿并将其作为位姿图中的约束进行局部优化,从而优化自身的局部地图以更精准地进行之后的定位建图过程。After the backend is optimized, the optimized pose will be sent to the corresponding sub-end. The sub-end will update the pose and use it as a constraint in the pose graph for local optimization, thereby optimizing its own local map to perform more accurately. The subsequent positioning and mapping process.
最终输出位姿轨迹和全局一致的地图。The final output pose trajectory and globally consistent map.
图2为本发明实施例提出的一种空地多模态多智能体协同定位与建图系统示意图,分为两部分:子端和后端。无人机和无人车都称为智能体,每个智能体都独立运行一个子端,子端搭载相机、IMU、云台及其控制系统和能够与中央服务器进行数据交换的通信单元和板载处理单元,运行一个独立的前端视觉惯性里程计,将关键帧和地图点发送给后端并维持一个规模较小的局部地图;后端接收各子端信息,通过中央服务器进行视觉标志物检测、回环检测、融合优化等计算量大的操作,输出位姿轨迹,并创建一个全局一致的地图。系统使用ROS进行子端和后端之间的通信,子端向后端传输捕捉到的关键帧和地图点,后端向子端传输更新后的位姿。Figure 2 is a schematic diagram of an air-ground multi-modal multi-agent collaborative positioning and mapping system proposed by an embodiment of the present invention, which is divided into two parts: a sub-end and a back-end. UAVs and unmanned vehicles are both called agents. Each agent runs a sub-end independently. The sub-end is equipped with a camera, IMU, gimbal and its control system, as well as communication units and boards that can exchange data with the central server. The onboard processing unit runs an independent front-end visual inertial odometry, sends key frames and map points to the back-end and maintains a smaller local map; the back-end receives information from each sub-end and performs visual marker detection through the central server. , loop closure detection, fusion optimization and other computationally intensive operations, output pose trajectories, and create a globally consistent map. The system uses ROS for communication between the sub-end and the back-end. The sub-end transmits the captured key frames and map points to the back-end, and the back-end transmits the updated pose to the sub-end.
本发明提出的一种空地多模态多智能体协同定位与建图方法,适用于多种应用场景中,并且显著提升单一机器人系统的感知能力与工作效率。尤其在搜索和救援领域中,空中无人机(UnmannedAerial Vehicle,UAV)以其空中视角的优势探测未知环境地形,引导地面机器人 (Unmanned ground vehicle,UGV)驶入目标区域实施精准的救援任务,以相较人力高效而低代价的方式实现搜索与营救。其中空中无人机拥有地面全局视野,结合地面机器人局部环境感知能力,能构建出全局一致的环境地图模型,此地图模型将为地面机器人提供全局的导航必备信息。The present invention proposes an open-ground multi-modal multi-agent collaborative positioning and mapping method, which is suitable for a variety of application scenarios and significantly improves the perception ability and work efficiency of a single robot system. Especially in the field of search and rescue, aerial drones (Unmanned Aerial Vehicle, UAV) use their aerial perspective to detect unknown environmental terrain, and guide ground robots (Unmanned ground vehicle, UGV) to drive into the target area to carry out precise rescue missions. Search and rescue is achieved in an efficient and low-cost way compared to manpower. Among them, aerial drones have a global field of view on the ground. Combined with the local environment perception capabilities of ground robots, they can build a globally consistent environmental map model. This map model will provide ground robots with necessary information for global navigation.
与现有技术相比,本发明的优点有:Compared with the existing technology, the advantages of the present invention are:
1)纯视觉SLAM有一些缺陷,视觉对于场景中的纹理特征捕捉能力较好,但对于环境中的结构特征捕捉能力不足,并且对初始化、光照敏感,并且单目相机不能得到位姿和地图的绝对尺度。本申请在视觉检测和跟踪的基础上,加入IMU传感器,进行多模态信息融合,IMU响应快,不受成像质量影响,可估计绝对尺度的特性,对室外表面无结构的物体的视觉定位进行补充。1) Pure visual SLAM has some shortcomings. Vision has good ability to capture texture features in the scene, but it has insufficient ability to capture structural features in the environment. It is also sensitive to initialization and lighting, and the monocular camera cannot obtain pose and map information. Absolute scale. On the basis of visual detection and tracking, this application adds an IMU sensor to perform multi-modal information fusion. The IMU responds quickly and is not affected by imaging quality. It can estimate the characteristics of absolute scale and perform visual positioning of objects with no structure on outdoor surfaces. Replenish.
2)在智能体运动过程中产生的震动会随着固定结构传递给相机,这会影响前端系统对特征点的采集,本申请对于相机引入云台及其控制系统,达到防抖防震的目的。2) The vibration generated during the movement of the intelligent body will be transmitted to the camera along with the fixed structure, which will affect the collection of feature points by the front-end system. This application introduces a gimbal and its control system to the camera to achieve the purpose of anti-shake and anti-shock.
为了实现上述实施例,本发明还提出空地多模态多智能体协同定位与建图装置。In order to implement the above embodiments, the present invention also proposes an air-ground multi-modal multi-agent collaborative positioning and mapping device.
图3为本发明实施例提供的一种空地多模态多智能体协同定位与建图装置的示意图。Figure 3 is a schematic diagram of an air-ground multi-modal multi-agent collaborative positioning and mapping device provided by an embodiment of the present invention.
如图3所示,该空地多模态多智能体协同定位与建图装置包括:获取模块100,建图模块200,空地视角融合模块300,相近视角融合模块400,输出模块500,其中,As shown in Figure 3, the air-ground multi-modal multi-agent collaborative positioning and mapping device includes: an acquisition module 100, a mapping module 200, an air-ground perspective fusion module 300, a similar perspective fusion module 400, and an output module 500, where,
获取模块,用于获取智能体的测量数据,智能体包括无人机和无人车,无人车设置有视觉标志物;The acquisition module is used to obtain measurement data of intelligent agents. Intelligent agents include drones and unmanned vehicles. The unmanned vehicles are equipped with visual markers;
建图模块,用于通过无人车的测量数据进行局部视角局部建图,通过无人机的测量数据进行全局视角局部建图;The mapping module is used to perform local mapping from a local perspective using the measurement data of unmanned vehicles, and perform local mapping from a global perspective using measurement data from UAVs;
空地视角融合模块,用于通过无人机对视觉标志物进行检测,当检测成功时获取无人机和无人车之间的相对位姿,利用相对位姿的转换关系进行空地视角地图融合与优化;The air-ground perspective fusion module is used to detect visual markers through drones. When the detection is successful, the relative pose between the drone and the unmanned vehicle is obtained, and the relative pose conversion relationship is used to perform air-ground perspective map fusion and optimization;
相近视角融合模块,用于基于回环不断检测智能体是否经过重叠区域,当检测到重叠区域时,通过匹配的关键帧建立智能体的两幅地图的关联,进行轨迹校准和相近视角地图融合与优化;The similar perspective fusion module is used to continuously detect whether the agent passes through the overlapping area based on loop closure. When the overlapping area is detected, the association between the two maps of the agent is established through matching key frames to perform trajectory calibration and similar perspective map fusion and optimization. ;
输出模块,用于根据空地视角融合与优化后的地图和相近视角融合与优化后的地图得到位姿轨迹和全局一致的地图。The output module is used to obtain a map with pose trajectory and global consistency based on the map fused and optimized from the air-ground perspective and the map fused and optimized from the similar perspective.
为达上述目的,本发明第三方面实施例提出了一种计算机设备,其特征在于,包括存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时,实现如上所述的空地多模态多智能体协同定位与建图方法。In order to achieve the above object, a third embodiment of the present invention provides a computer device, which is characterized in that it includes a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the computer program, the air-ground multi-modal multi-agent collaborative positioning and mapping method as described above is implemented.
为达上述目的,本发明第四方面实施例提出了一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如上所述的空地多模态多智能体协同定位与建图方法。In order to achieve the above object, a fourth embodiment of the present invention provides a computer-readable storage medium on which a computer program is stored. It is characterized in that when the computer program is executed by a processor, the above-mentioned air-ground multi-mode is implemented. State-of-the-art multi-agent collaborative localization and mapping method.
在本说明书的描述中,参考术语“一个实施例”、“一些实施例”、 “示例”、“具体示例”、或“一些示例”等的描述意指结合该实施例或示例描述的具体特征、结构、材料或者特点包含于本发明的至少一个实施例或示例中。在本说明书中,对上述术语的示意性表述不必须针对的是相同的实施例或示例。而且,描述的具体特征、结构、材料或者特点可以在任一个或多个实施例或示例中以合适的方式结合。此外,在不相互矛盾的情况下,本领域的技术人员可以将本说明书中描述的不同实施例或示例以及不同实施例或示例的特征进行结合和组合。In the description of this specification, reference to the terms "one embodiment," "some embodiments," "an example," "specific examples," or "some examples" or the like means that specific features are described in connection with the embodiment or example. , structures, materials or features are included in at least one embodiment or example of the invention. In this specification, the schematic expressions of the above terms are not necessarily directed to the same embodiment or example. Furthermore, the specific features, structures, materials or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Furthermore, those skilled in the art may combine and combine different embodiments or examples and features of different embodiments or examples described in this specification unless they are inconsistent with each other.
此外,术语“第一”、“第二”仅用于描述目的,而不能理解为指示或暗示相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括至少一个该特征。在本发明的描述中,“多个”的含义是至少两个,例如两个,三个等,除非另有明确具体的限定。In addition, the terms “first” and “second” are used for descriptive purposes only and cannot be understood as indicating or implying relative importance or implicitly indicating the quantity of indicated technical features. Therefore, features defined as "first" and "second" may explicitly or implicitly include at least one of these features. In the description of the present invention, "plurality" means at least two, such as two, three, etc., unless otherwise expressly and specifically limited.
尽管上面已经示出和描述了本发明的实施例,可以理解的是,上述实施例是示例性的,不能理解为对本发明的限制,本领域的普通技术人员在本发明的范围内可以对上述实施例进行变化、修改、替换和变型。Although the embodiments of the present invention have been shown and described above, it can be understood that the above-mentioned embodiments are illustrative and should not be construed as limitations of the present invention. Those of ordinary skill in the art can make modifications to the above-mentioned embodiments within the scope of the present invention. The embodiments are subject to changes, modifications, substitutions and variations.
Claims (8)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311248105.4A CN116989772B (en) | 2023-09-26 | 2023-09-26 | An air-ground multi-modal multi-agent collaborative positioning and mapping method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311248105.4A CN116989772B (en) | 2023-09-26 | 2023-09-26 | An air-ground multi-modal multi-agent collaborative positioning and mapping method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116989772A CN116989772A (en) | 2023-11-03 |
CN116989772B true CN116989772B (en) | 2024-01-02 |
Family
ID=88526957
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311248105.4A Active CN116989772B (en) | 2023-09-26 | 2023-09-26 | An air-ground multi-modal multi-agent collaborative positioning and mapping method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116989772B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117191005B (en) * | 2023-11-08 | 2024-01-30 | 中国人民解放军国防科技大学 | An air-ground heterogeneous collaborative mapping method, device, equipment and storage medium |
CN118424256A (en) * | 2024-04-18 | 2024-08-02 | 北京化工大学 | Map building and positioning method and device for distributed multi-resolution map fusion |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110044354A (en) * | 2019-03-28 | 2019-07-23 | 东南大学 | A kind of binocular vision indoor positioning and build drawing method and device |
CN210377164U (en) * | 2019-06-17 | 2020-04-21 | 酷黑科技(北京)有限公司 | Air-ground cooperative operation system |
CN112802104A (en) * | 2021-02-04 | 2021-05-14 | 华南理工大学 | Loop detection method based on RGB-D camera |
CN114485619A (en) * | 2022-01-26 | 2022-05-13 | 清华大学 | Multi-robot positioning and navigation method and device based on air-ground cooperation |
CN114966789A (en) * | 2022-05-12 | 2022-08-30 | 华中科技大学 | Mapping method and system fusing GNSS and multi-view vision |
CN115993825A (en) * | 2022-12-27 | 2023-04-21 | 中国人民解放军陆军工程大学 | Unmanned vehicle cluster control system based on air-ground cooperation |
WO2023065342A1 (en) * | 2021-10-22 | 2023-04-27 | 华为技术有限公司 | Vehicle, vehicle positioning method and apparatus, device, and computer-readable storage medium |
CN116358520A (en) * | 2023-03-16 | 2023-06-30 | 湖南大学 | A human-machine multi-node collaborative semantic laser SLAM system and method |
-
2023
- 2023-09-26 CN CN202311248105.4A patent/CN116989772B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110044354A (en) * | 2019-03-28 | 2019-07-23 | 东南大学 | A kind of binocular vision indoor positioning and build drawing method and device |
CN210377164U (en) * | 2019-06-17 | 2020-04-21 | 酷黑科技(北京)有限公司 | Air-ground cooperative operation system |
CN112802104A (en) * | 2021-02-04 | 2021-05-14 | 华南理工大学 | Loop detection method based on RGB-D camera |
WO2023065342A1 (en) * | 2021-10-22 | 2023-04-27 | 华为技术有限公司 | Vehicle, vehicle positioning method and apparatus, device, and computer-readable storage medium |
CN114485619A (en) * | 2022-01-26 | 2022-05-13 | 清华大学 | Multi-robot positioning and navigation method and device based on air-ground cooperation |
CN114966789A (en) * | 2022-05-12 | 2022-08-30 | 华中科技大学 | Mapping method and system fusing GNSS and multi-view vision |
CN115993825A (en) * | 2022-12-27 | 2023-04-21 | 中国人民解放军陆军工程大学 | Unmanned vehicle cluster control system based on air-ground cooperation |
CN116358520A (en) * | 2023-03-16 | 2023-06-30 | 湖南大学 | A human-machine multi-node collaborative semantic laser SLAM system and method |
Non-Patent Citations (3)
Title |
---|
基于无人机与无人车协同的车辆路径规划;胡秋霞;《中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑》 第5期;全文 * |
微小型空地无人系统协同导航技术研究;邹鹏程;《中国优秀硕士学位论文全文数据库 信息科技辑》 第3期;全文 * |
非确定环境下无人机与无人车动态协同设计;徐文菁;《洛阳理工学院学报( 自然科学版)》;第29卷(第4期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN116989772A (en) | 2023-11-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110068335B (en) | A method and system for real-time positioning of UAV swarms in GPS-denied environment | |
CN108711166B (en) | A Monocular Camera Scale Estimation Method Based on Quadrotor UAV | |
Heng et al. | Self-calibration and visual slam with a multi-camera system on a micro aerial vehicle | |
CN109520497B (en) | Unmanned aerial vehicle autonomous positioning method based on vision and imu | |
CN109307508B (en) | Panoramic inertial navigation SLAM method based on multiple key frames | |
CN109211241B (en) | Autonomous positioning method of UAV based on visual SLAM | |
CN112734765B (en) | Mobile robot positioning method, system and medium based on fusion of instance segmentation and multiple sensors | |
Teuliere et al. | Chasing a moving target from a flying UAV | |
CN116989772B (en) | An air-ground multi-modal multi-agent collaborative positioning and mapping method | |
CN106873619B (en) | Processing method of flight path of unmanned aerial vehicle | |
CN111288989B (en) | Visual positioning method for small unmanned aerial vehicle | |
CN108759826B (en) | A UAV motion tracking method based on the fusion of multi-sensing parameters of mobile phones and UAVs | |
CN110726406A (en) | An Improved Nonlinear Optimization Method for Monocular Inertial Navigation SLAM | |
CN114966734A (en) | Bidirectional depth vision inertial pose estimation method combined with multi-line laser radar | |
Eynard et al. | Real time UAV altitude, attitude and motion estimation from hybrid stereovision | |
Zhou et al. | On-board inertial-assisted visual odometer on an embedded system | |
CN114459467A (en) | A target localization method based on VI-SLAM in unknown rescue environment | |
CN114485640A (en) | Monocular visual-inertial synchronous positioning and mapping method and system based on point and line features | |
CN117152249A (en) | Multi-UAV collaborative mapping and perception method and system based on semantic consistency | |
Huang et al. | Monocular vision-based autonomous navigation system on a toy quadcopter in unknown environments | |
Lu et al. | Vision-based localization methods under GPS-denied conditions | |
Jian et al. | Lvcp: Lidar-vision tightly coupled collaborative real-time relative positioning | |
Hosen et al. | Vision-aided nonlinear observer for fixed-wing unmanned aerial vehicle navigation | |
Ling et al. | RGB-D inertial odometry for indoor robot via keyframe-based nonlinear optimization | |
CN117870652B (en) | A land-air online collaborative positioning method based on radar, inertial and vision fusion |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |