CN116993933A - Live-action map construction method, device and equipment under emergency scene and storage medium - Google Patents
Live-action map construction method, device and equipment under emergency scene and storage medium Download PDFInfo
- Publication number
- CN116993933A CN116993933A CN202310719078.8A CN202310719078A CN116993933A CN 116993933 A CN116993933 A CN 116993933A CN 202310719078 A CN202310719078 A CN 202310719078A CN 116993933 A CN116993933 A CN 116993933A
- Authority
- CN
- China
- Prior art keywords
- image
- real
- map
- disaster
- communication positioning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/05—Geographic models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/217—Validation; Performance evaluation; Active pattern learning techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
- G06Q50/26—Government or public services
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/757—Matching configurations of points or features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/761—Proximity, similarity or dissimilarity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/17—Terrestrial scenes taken from planes or by drones
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Life Sciences & Earth Sciences (AREA)
- Multimedia (AREA)
- Remote Sensing (AREA)
- Evolutionary Biology (AREA)
- Biophysics (AREA)
- Tourism & Hospitality (AREA)
- Mathematical Physics (AREA)
- Geometry (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Biomedical Technology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Development Economics (AREA)
- Educational Administration (AREA)
- Computer Graphics (AREA)
- Economics (AREA)
- Human Resources & Organizations (AREA)
- Marketing (AREA)
- Primary Health Care (AREA)
- Strategic Management (AREA)
- General Business, Economics & Management (AREA)
- Image Processing (AREA)
Abstract
Description
技术领域Technical Field
本发明涉及人工智能技术领域,尤其涉及一种应急场景下的实景地图构建方法、装置、设备及存储介质。The present invention relates to the field of artificial intelligence technology, and in particular to a method, device, equipment and storage medium for constructing a real-scene map in an emergency scenario.
背景技术Background Art
自然灾害的发生通常让人始料未及,由于灾区基础设施薄弱,地形复杂等,导致应急救援严重滞后。全局的感知、可靠的通信、精确的定位作为应急体系的重要组成部分,对及时做出救援决策尤为重要。Natural disasters usually happen unexpectedly. Due to weak infrastructure and complex terrain in disaster areas, emergency rescue is seriously delayed. Global perception, reliable communication, and accurate positioning are important components of the emergency response system and are particularly important for making timely rescue decisions.
目前,主要采用以经验为主的专家评判手段,即指挥人员通过对各个应急救援区域的灾情进行判断,从而指导通信定位需求的分级。然而,灾情瞬息万变,设施损毁,受灾区域地形复杂,专家资源不足,难以及时准确获取全局态势并正确分配救援资源。At present, the main method used is expert judgment based on experience, that is, the commander judges the disaster situation in each emergency rescue area to guide the classification of communication positioning needs. However, the disaster situation changes rapidly, facilities are damaged, the terrain of the disaster area is complex, and expert resources are insufficient, making it difficult to obtain the overall situation in a timely and accurate manner and correctly allocate rescue resources.
因此,现有技术的不足在于:实时灾情感知效率较低,且通信定位需求分级的准确性不高,导致指挥人员难以及时准确获取全局态势并正确分配救援资源。Therefore, the shortcomings of the existing technology are: the efficiency of real-time disaster perception is low, and the accuracy of communication positioning demand classification is not high, which makes it difficult for commanders to obtain the overall situation in a timely and accurate manner and correctly allocate rescue resources.
发明内容Summary of the invention
本发明提供一种应急场景下的实景地图构建方法、装置、设备及存储介质,用以解决现有技术中实时灾情感知效率较低,且通信定位需求分级的准确性不高,导致指挥人员难以及时准确获取全局态势并正确分配救援资源的缺陷,实现提升实时灾情感知效率和通信定位需求分级的准确性,从而更好地辅助指挥人员对全局态势的感知,并合理分配救援资源的目的。The present invention provides a method, device, equipment and storage medium for constructing a real-life map in an emergency scenario, which is used to solve the defects in the prior art that the efficiency of real-time disaster perception is low and the accuracy of communication positioning demand classification is not high, resulting in difficulty for commanders to timely and accurately obtain the overall situation and correctly allocate rescue resources, thereby improving the efficiency of real-time disaster perception and the accuracy of communication positioning demand classification, thereby better assisting commanders in perceiving the overall situation and reasonably allocating rescue resources.
本发明提供一种应急场景下的实景地图构建方法,包括:The present invention provides a method for constructing a real scene map in an emergency scenario, comprising:
获取多无人机拍摄的多个受灾区域的实景图像;Obtain real-life images of multiple disaster-stricken areas taken by multiple drones;
基于各所述实景图像,构建实时灾情感知态势图;Based on the real-scene images, a real-time disaster awareness situation map is constructed;
获取通信定位需求分级模型,所述通信定位需求分级模型是基于应急场景下的历史分级数据训练得到的;Acquire a communication positioning demand grading model, wherein the communication positioning demand grading model is obtained by training based on historical grading data in emergency scenarios;
基于所述实时灾情感知态势图和所述通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图。Based on the real-time disaster awareness situation map and the communication positioning demand grading model, a real-life map based on communication positioning demand grading calibration and disaster point calibration is constructed.
根据本发明提供的一种应急场景下的实景地图构建方法,所述基于各所述实景图像,构建实时灾情感知态势图,包括:According to a method for constructing a real-scene map in an emergency scenario provided by the present invention, the real-time disaster awareness situation map is constructed based on each of the real-scene images, including:
将所述实景图像从像素坐标系转换到世界坐标系,得到目标图像;Convert the real scene image from the pixel coordinate system to the world coordinate system to obtain a target image;
基于各所述目标图像的世界坐标,将各所述目标图像进行拼接处理,得到拼接图像;Based on the world coordinates of each of the target images, stitching the target images to obtain a stitched image;
将所述拼接图像中的空缺区域进行补全处理,得到所述实时灾情感知态势图。The missing areas in the spliced image are completed to obtain the real-time disaster awareness situation map.
根据本发明提供的一种应急场景下的实景地图构建方法,所述基于各所述目标图像的世界坐标,将各所述目标图像进行拼接处理,得到拼接图像,包括:According to a method for constructing a real scene map in an emergency scenario provided by the present invention, the target images are spliced based on the world coordinates of the target images to obtain a spliced image, including:
针对任意两张所述目标图像,根据两张所述目标图像的世界坐标判断两张所述目标图像是否存在关联区域;For any two target images, judging whether there is a related area between the two target images according to the world coordinates of the two target images;
在两张所述目标图像存在关联区域的情况下,采用尺度不变的特征点检测算法对两张所述目标图像分别进行特征检测,得到特征点集;In the case where there are associated regions between the two target images, a scale-invariant feature point detection algorithm is used to perform feature detection on the two target images respectively to obtain a feature point set;
计算所述特征点集中每个特征点的描述符,得到描述符集;Calculating the descriptor of each feature point in the feature point set to obtain a descriptor set;
基于两张所述目标图像对应的所述描述符集,对两张所述目标图像对应的所述特征点集进行相似特征点的匹配处理,得到多个匹配点;Based on the descriptor sets corresponding to the two target images, matching processing of similar feature points is performed on the feature point sets corresponding to the two target images to obtain a plurality of matching points;
将所述多个匹配点中的异常匹配点进行过滤,以对所述多个匹配点进行矫正;Filtering abnormal matching points among the multiple matching points to correct the multiple matching points;
将两张所述目标图像分别进行透视变换;Performing perspective transformation on the two target images respectively;
基于矫正后的所述多个匹配点,对透视变换后的两张所述目标图像进行图像融合,得到所述拼接图像;其中,在所述图像融合的过程中,采用羽化技术进行平滑的图像转换,以及通过加权平均颜色值对重叠的像素进行融合。Based on the corrected multiple matching points, the two target images after perspective transformation are fused to obtain the stitched image; wherein, in the process of image fusion, feathering technology is used to perform smooth image conversion, and overlapping pixels are fused by weighted average color value.
根据本发明提供的一种应急场景下的实景地图构建方法,所述将所述拼接图像中的空缺区域进行补全处理,得到所述实时灾情感知态势图,包括:According to a method for constructing a real-scene map in an emergency scenario provided by the present invention, the method of completing the missing areas in the spliced image to obtain the real-time disaster awareness situation map includes:
对所述拼接图像中的空缺区域进行低维视觉先验信息的重建;Reconstructing low-dimensional visual prior information of the vacant area in the stitched image;
基于所述低维视觉先验信息,向重建后的所述拼接图像中补充纹理细节信息,得到所述实时灾情感知态势图。Based on the low-dimensional visual prior information, texture detail information is added to the reconstructed spliced image to obtain the real-time disaster awareness situation map.
根据本发明提供的一种应急场景下的实景地图构建方法,所述对所述拼接图像中的空缺区域进行低维视觉先验信息的重建,包括:According to a method for constructing a real scene map in an emergency scenario provided by the present invention, the reconstruction of low-dimensional visual prior information of the vacant area in the spliced image includes:
将所述拼接图像降采样为低维图像;Downsampling the stitched image into a low-dimensional image;
将所述低维图像确定为所述拼接图像的视觉先验图像;Determining the low-dimensional image as a visual prior image of the spliced image;
采用聚类算法在预设图像数据集的颜色空间中生成颜色字典;A clustering algorithm is used to generate a color dictionary in the color space of a preset image data set;
针对所述视觉先验图像中的每个像素,在所述颜色字典中查找与所述像素最接近的元素索引,得到所述像素的离散表示;For each pixel in the visual prior image, searching the color dictionary for an element index closest to the pixel to obtain a discrete representation of the pixel;
将所述视觉先验图像中各所述像素的离散表示输入Transformer模型中进行迭代计算,将所述拼接图像中的空缺区域中各所述像素的离散表示中的元素替换为吉布斯采样标记,得到离散序列;Inputting the discrete representation of each pixel in the visual prior image into the Transformer model for iterative calculation, replacing the elements in the discrete representation of each pixel in the vacant area in the spliced image with Gibbs sampling marks, to obtain a discrete sequence;
针对每个所述离散序列,通过查询所述颜色字典来重建所述低维视觉先验信息。For each of the discrete sequences, the low-dimensional visual prior information is reconstructed by querying the color dictionary.
根据本发明提供的一种应急场景下的实景地图构建方法,所述基于所述低维视觉先验信息,向重建后的所述拼接图像中补充纹理细节信息,得到所述实时灾情感知态势图,包括:According to a method for constructing a real-life map in an emergency scenario provided by the present invention, the method of supplementing texture detail information to the reconstructed spliced image based on the low-dimensional visual prior information to obtain the real-time disaster awareness situation map includes:
将重建后的所述视觉先验图像的图像矩阵和所述图像矩阵的双线性插值结果输入到引导上采样网络中进行处理,得到预测值;Inputting the reconstructed image matrix of the visual prior image and the bilinear interpolation result of the image matrix into a guided upsampling network for processing to obtain a predicted value;
基于所述预测值和真实值之间的损失函数来调整所述引导上采样网络的模型参数;所述损失函数为L1损失函数和对抗性损失函数的加权和;所述对抗性损失函数与判别器对所述预测值的判别值相关;Adjusting the model parameters of the guided upsampling network based on a loss function between the predicted value and the true value; the loss function is a weighted sum of an L1 loss function and an adversarial loss function; the adversarial loss function is related to a discriminant value of the discriminator to the predicted value;
联合训练所述引导上采样网络和所述判别器,得到多元化还原结果;Jointly training the guided upsampling network and the discriminator to obtain diversified restoration results;
从所述多元化还原结果中确定鲁棒性最高的结果;Determining the most robust result from the plurality of reduction results;
基于所述鲁棒性最高的结果向重建后的所述拼接图像中补充所述纹理细节信息,得到所述实时灾情感知态势图。The texture detail information is supplemented to the reconstructed spliced image based on the result with the highest robustness to obtain the real-time disaster awareness situation map.
根据本发明提供的一种应急场景下的实景地图构建方法,获取通信定位需求分级模型,包括:According to a method for constructing a real scene map in an emergency scenario provided by the present invention, a communication positioning demand grading model is obtained, including:
获取所述应急场景下的历史分级数据;所述历史分级数据包括分级指标和针对每个分级指标的专家评语标签;Acquire historical classification data under the emergency scenario; the historical classification data includes classification indicators and expert comment labels for each classification indicator;
将所述历史分级数据划分为训练集和测试集;Dividing the historical classification data into a training set and a test set;
将所述训练集输入所述通信定位需求分级模型中进行训练;Inputting the training set into the communication positioning demand grading model for training;
将所述测试集输入训练好的所述通信定位需求分级模型中进行测试;Inputting the test set into the trained communication positioning demand grading model for testing;
在所述通信定位需求分级模型的测试结果未通过的情况下,对所述通信定位需求分级模型的模型参数进行微调,直至所述通信定位需求分级模型的测试结果通过。When the test result of the communication positioning requirement grading model fails, the model parameters of the communication positioning requirement grading model are fine-tuned until the test result of the communication positioning requirement grading model passes.
根据本发明提供的一种应急场景下的实景地图构建方法,所述基于所述实时灾情感知态势图和所述通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图,包括:According to a method for constructing a real scene map in an emergency scenario provided by the present invention, based on the real-time disaster awareness situation map and the communication positioning demand grading model, a real scene map based on communication positioning demand grading calibration and disaster point calibration is constructed, comprising:
确定所述通信定位需求分级模型的输入指标;所述输入指标包括服务对象、应急场景、处置阶段和事件客观因素的细化类型以及通信定位需求的各个等级;Determine the input indicators of the communication positioning demand grading model; the input indicators include detailed types of service objects, emergency scenarios, handling stages and objective factors of events and various levels of communication positioning demand;
将所述输入指标输入所述通信定位需求分级模型中进行处理,输出通信定位需求分级标定结果;Inputting the input index into the communication positioning demand grading model for processing, and outputting a communication positioning demand grading calibration result;
基于所述实时灾情感知态势图确定灾点标定结果;Determine the disaster point calibration result based on the real-time disaster awareness situation map;
将所述通信定位需求分级标定结果和所述灾点标定结果添加至所述实时灾情感知态势图中,得到所述实景地图。The communication positioning demand classification calibration result and the disaster point calibration result are added to the real-time disaster awareness situation map to obtain the real-scene map.
本发明还提供一种应急场景下的实景地图构建装置,包括:The present invention also provides a real scene map construction device in an emergency scenario, comprising:
第一获取模块,用于获取多无人机拍摄的受灾区域的实景图像;The first acquisition module is used to acquire real-life images of the disaster-stricken area taken by multiple drones;
第一构建模块,用于基于各所述实景图像,构建实时灾情感知态势图;A first construction module is used to construct a real-time disaster awareness situation map based on each of the real scene images;
第二获取模块,用于获取通信定位需求分级模型,所述通信定位需求分级模型是基于应急场景下的历史分级数据训练得到的;A second acquisition module is used to acquire a communication positioning demand classification model, where the communication positioning demand classification model is trained based on historical classification data in emergency scenarios;
第二构建模块,用于基于所述实时灾情感知态势图和所述通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图。The second construction module is used to construct a real-life map based on communication positioning demand grading calibration and disaster point calibration based on the real-time disaster awareness situation map and the communication positioning demand grading model.
本发明还提供一种电子设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述程序时实现如上述任一种所述的应急场景下的实景地图构建方法的步骤。The present invention also provides an electronic device, comprising a memory, a processor, and a computer program stored in the memory and executable on the processor, wherein when the processor executes the program, the steps of the method for constructing a real-life map in an emergency scenario as described in any one of the above are implemented.
本发明还提供一种非暂态计算机可读存储介质,其上存储有计算机程序,该计算机程序被处理器执行时实现如上述任一种所述的应急场景下的实景地图构建方法的步骤。The present invention also provides a non-transitory computer-readable storage medium having a computer program stored thereon, which, when executed by a processor, implements the steps of the method for constructing a real-scene map in an emergency scenario as described in any one of the above.
本发明提供的应急场景下的实景地图构建方法、装置、设备及存储介质,首先,获取多无人机拍摄的多个受灾区域的实景图像,并基于各实景图像构建实时灾情感知态势图,可以提升实时灾情感知效率;然后,获取通信定位需求分级模型,通信定位需求分级模型是基于应急场景下的历史分级数据训练得到的;基于实时灾情感知态势图和通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图,可以提升通信定位需求分级的准确性,从而更好地辅助指挥人员对全局态势的感知,并合理分配救援资源。The present invention provides a method, device, equipment and storage medium for constructing a real-life map in an emergency scenario. First, real-life images of multiple disaster-stricken areas taken by multiple drones are obtained, and a real-time disaster awareness situation map is constructed based on each real-life image, which can improve the efficiency of real-time disaster awareness. Then, a communication positioning demand grading model is obtained, and the communication positioning demand grading model is obtained by training based on historical grading data in emergency scenarios. Based on the real-time disaster awareness situation map and the communication positioning demand grading model, a real-life map based on communication positioning demand grading calibration and disaster point calibration is constructed, which can improve the accuracy of communication positioning demand grading, thereby better assisting commanders in perceiving the overall situation and reasonably allocating rescue resources.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
为了更清楚地说明本发明或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in the present invention or the prior art, the drawings required for use in the embodiments or the description of the prior art will be briefly introduced below. Obviously, the drawings described below are some embodiments of the present invention. For ordinary technicians in this field, other drawings can be obtained based on these drawings without paying creative work.
图1是本发明实施例提供的应急场景下的实景地图构建方法的流程示意图;FIG1 is a schematic diagram of a flow chart of a method for constructing a real scene map in an emergency scenario provided by an embodiment of the present invention;
图2是本发明实施例提供的应急场景下的实景地图构建方法的场景示意图;2 is a schematic diagram of a scene of a method for constructing a real scene map in an emergency scenario provided by an embodiment of the present invention;
图3是本发明实施例提供的无人机采样图像坐标转换流程示意图;FIG3 is a schematic diagram of a coordinate conversion process of a drone sampling image provided by an embodiment of the present invention;
图4是本发明实施例提供的无人机采样图像坐标转换示意图;FIG4 is a schematic diagram of coordinate conversion of a drone sampling image provided by an embodiment of the present invention;
图5是本发明实施例提供的无人机采样图像拼接流程示意图;FIG5 is a schematic diagram of a UAV sampling image stitching process according to an embodiment of the present invention;
图6a是本发明实施例提供的待检测的目标图像1的示意图;FIG6a is a schematic diagram of a target image 1 to be detected provided by an embodiment of the present invention;
图6b是本发明实施例提供的待检测的目标图像2的示意图;FIG6b is a schematic diagram of a target image 2 to be detected provided by an embodiment of the present invention;
图6c是本发明实施例提供的目标图像1的特征点检测示意图;FIG6c is a schematic diagram of feature point detection of a target image 1 provided by an embodiment of the present invention;
图6d是本发明实施例提供的目标图像2的特征点检测示意图;FIG6d is a schematic diagram of feature point detection of a target image 2 provided by an embodiment of the present invention;
图7是本发明实施例提供的特征点匹配示意图;FIG7 is a schematic diagram of feature point matching provided by an embodiment of the present invention;
图8是本发明实施例提供的图像拼接示意图;FIG8 is a schematic diagram of image stitching provided by an embodiment of the present invention;
图9是本发明实施例提供的无人机采样图像补全流程示意图;FIG9 is a schematic diagram of a UAV sampling image completion process according to an embodiment of the present invention;
图10是本发明实施例提供的基于Transformer的视觉先验生成算法流程示意图;FIG10 is a flow chart of a Transformer-based visual prior generation algorithm according to an embodiment of the present invention;
图11是本发明实施例提供的基于CNN的细节纹理修复流程示意图;FIG11 is a schematic diagram of a detail texture restoration process based on CNN provided by an embodiment of the present invention;
图12是本发明实施例提供的分级指标示意图;FIG12 is a schematic diagram of a grading index provided in an embodiment of the present invention;
图13是本发明实施例提供的服务对象指标示意图;13 is a schematic diagram of service object indicators provided in an embodiment of the present invention;
图14是本发明实施例提供的应急场景指标示意图;FIG14 is a schematic diagram of emergency scenario indicators provided by an embodiment of the present invention;
图15是本发明实施例提供的处置阶段指标示意图;15 is a schematic diagram of treatment stage indicators provided by an embodiment of the present invention;
图16是本发明实施例提供的事件客观因素指标示意图;FIG16 is a schematic diagram of objective factor indicators of events provided by an embodiment of the present invention;
图17是本发明实施例提供的通信定位需求等级示意图;17 is a schematic diagram of the communication positioning requirement level provided by an embodiment of the present invention;
图18是本发明实施例提供的需求分级算法流程示意图;FIG18 is a schematic diagram of a demand classification algorithm flow chart provided by an embodiment of the present invention;
图19是本发明实施例提供的无人机灾情航拍数据集;FIG19 is a drone disaster aerial photography dataset provided by an embodiment of the present invention;
图20是本发明实施例提供的无人机采集拼接结果;FIG20 is a mosaic result of drone acquisition provided by an embodiment of the present invention;
图21是本发明实施例提供的无人机采集拼接裁剪结果;FIG21 is a result of drone acquisition, stitching and cropping provided by an embodiment of the present invention;
图22是本发明实施例提供的无人机采集图像补全结果;FIG22 is a completion result of an image collected by a drone provided in an embodiment of the present invention;
图23是本发明实施例提供的不同带宽与不同采集区域条件下的实景地图生成质量变化曲线;FIG23 is a curve showing the quality change of the real scene map generated under different bandwidths and different acquisition area conditions provided by an embodiment of the present invention;
图24是本发明实施例提供的实景地图灾点检测与标定;FIG24 is a diagram of disaster point detection and calibration of a real-life map provided by an embodiment of the present invention;
图25a是本发明实施例提供的通信定位需求分级模型混淆矩阵之一;FIG25a is one of the confusion matrices of the communication positioning requirement classification model provided by an embodiment of the present invention;
图25b是本发明实施例提供的通信定位需求分级模型混淆矩阵之二;FIG25b is a second confusion matrix of the communication positioning requirement classification model provided by an embodiment of the present invention;
图26是本发明实施例提供的基于通信定位需求分级标定与灾点标定的实景地图;26 is a real-life map based on communication positioning demand classification calibration and disaster point calibration provided by an embodiment of the present invention;
图27是本发明实施例提供的应急场景下的实景地图构建装置的结构示意图;27 is a schematic diagram of the structure of a real scene map construction device for emergency scenarios provided by an embodiment of the present invention;
图28是本发明实施例提供的电子设备的结构示意图。FIG. 28 is a schematic diagram of the structure of an electronic device provided in an embodiment of the present invention.
具体实施方式DETAILED DESCRIPTION
为使本发明的目的、技术方案和优点更加清楚,下面将结合本发明中的附图,对本发明中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。In order to make the purpose, technical solution and advantages of the present invention clearer, the technical solution of the present invention will be clearly and completely described below in conjunction with the drawings of the present invention. Obviously, the described embodiments are part of the embodiments of the present invention, not all of the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by ordinary technicians in this field without creative work are within the scope of protection of the present invention.
下面结合图1至图26描述本发明的应急场景下的实景地图构建方法。The following describes the real scene map construction method for emergency scenarios of the present invention in conjunction with Figures 1 to 26.
请参照图1,图1是本发明实施例提供的应急场景下的实景地图构建方法的流程示意图。如图1所示,该方法可以包括以下步骤:Please refer to Figure 1, which is a flow chart of a method for constructing a real scene map in an emergency scenario provided by an embodiment of the present invention. As shown in Figure 1, the method may include the following steps:
步骤101、获取多无人机拍摄的多个受灾区域的实景图像;Step 101: Acquire real-life images of multiple disaster-stricken areas taken by multiple drones;
步骤102、基于各实景图像,构建实时灾情感知态势图;Step 102: construct a real-time disaster awareness situation map based on each real scene image;
步骤103、获取通信定位需求分级模型,通信定位需求分级模型是基于应急场景下的历史分级数据训练得到的;Step 103: Obtain a communication positioning demand classification model, where the communication positioning demand classification model is trained based on historical classification data in emergency scenarios;
步骤104、基于实时灾情感知态势图和通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图。Step 104: Based on the real-time disaster awareness situation map and the communication positioning demand grading model, a real-life map based on communication positioning demand grading calibration and disaster point calibration is constructed.
在步骤101中,如图2所示,受灾区域有多个,无人机有多架,获取多架无人机拍摄的多个受灾区域的实景图像。In step 101, as shown in FIG. 2, there are multiple disaster-stricken areas and multiple drones, and real-scene images of the multiple disaster-stricken areas taken by the multiple drones are acquired.
示例性地,假设有k架无人机M×N的矩形区域内执行感知任务,并将拍摄的多张实景图像回传至地面指挥车与服务器,无人机集合可表示为第k架无人机的位置可用坐标uk={xk,yk,zk}表示,无人机采集的实景图像集合可表示为 For example, suppose there are k drones performing perception tasks in a rectangular area of M×N, and transmitting multiple real-scene images taken back to the ground command vehicle and server. The drone set can be represented as The position of the kth UAV can be represented by the coordinates u k ={x k ,y k ,z k }, and the set of real-life images collected by the UAV can be represented as
在本步骤中,一方面,由于受灾区域地形复杂,搜救区域广泛,单无人机拍摄全局态势图效率低,难以保证应急救援对实景地图新鲜度高的需求。因此,需要引入多无人机进行协同感知。另一方面,由于受灾区域内地面基站等设施的损毁,导致带宽资源匮乏,无人机仅支持回传部分高清图片。因此,需要多无人机拍摄多张实景图像进行拼接。In this step, on the one hand, due to the complex terrain of the disaster-stricken area and the wide search and rescue area, the efficiency of a single drone in shooting a global situation map is low, and it is difficult to ensure the high freshness of the real-life map required for emergency rescue. Therefore, it is necessary to introduce multiple drones for collaborative perception. On the other hand, due to the damage of ground base stations and other facilities in the disaster-stricken area, bandwidth resources are scarce, and drones only support the transmission of some high-definition pictures. Therefore, multiple drones are needed to shoot multiple real-life images for stitching.
在步骤102中,可以通过将多架无人机拍摄的各实景图像进行坐标转换、拼接处理和补全处理,构建实时灾情感知态势图,可以提升实时灾情感知效率。In step 102, the real-time disaster awareness situation map can be constructed by performing coordinate conversion, splicing processing and completion processing on the real-scene images taken by multiple drones, thereby improving the efficiency of real-time disaster awareness.
在步骤103中,应急场景可以包括火灾、烟雾、地震等,应急场景下的历史分级数据可以包括大量的火灾、烟雾、地震等应急场景下的灾情数据集。In step 103, emergency scenarios may include fire, smoke, earthquake, etc., and the historical hierarchical data under emergency scenarios may include a large number of disaster data sets under emergency scenarios such as fire, smoke, earthquake, etc.
具体地,基于应急场景下的历史分级数据训练通信定位需求分级模型,获取训练好的通信定位需求分级模型。Specifically, a communication positioning demand classification model is trained based on historical classification data in emergency scenarios to obtain a trained communication positioning demand classification model.
在一种可能的实施方式中,步骤103可以包括以下子步骤:In a possible implementation, step 103 may include the following sub-steps:
步骤1031、获取应急场景下的历史分级数据;历史分级数据包括分级指标和针对每个分级指标的专家评语标签;Step 1031: Obtain historical classification data in emergency scenarios; the historical classification data includes classification indicators and expert comment labels for each classification indicator;
步骤1032、将历史分级数据划分为训练集和测试集;Step 1032: Divide the historical classification data into a training set and a test set;
步骤1033、将训练集输入通信定位需求分级模型中进行训练;Step 1033: input the training set into the communication positioning demand grading model for training;
步骤1034、将测试集输入训练好的通信定位需求分级模型中进行测试;Step 1034: input the test set into the trained communication positioning requirement classification model for testing;
步骤1035、在通信定位需求分级模型的测试结果未通过的情况下,对通信定位需求分级模型的模型参数进行微调,直至通信定位需求分级模型的测试结果通过。Step 1035: When the test result of the communication positioning requirement grading model fails, fine-tune the model parameters of the communication positioning requirement grading model until the test result of the communication positioning requirement grading model passes.
在步骤104中,通过实时灾情感知态势图可以进行灾点标定,通过通信定位需求分级模型可以进行通信定位需求分级标定,从而得到基于实时灾情感知态势图和通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图,可以提升通信定位需求分级的准确性。In step 104, disaster points can be calibrated through the real-time disaster awareness situation map, and communication positioning demand grading calibration can be performed through the communication positioning demand grading model, so as to obtain a real-time disaster awareness situation map and the communication positioning demand grading model, and construct a real-life map based on the communication positioning demand grading calibration and disaster point calibration, which can improve the accuracy of communication positioning demand grading.
本发明实施例提供的应急场景下的实景地图构建方法,首先,获取多无人机拍摄的多个受灾区域的实景图像,并基于各实景图像构建实时灾情感知态势图,可以提升实时灾情感知效率;然后,获取通信定位需求分级模型,通信定位需求分级模型是基于应急场景下的历史分级数据训练得到的;基于实时灾情感知态势图和通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图,可以提升通信定位需求分级的准确性,从而更好地辅助指挥人员对全局态势的感知,并合理分配救援资源。The method for constructing a real-life map in an emergency scenario provided by an embodiment of the present invention first obtains real-life images of multiple disaster-stricken areas taken by multiple drones, and constructs a real-time disaster awareness situation map based on each real-life image, which can improve the efficiency of real-time disaster awareness; then, a communication positioning demand grading model is obtained, and the communication positioning demand grading model is trained based on historical grading data in the emergency scenario; based on the real-time disaster awareness situation map and the communication positioning demand grading model, a real-life map based on communication positioning demand grading calibration and disaster point calibration is constructed, which can improve the accuracy of communication positioning demand grading, thereby better assisting commanders in perceiving the overall situation and reasonably allocating rescue resources.
基于图1对应实施例的应急场景下的实景地图构建方法,在一种示例实施例中,步骤102可以包括以下子步骤:Based on the method for constructing a real scene map in an emergency scenario according to the embodiment of FIG. 1 , in an exemplary embodiment, step 102 may include the following sub-steps:
步骤1021、将实景图像从像素坐标系转换到世界坐标系,得到目标图像;Step 1021, converting the real scene image from the pixel coordinate system to the world coordinate system to obtain a target image;
步骤1022、基于各目标图像的世界坐标,将各目标图像进行拼接处理,得到拼接图像;Step 1022: based on the world coordinates of each target image, stitching the target images together to obtain a stitched image;
步骤1023、将拼接图像中的空缺区域进行补全处理,得到实时灾情感知态势图。Step 1023: Fill in the missing areas in the spliced image to obtain a real-time disaster awareness situation map.
在步骤1021中,如图3所示,相机成像系统的坐标转换涉及到不同的坐标系,分别是世界坐标系、相机坐标系、理想图像坐标系、真实图像坐标系和像素坐标系。这些坐标系之间的转换和映射是实现从世界坐标系到像素坐标系转换的必要步骤。In step 1021, as shown in FIG3, the coordinate transformation of the camera imaging system involves different coordinate systems, namely the world coordinate system, the camera coordinate system, the ideal image coordinate system, the real image coordinate system and the pixel coordinate system. The transformation and mapping between these coordinate systems are necessary steps to achieve the transformation from the world coordinate system to the pixel coordinate system.
示例性地,对k-th无人机拍摄的实景图像进行坐标转换,得到M×N的目标图像。其中,第k-th无人机采集的t张实景图像集合可表示为对于所有采集的实景图像,从像素坐标[ut,vt,1]T到世界坐标转换关系可表示为:For example, coordinate transformation is performed on the real scene image captured by the k-th drone to obtain an M×N target image. The set of t real scene images captured by the k-th drone can be expressed as For all collected real-world images, from pixel coordinates [u t ,v t ,1] T to world coordinates The conversion relationship can be expressed as:
其中,R表示旋转矩阵,T表示平移向量,K表示相机的内参矩阵,Zc表示成像的比例因子。Among them, R represents the rotation matrix, T represents the translation vector, K represents the intrinsic parameter matrix of the camera, and Zc represents the imaging scale factor.
如图4所示,A(Xw,Yw,Zw)表示世界坐标系中的点A坐标;a(x,y)表示A在图像中的成像点,并以像素坐标系中的坐标(u,v)表示;Oc-XcYcZc表示相机坐标系,坐标原点为光心Oc;o-xy表示图像坐标系,以图像中点为光心;uv表示像素坐标系,而f表示相机焦距,即从图像中点o到相机光心Oc的距离:As shown in Figure 4, A( Xw , Yw , Zw ) represents the coordinates of point A in the world coordinate system; a(x, y) represents the imaging point of A in the image, and is expressed in coordinates (u, v) in the pixel coordinate system; Oc - XcYcZc represents the camera coordinate system , with the origin of the coordinate system as the optical center Oc ; o-xy represents the image coordinate system, with the midpoint of the image as the optical center; uv represents the pixel coordinate system, and f represents the focal length of the camera, that is, the distance from the midpoint o of the image to the optical center Oc of the camera:
f=||o-Oc|| (2)f=||oO c || (2)
关于从像素坐标系转换到世界坐标系的原理,下面进行详细介绍:The principle of converting from pixel coordinate system to world coordinate system is introduced in detail below:
首先,为实现从世界到相机坐标系的转换,可采用刚体变换。刚体变换可以通过旋转、平移等方式改变相机坐标系的空间位置与朝向。常用的表示刚体变换的方法是旋转矩阵R和平移向量T:First, to achieve the conversion from the world to the camera coordinate system, a rigid body transformation can be used. Rigid body transformation can change the spatial position and orientation of the camera coordinate system by rotation, translation, etc. The commonly used method to represent rigid body transformation is the rotation matrix R and the translation vector T:
其次,为实现相机坐标系到图像坐标系的转换,可使用透视投影。透视投影可以将三维空间中的点P转换为二维像平面上的点,与像平面的交点即为点p,过程可表示为:Secondly, to achieve the conversion from the camera coordinate system to the image coordinate system, perspective projection can be used. Perspective projection can convert point P in three-dimensional space into a point on a two-dimensional image plane. The intersection point with the image plane is point p. The process can be expressed as:
再次,为修正非线性成像的坐标偏差,可采用畸变矫正。畸变矫正能够将图像中物体的形状和大小能够在平面上准确地呈现出来。同时,需要将光信号数字化为电信号,以完成从真实世界到像素坐标系的转换,过程可表示如下:Thirdly, to correct the coordinate deviation of nonlinear imaging, distortion correction can be used. Distortion correction can accurately present the shape and size of objects in the image on a plane. At the same time, the optical signal needs to be digitized into an electrical signal to complete the conversion from the real world to the pixel coordinate system. The process can be expressed as follows:
其中,o(uo,vo)表示在uv坐标系下的坐标,dx、dy为相机内参,分别表示像平面坐标系中每个像素在横纵轴上的物理尺寸。Among them, o(u o , vo ) represents the coordinate in the uv coordinate system, dx and dy are the camera intrinsic parameters, which respectively represent the physical size of each pixel in the image plane coordinate system on the horizontal and vertical axes.
因此,世界到像素坐标系的关系如下:Therefore, the relationship between world and pixel coordinates is as follows:
从而,可通过下述表达式推算得到像素坐标到真实世界坐标的转换关系:Therefore, the conversion relationship from pixel coordinates to real-world coordinates can be calculated by the following expression:
其中,K-1=(EF)-1,E表示像素与像平面的转换矩阵,F表示相机内参矩阵。Wherein, K -1 =(EF) -1 , E represents the transformation matrix between the pixel and the image plane, and F represents the camera intrinsic parameter matrix.
在步骤1022中,根据各目标图像的世界坐标对各目标图像进行拼接处理,得到拼接图像。In step 1022, each target image is spliced according to its world coordinates to obtain a spliced image.
在一种可能的实施方式中,步骤1022可以包括以下子步骤:In a possible implementation, step 1022 may include the following sub-steps:
步骤10221、针对任意两张目标图像,根据两张目标图像的世界坐标判断两张目标图像是否存在关联区域;Step 10221: for any two target images, determine whether there is a related area between the two target images according to the world coordinates of the two target images;
步骤10222、在两张目标图像存在关联区域的情况下,采用尺度不变的特征点检测算法对两张目标图像分别进行特征检测,得到特征点集;Step 10222: When there are associated regions between the two target images, a scale-invariant feature point detection algorithm is used to perform feature detection on the two target images to obtain a feature point set.
步骤10223、计算特征点集中每个特征点的描述符,得到描述符集;Step 10223, calculating the descriptor of each feature point in the feature point set to obtain a descriptor set;
步骤10224、基于两张目标图像对应的描述符集,对两张目标图像对应的特征点集进行相似特征点的匹配处理,得到多个匹配点;Step 10224: Based on the descriptor sets corresponding to the two target images, matching processing of similar feature points is performed on the feature point sets corresponding to the two target images to obtain multiple matching points;
步骤10225、将多个匹配点中的异常匹配点进行过滤,以对多个匹配点进行矫正;Step 10225: filtering abnormal matching points among the multiple matching points to correct the multiple matching points;
步骤10226、将两张目标图像分别进行透视变换;Step 10226, performing perspective transformation on the two target images respectively;
步骤10227、基于矫正后的多个匹配点,对透视变换后的两张目标图像进行图像融合,得到拼接图像;其中,在图像融合的过程中,采用羽化技术进行平滑的图像转换,以及通过加权平均颜色值对重叠的像素进行融合。Step 10227: Based on the corrected multiple matching points, the two target images after perspective transformation are fused to obtain a stitched image; wherein, in the process of image fusion, feathering technology is used to perform smooth image conversion, and overlapping pixels are fused by weighted average color value.
下面结合图5对上述步骤10221-步骤10227进行说明:The above steps 10221 to 10227 are described below in conjunction with FIG. 5 :
在步骤10221中,针对任意两张目标图像,根据两张目标图像的世界坐标判断这两张目标图像是否邻接,也即是否存在关联区域。具体地,可以通过判断一张目标图像的四个端点坐标填入区域是否被另一张目标图像覆盖,来判断这两张目标图像是否邻接。其中,端点坐标(xi,yi)对应的覆盖标志可表示如下:以此决定多无人机拍摄图片拼接的执行与否In step 10221, for any two target images, it is determined whether the two target images are adjacent, that is, whether there is a related area, based on the world coordinates of the two target images. Specifically, it is possible to determine whether the two target images are adjacent by determining whether the four endpoint coordinates of one target image are filled in by another target image. Among them, the coverage flag corresponding to the endpoint coordinates (x i , y i ) can be expressed as follows: This determines whether the stitching of multiple drone images is executed or not.
可以看出,端点坐标(xi,yi)对应的覆盖标志为True,则表示这两张目标图像邻接,需要执行这两张目标图像的拼接;端点坐标(xi,yi)对应的覆盖标志为False,则表示这两张目标图像不邻接,无需需执行这两张目标图像的拼接。It can be seen that if the coverage flag corresponding to the endpoint coordinates (x i , y i ) is True, it means that the two target images are adjacent and need to be stitched together; if the coverage flag corresponding to the endpoint coordinates (x i , y i ) is False, it means that the two target images are not adjacent and do not need to be stitched together.
在步骤10222中,示例性地,以图6a所示的目标图像1和图6b所示的目标图像2为两张目标图像,在两张目标图像存在关联区域的情况下,采用尺度不变的特征点检测算法对两张目标图像分别进行特征检测,得到如图6c和图6d所示的特征点集。In step 10222, exemplarily, target image 1 shown in FIG. 6a and target image 2 shown in FIG. 6b are taken as two target images. When there are associated areas between the two target images, a scale-invariant feature point detection algorithm is used to perform feature detection on the two target images respectively, and feature point sets as shown in FIG. 6c and FIG. 6d are obtained.
在步骤10224中,如图6c和图6d所示,两幅图像中存在着大量的相似特征点。可以通过暴力匹配法(BF Matcher,BF)、K最邻近法(k-NearestNeighbor,KNN)等线性回归模型算法进行相似特征点的匹配处理,得到多个匹配点。In step 10224, as shown in FIG6c and FIG6d, there are a large number of similar feature points in the two images. Similar feature points can be matched by using a linear regression model algorithm such as a brute force matching method (BF Matcher, BF) and a K-Nearest Neighbor (KNN) method to obtain multiple matching points.
关于BF匹配法,通过迭代求解欧式距离最近的两组描述符作为匹配项。Regarding the BF matching method, the two groups of descriptors with the closest Euclidean distance are iteratively solved as matching items.
关于KNN算法,对于一幅图像中的每个描述符,在另一幅图像的描述符集中找到K个最近邻的点,其中,最近邻被认为是两幅图像中关键点之间的真实对应关系。KNN算法由于KNN特征点匹配的高效性,采取KNN进行特征匹配的结果如图7所示。Regarding the KNN algorithm, for each descriptor in one image, the K nearest neighbor points are found in the descriptor set of the other image, where the nearest neighbor is considered to be the true correspondence between the key points in the two images. Due to the high efficiency of KNN feature point matching, the result of using KNN for feature matching is shown in Figure 7.
在步骤10225中,为了应对线性回归模型对异常值不敏感的问题,本步骤需要进一步对异常匹配点进行过滤。因此,需要使用一种变换矩阵来描述并获取匹配点之间的关系,这种变换矩阵被称为单应矩阵。在单应矩阵的估计中,不属于重要区域的不需要的特征点将被删除,以避免噪声的影响。为了排除线性回归模型引入的异常值影响,采用随机抽样一致(RANdom Sample Consensus,RANSAC)算法进行单应矩阵的计算。RANSAC是一种经典的拟合算法,它可以在存在噪声和异常值的数据集中估计模型参数。基本思想是随机选择小部分数据,拟合模型并计算误差。多次迭代后,选择误差最小的模型作为输出,以此获取更为鲁棒的匹配结果。In step 10225, in order to deal with the problem that the linear regression model is insensitive to outliers, this step needs to further filter the abnormal matching points. Therefore, it is necessary to use a transformation matrix to describe and obtain the relationship between the matching points. This transformation matrix is called a homography matrix. In the estimation of the homography matrix, unnecessary feature points that do not belong to important areas will be deleted to avoid the influence of noise. In order to eliminate the influence of outliers introduced by the linear regression model, the random sampling consensus (RANdom Sample Consensus, RANSAC) algorithm is used to calculate the homography matrix. RANSAC is a classic fitting algorithm that can estimate model parameters in a data set with noise and outliers. The basic idea is to randomly select a small part of the data, fit the model and calculate the error. After multiple iterations, the model with the smallest error is selected as the output to obtain a more robust matching result.
在步骤10226和步骤10227中,为获得期望的输出图像形状,采用透视变换对所有输入图像进行变形,并在经过变换的图像上进行融合。该过程需要计算每个输入图像的变形图像坐标范围,以确定输出图像的尺寸,并映射每个源图像的四个角到变换后的图像上。接下来,使用羽化等技术实现平滑的图像转换,通过加权平均颜色值来融合重叠的像素,以避免拼接处的不连续和缝隙出现。从而完成了所有的图像拼接步骤,得到如图8所示的拼接图像。In step 10226 and step 10227, in order to obtain the desired output image shape, all input images are deformed by perspective transformation and fused on the transformed images. This process requires calculating the deformed image coordinate range of each input image to determine the size of the output image and mapping the four corners of each source image to the transformed image. Next, feathering and other techniques are used to achieve smooth image conversion, and the overlapping pixels are fused by weighted average color values to avoid discontinuity and gaps at the splicing. All image splicing steps are completed, and a spliced image as shown in Figure 8 is obtained.
在该实施方式中,由于在特征匹配时过滤了异常匹配点,可以避免噪声的影响;以及,在图像融合时采用羽化技术进行平滑的图像转换,以及通过加权平均颜色值对重叠的像素进行融合,可以避免拼接处的不连续和缝隙出现,从而可以提升图像拼接的准确性。In this embodiment, since abnormal matching points are filtered during feature matching, the influence of noise can be avoided; and, feathering technology is used to perform smooth image conversion during image fusion, and overlapping pixels are fused by weighted average color values, which can avoid discontinuities and gaps at the splicing points, thereby improving the accuracy of image stitching.
在步骤1023中,向拼接图像未采集到的空缺区域进行像素和纹理的补全,得到实时灾情感知态势图。In step 1023, pixels and textures are supplemented in the missing areas that are not captured in the stitched image to obtain a real-time disaster awareness situation map.
在一种可能的实施方式中,步骤1023可以包括以下子步骤:In a possible implementation, step 1023 may include the following sub-steps:
步骤10231、对拼接图像中的空缺区域进行低维视觉先验信息的重建;Step 10231, reconstructing low-dimensional visual prior information of the vacant area in the spliced image;
步骤10232、基于低维视觉先验信息,向重建后的拼接图像中补充纹理细节信息,得到实时灾情感知态势图。Step 10232: Based on the low-dimensional visual prior information, texture detail information is added to the reconstructed spliced image to obtain a real-time disaster awareness situation map.
下面结合图9对步骤10231和步骤10232进行说明:Step 10231 and step 10232 are described below in conjunction with FIG. 9 :
在步骤10231之前,首先针对拼接图像进行判断,即针对拼接图像中的位置坐标(xi,yi),通过掩码进行空缺区域表示,如下式所示:Before step 10231, the stitched image is first judged, that is, the position coordinates (x i , y i ) in the stitched image are represented by a mask to indicate the missing area, as shown in the following formula:
进而根据算法对空缺区域,即掩码为0的部分执行补全算法。Then, the algorithm is used to complete the missing area, that is, the part where the mask is 0.
可以看出,掩码为1,表示存在采集点,无需执行补全算法;掩码为0,表示缺失采集点,执行补全算法。It can be seen that when the mask is 1, it means that there are acquisition points and there is no need to execute the completion algorithm; when the mask is 0, it means that the acquisition points are missing and the completion algorithm is executed.
在步骤10231中,低维视觉先验信息可以包括低维的初步重建结构和粗纹理。对拼接图像中的空缺区域进行初步的重建,得到拼接图像中的空缺区域的低维的初步重建结构和粗纹理。In step 10231, the low-dimensional visual prior information may include a low-dimensional preliminary reconstruction structure and a coarse texture. The vacant area in the spliced image is preliminarily reconstructed to obtain a low-dimensional preliminary reconstruction structure and a coarse texture of the vacant area in the spliced image.
示例性地,图像补全旨在将具有缺失像素的拼接图像转换为完整图像图像补全任务具有随机性,因此,给定拼接图像Im,存在条件分布p(I∣Im)。由于在给定I和Im的情况下,粗略的先验信息X是确定的,因此p(I∣Im)可以重写为:For example, image completion aims to stitch images with missing pixels. Convert to full image The image completion task is stochastic, so given the spliced image I m , there is a conditional distribution p(I|I m ). Since the rough prior information X is certain given I and I m , p(I|I m ) can be rewritten as:
p(I∣Im)=p(I∣Im)·p(X∣I,Im)=p(X∣Im)·p(I∣X,Im) (10)p(I∣I m )=p(I∣I m )·p(X∣I,I m )=p(X∣I m )·p(I∣X,I m ) (10)
不同于直接从p(I∣Im)采样,本步骤首先使用Transformer模型对给定Im的视觉先验的基本分布进行建模,并将其表示为p(X∣Im)。由于Transformer具有强大的表示能力,这些重构的视觉先验包含全局结构和粗糙纹理的充分信息。Different from sampling directly from p(I|I m ), this step first uses the Transformer model to model the basic distribution of the visual prior given I m and represents it as p(X|I m ). Due to the powerful representation ability of Transformer, these reconstructed visual priors contain sufficient information of global structure and rough texture.
在步骤10232中,示例性地,采用卷积神经网络(Convolutional NeuralNetworks,CNN),在低维视觉先验信息和未Masked像素的指导下,补充精细的纹理细节,并将其表示为p(I∣X,Im)。In step 10232, illustratively, a convolutional neural network (CNN) is used to supplement fine texture details under the guidance of low-dimensional visual prior information and unmasked pixels, and it is represented as p(I|X,I m ).
在该实施方式中,首先对拼接图像中的空缺区域进行低维视觉先验信息的重建,然后基于低维视觉先验信息,向重建后的拼接图像中补充纹理细节信息,得到高质量图像,即实时灾情感知态势图,可以实现高保真度的多元图像补全性能。In this implementation, low-dimensional visual prior information is firstly reconstructed for the missing areas in the stitched image, and then, based on the low-dimensional visual prior information, texture detail information is supplemented to the reconstructed stitched image to obtain a high-quality image, i.e., a real-time disaster awareness situation map, which can achieve high-fidelity multivariate image completion performance.
在本实施例中,可以通过将多架无人机拍摄的各实景图像进行坐标转换、拼接处理和补全处理,构建实时灾情感知态势图,可以提升实时灾情感知效率。In this embodiment, the real-time disaster awareness situation map can be constructed by performing coordinate conversion, splicing processing and completion processing on the real-scene images taken by multiple drones, which can improve the efficiency of real-time disaster awareness.
在一种示例实施例中,步骤10231包括以下子步骤:In an exemplary embodiment, step 10231 includes the following sub-steps:
步骤102311、将拼接图像降采样为低维图像;Step 102311, downsampling the spliced image into a low-dimensional image;
步骤102312、将低维图像确定为拼接图像的视觉先验图像;Step 102312, determining the low-dimensional image as a visual prior image of the spliced image;
步骤102313、采用聚类算法在预设图像数据集的颜色空间中生成颜色字典;Step 102313, using a clustering algorithm to generate a color dictionary in the color space of a preset image data set;
步骤102314、针对视觉先验图像中的每个像素,在颜色字典中查找与像素最接近的元素索引,得到像素的离散表示;Step 102314: for each pixel in the visual prior image, find the element index closest to the pixel in the color dictionary to obtain a discrete representation of the pixel;
步骤102315、将视觉先验图像中各像素的离散表示输入Transformer模型中进行迭代计算,将拼接图像中的空缺区域中各像素的离散表示中的元素替换为吉布斯采样标记,得到离散序列;Step 102315, input the discrete representation of each pixel in the visual prior image into the Transformer model for iterative calculation, replace the elements in the discrete representation of each pixel in the vacant area in the spliced image with Gibbs sampling marks, and obtain a discrete sequence;
步骤102316、针对每个离散序列,通过查询颜色字典来重建低维视觉先验信息。Step 102316: For each discrete sequence, reconstruct low-dimensional visual prior information by querying the color dictionary.
下面结合图10对上述步骤102311-步骤102316进行说明:The above steps 102311 to 102316 are described below in conjunction with FIG. 10 :
在步骤102311中,示例性地,将拼接图像降采样为32×32或48×48的低维图像,本实施例不限于此。In step 102311, illustratively, the stitched image is downsampled to a low-dimensional image of 32×32 or 48×48, but this embodiment is not limited thereto.
在步骤102312中,将低维图像确定为拼接图像的视觉先验图像,视觉先验图像仅包含结构信息和粗略的纹理。In step 102312, the low-dimensional image is determined as a visual prior image of the stitched image, and the visual prior image only contains structural information and rough texture.
在步骤102313中,示例性地,聚类算法可以为K-Means聚类算法,预设图像数据集可以为ImageNet数据集,颜色空间可以为RGB空间。采用K-Means聚类算法在ImageNet数据集的RGB空间中生成一个大小为512×3的颜色字典,本实施例不限于此。In step 102313, illustratively, the clustering algorithm may be a K-Means clustering algorithm, the preset image data set may be an ImageNet data set, and the color space may be an RGB space. A color dictionary of size 512×3 is generated in the RGB space of the ImageNet data set using the K-Means clustering algorithm, but the present embodiment is not limited thereto.
在步骤102315中,Transformer模型的学习目标是将空缺区域的表示序列中的元素替换为特殊掩码标记(Masked Token),从而将视觉先验图像转换为离散序列。In step 102315, the learning goal of the Transformer model is to replace the elements in the representation sequence of the vacant area with a special masked token, thereby converting the visual prior image into a discrete sequence.
给定离散序列X={x1,x2,...,xL}中的每个标记,其中L是序列的长度,本实施例使用预定义的嵌入(Embedding)将其映射到d维特征向量中。为了将位置信息编码到输入中,本实施例添加了可学习的位置嵌入到每个位置标记的特征中,使得最终的Transformer模型的输入格式为 Given each token in a discrete sequence X = {x 1 ,x 2 ,...,x L }, where L is the length of the sequence, this embodiment uses a predefined embedding to map it into a d-dimensional feature vector. In order to encode position information into the input, this embodiment adds a learnable position embedding to each position The marked features make the input format of the final Transformer model
模型的网络架构基于唯一解码器(Decoder-only)的Transformer算法,主要由N个双向自注意力机制的Transformer层组成,在每个Transformer层中,计算公式如下:The network architecture of the model is based on the decoder-only Transformer algorithm, which mainly consists of N bidirectional self-attention Transformer layers. In each Transformer layer, the calculation formula is as follows:
Zl-1=LN(BMSA(Dl-1))+Dl-1 (11)Z l-1 =LN(BMSA(D l-1 ))+D l-1 (11)
Dl=LN(FC(Zl-1))+Zl-1 (12)D l =LN(FC(Z l-1 ))+Z l-1 (12)
其中,LN、BMSA、FC分别表示层归一化、多头自注意力层和全连接层。更具体地,给定输入D,BMSA可以计算为:Among them, LN, BMSA, and FC represent layer normalization, multi-head self-attention layer, and fully connected layer, respectively. More specifically, given the input D, BMSA can be calculated as:
BMSA=[head1;...;headh]WO (13)BMSA=[head 1 ;...;head h ]W O (13)
其中,head是头部,h是头部的数量,d为超参数,表示向量维度,和是三个可学习的线性投影层,1≤i≤h,WO为可学习的全连接层,其目标是融合不同头部的输出结果。Among them, head is the head, h is the number of heads, d is a hyperparameter, indicating the vector dimension, and are three learnable linear projection layers, 1≤i≤h, and WO is a learnable fully connected layer whose goal is to fuse the output results of different heads.
最终的Transformer层的输出会被映射到512个颜色字典元素上。这个映射通过一个全连接层和Softmax函数来完成。采用类似于掩码语言模型(Masked Language Model,MLM)的目标函数来优化Transformer模型。具体的,设Π={π1,π2,…,πK}表示离散输入中MaskedToken的索引,其中K是Masked Token的数量。设XΠ表示X中Masked Token的集合,X-Π表示未MaskedToken的集合。MLM的目标是在给定已知Token的情况下最小化Masked Token的负对数似然:The output of the final Transformer layer is mapped to 512 color dictionary elements. This mapping is done through a fully connected layer and a Softmax function. An objective function similar to the Masked Language Model (MLM) is used to optimize the Transformer model. Specifically, let Π = {π 1 ,π 2 ,…,π K } represent the index of the MaskedToken in the discrete input, where K is the number of Masked Tokens. Let X Π represent the set of Masked Tokens in X, and X -Π represent the set of unmaskedTokens. The goal of MLM is to minimize the negative log-likelihood of the Masked Token given a known Token:
其中,LMLM是MLM的似然求解值,θ是Transformer的参数。结合双向注意的MLM目标确保了Transformer模型能够捕获整个上下文信息,以预测缺失区域的概率分布。为了生成可信的Transformer模型的分布,直接采样整个Masked位置集合会产生不可信的结果,因为这些位置是相互独立的。相反,本实施例使用吉布斯采样来迭代地在不同的位置上采样Token。具体的,在每次迭代时,依次从条件分布中使用前k个预测元素来采样网格位置,其中表示之前生成的Token。然后,本实施例将相应的Masked Token替换为采样的Token,并重复此过程,直到更新所有位置。通过采样,可以得到一组完整的Token序列。Where L MLM is the likelihood solution of MLM, and θ is the parameter of Transformer. The MLM objective combined with bidirectional attention ensures that the Transformer model can capture the entire context information to predict the probability distribution of the missing region. In order to generate a credible distribution of the Transformer model, directly sampling the entire set of Masked positions will produce untrustworthy results because these positions are independent of each other. Instead, this embodiment uses Gibbs sampling to iteratively sample tokens at different positions. Specifically, at each iteration, the conditional distribution is sequentially sampled from The first k predicted elements are used to sample grid positions, where Represents the previously generated Token. Then, this embodiment replaces the corresponding Masked Token with the sampled Token and repeats this process until all positions are updated. Through sampling, a complete set of Token sequences can be obtained.
在步骤102316中,对于从Transformer模型中采样得到的每个完整离散序列,可以通过查询颜色字典来重建其视觉先验从而获得合理和多样化的视觉先验信息。因此便完成了低维视觉先验信息的构建。In step 102316, for each complete discrete sequence sampled from the Transformer model, its visual prior can be reconstructed by querying the color dictionary Thus, reasonable and diverse visual prior information is obtained, thus completing the construction of low-dimensional visual prior information.
在本实施例中,利用Transformer模型强大的表示能力,重建低维视觉先验信息。In this embodiment, the powerful representation capability of the Transformer model is used to reconstruct low-dimensional visual prior information.
在一种示例实施例中,步骤10232包括以下子步骤:In an exemplary embodiment, step 10232 includes the following sub-steps:
步骤102321、将重建后的视觉先验图像的图像矩阵和图像矩阵的双线性插值结果输入到引导上采样网络中进行处理,得到预测值;Step 102321, input the image matrix of the reconstructed visual prior image and the bilinear interpolation result of the image matrix into the guided upsampling network for processing to obtain a predicted value;
步骤102322、基于预测值和真实值之间的损失函数来调整引导上采样网络的模型参数;损失函数为L1损失函数和对抗性损失函数的加权和;对抗性损失函数与判别器对预测值的判别值相关;Step 102322, adjusting the model parameters of the guided upsampling network based on the loss function between the predicted value and the true value; the loss function is a weighted sum of the L1 loss function and the adversarial loss function; the adversarial loss function is related to the discriminant value of the discriminator to the predicted value;
步骤102323、联合训练引导上采样网络和判别器,得到多元化还原结果;Step 102323, jointly train the guided upsampling network and the discriminator to obtain diversified restoration results;
步骤102324、从多元化还原结果中确定鲁棒性最高的结果;Step 102324, determining the most robust result from the diversified restoration results;
步骤102325、基于鲁棒性最高的结果向重建后的拼接图像中补充纹理细节信息,得到实时灾情感知态势图。Step 102325: based on the result with the highest robustness, add texture detail information to the reconstructed spliced image to obtain a real-time disaster awareness situation map.
下面结合图11对上述步骤102321-步骤102325进行说明:The above steps 102321 to 102325 are described below in conjunction with FIG. 11 :
在步骤102321中,完成重建低维视觉先验后,为恢复原始分辨率为H×W×3的数据,并且保持Masked和未Masked区域之间的边界一致性,引入了引导上采样网络。该网络利用神经网络建模纹理模式,通过Masked输入Im的指导下,渲染重建的视觉先验,从而实现高保真细节的重建。其中,引导上采样处理可表示为:In step 102321, after completing the reconstruction of the low-dimensional visual prior, in order to restore the data with the original resolution of H×W×3 and maintain the boundary consistency between the masked and unmasked areas, a guided upsampling network is introduced. The network uses a neural network to model the texture pattern, and renders the reconstructed visual prior under the guidance of the masked input Im , thereby achieving reconstruction of high-fidelity details. Among them, the guided upsampling process can be expressed as:
其中,表示原始数据X重塑后的矩阵,是It的双线性插值结果,∩表示级联操作,Ipred表示引导上采样网络的预测值。是由δ参数化的引导上采样网络的主干,包含编码器、解码器以及多个残差块。in, Represents the matrix after the original data X is reshaped, is the bilinear interpolation result of It , ∩ represents the cascade operation, and Ipred represents the predicted value of the guided upsampling network. It is the backbone of the guided upsampling network parameterized by δ, consisting of an encoder, a decoder, and multiple residual blocks.
在步骤102322中,通过最小化引导上采样网络预测值Ipred和真实值I之间的l1损失来进行优化网络性能:In step 102322, the network performance is optimized by minimizing the l 1 loss between the guided upsampling network prediction value I pred and the true value I:
其中,表示l1损失的期望。in, represents the expectation of l1 loss.
为了在训练过程中生成更真实的细节,引入了额外的对抗性损失:In order to generate more realistic details during training, an additional adversarial loss is introduced:
其中,Ladv表示对抗性损失的期望,是由ω参数化的判别器。Among them, L adv represents the expectation of adversarial loss, is the discriminator parameterized by ω.
在步骤102323中,通过解决以下优化问题,联合训练上采样网络F和判别器 In step 102323, the upsampling network F and the discriminator are jointly trained by solving the following optimization problem
其中,Lupsample表示l1损失与对抗性损失的上采样的期望,α1和α2分别为l1损失权重与对抗性损失权重。Among them, L upsample represents the expectation of upsampling of l 1 loss and adversarial loss, α1 and α2 are the l 1 loss weight and adversarial loss weight, respectively.
在步骤102324中,该算法结合多种因素,如灾情蔓延态势、灾点置信度、图像生成指标和指挥人员评估等,在生成的多元化还原结果中选择鲁棒性最高的结果进行输出。In step 102324, the algorithm combines multiple factors, such as the spread of the disaster, the confidence level of the disaster site, image generation indicators, and commander evaluation, and selects the most robust result from the generated diversified restoration results for output.
在步骤102325中,基于鲁棒性最高的结果向重建后的拼接图像中补充纹理细节信息,得到实时灾情感知态势图。图11展示了根据灾点检测置信度所输出的最佳还原效果。In step 102325, texture detail information is added to the reconstructed spliced image based on the result with the highest robustness, so as to obtain a real-time disaster awareness situation map. FIG11 shows the best restoration effect output according to the confidence of disaster point detection.
在本实施例中,可以向重建后的拼接图像中补充纹理细节信息。In this embodiment, texture detail information may be supplemented into the reconstructed stitched image.
在一种示例实施例中,步骤104可以包括以下子步骤:In an exemplary embodiment, step 104 may include the following sub-steps:
步骤1041、确定通信定位需求分级模型的输入指标;输入指标包括服务对象、应急场景、处置阶段和事件客观因素的细化类型以及通信定位需求的各个等级;Step 1041: Determine input indicators of the communication positioning demand grading model; the input indicators include service objects, emergency scenarios, handling stages, detailed types of objective factors of events, and various levels of communication positioning demand;
步骤1042、将输入指标输入通信定位需求分级模型中进行处理,输出通信定位需求分级标定结果;Step 1042: input the input index into the communication positioning demand grading model for processing, and output the communication positioning demand grading calibration result;
步骤1043、基于实时灾情感知态势图确定灾点标定结果;Step 1043, determining the disaster point calibration result based on the real-time disaster awareness situation map;
步骤1044、将通信定位需求分级标定结果和灾点标定结果添加至实时灾情感知态势图中,得到实景地图。Step 1044: Add the communication positioning demand classification calibration results and the disaster point calibration results to the real-time disaster awareness situation map to obtain a real-scene map.
在步骤1041中,历史分级数据包括分级指标和针对每个分级指标的专家评语标签。示例性地,如图12所示,分级指标可以包括:一级指标、二级指标、…、五级指标;其中,一级指标包括通信和定位任务的优先级,二级指标包括A服务对象、B应急场景、C处置阶段、D事件客观因素、E灾点检测置信度,以此类推。In step 1041, the historical classification data includes classification indicators and expert comment labels for each classification indicator. Exemplarily, as shown in FIG12, the classification indicators may include: first-level indicators, second-level indicators, ..., and fifth-level indicators; wherein the first-level indicators include the priority of communication and positioning tasks, and the second-level indicators include A service objects, B emergency scenarios, C handling stages, D objective factors of events, E disaster point detection confidence, and so on.
如图13所示,对A服务对象进行了细化,A服务对象可以包括A1人员类型和A2载具类型;其中,A1人员类型可以包括A1-1一线小队、A1-2前方指挥员、A1-3后方指挥员和A1-4保障人员;A2载具类型可以包括A2-1通信车和A2-2指挥车。As shown in Figure 13, the A service objects are refined, and the A service objects can include A1 personnel types and A2 vehicle types; among them, A1 personnel types can include A1-1 front-line team, A1-2 front commander, A1-3 rear commander and A1-4 support personnel; A2 vehicle types can include A2-1 communication vehicle and A2-2 command vehicle.
如图14所示,对B应急场景进行了细化,B应急场景可以包括B1地形地貌和B2天气气候;其中,B1地形地貌可以包括B1-1丘陵、B1-2平原、B1-3盆地、B1-4山地、B1-5林区和B1-X其他;B1-5林区可以包括B1-5-1林相、B1-5-2树种和B1-5-X其他;B2天气气候可以包括B2-1气温、B2-2湿度、B2-3风、B2-4降雨和B2-X其他;B2-3风可以包括B2-3-1风力和B2-3-2风向。As shown in Figure 14, the B emergency scenario is refined and can include B1 topography and B2 weather and climate; among them, B1 topography can include B1-1 hills, B1-2 plains, B1-3 basins, B1-4 mountains, B1-5 forest areas and B1-X others; B1-5 forest areas can include B1-5-1 forest phases, B1-5-2 tree species and B1-5-X others; B2 weather and climate can include B2-1 temperature, B2-2 humidity, B2-3 wind, B2-4 rainfall and B2-X others; B2-3 wind can include B2-3-1 wind force and B2-3-2 wind direction.
如图15所示,对C处置阶段进行了细化,C处置阶段可以包括C1事前、C2事发、C3事中和C4事后;其中,C1事前可以包括C1-1灾情接报和C1-2预先勘察,C2事发可以包括C2-1兵力接送,C3事中可以包括C3-1探测阶段和C3-2处置阶段,C4事后可以包括C4-1清理灾区、C4-2撤离返营和C4-3灾后重建。As shown in Figure 15, the C handling stage is refined, and the C handling stage can include C1 before the event, C2 after the event, C3 during the event, and C4 after the event; among them, C1 before the event can include C1-1 disaster reporting and C1-2 preliminary investigation, C2 after the event can include C2-1 troop transfer, C3 during the event can include C3-1 detection stage and C3-2 handling stage, and C4 after the event can include C4-1 clearing the disaster area, C4-2 evacuation and return to the camp, and C4-3 post-disaster reconstruction.
如图16所示,对D事件客观因素进行了细化,D事件客观因素可以包括D1事件类型、D2灾情态势和D3事件规模;其中,D1事件类型可以包括D1-1平时和D1-2战时,D1-1平时可以包括日常巡护,D1-2战时可以包括D1-2-1火灾、D1-2-2泥石流、D1-2-3地震和D1-2-X其他;D2灾情态势可以包括D2-1灾点、D2-2潜在灾点和D2-3非灾点,D3事件规模可以包括D3-1受灾面积和D3-2受灾人数。As shown in Figure 16, the objective factors of D events are refined, and the objective factors of D events may include D1 event type, D2 disaster situation and D3 event scale; among them, D1 event type may include D1-1 peacetime and D1-2 wartime, D1-1 peacetime may include daily patrols, D1-2 wartime may include D1-2-1 fire, D1-2-2 mudslide, D1-2-3 earthquake and D1-2-X others; D2 disaster situation may include D2-1 disaster point, D2-2 potential disaster point and D2-3 non-disaster point, and D3 event scale may include D3-1 disaster-affected area and D3-2 number of affected people.
如图17所示,通信定位需求的各个等级分为五个级别,随着等级从L1到L5递增,对应区域需求的重要性程度也逐渐增加。例如,在救援现场相对安全区域,仅需基本的语音传输和定位需求;而越靠近灾点,现场实时信息更重要,需要更快的通信速率以保障图像和视频传输,需要更高的定位精度以保障救援的安全性。As shown in Figure 17, the various levels of communication and positioning requirements are divided into five levels. As the level increases from L1 to L5, the importance of the corresponding regional requirements also gradually increases. For example, in a relatively safe area at the rescue site, only basic voice transmission and positioning requirements are required; the closer to the disaster site, the more important the real-time information on the scene is, and a faster communication rate is required to ensure image and video transmission, and a higher positioning accuracy is required to ensure the safety of the rescue.
在步骤1042中,示例性地,如图18所示,首先,根据历史的应急救援中收集的指标数据以及专家分级判断将数据录入灾情分级专家知识库,并通过一维卷积神经网络进行模型训练。一维卷积网络是指对一维输入向量进行卷积、池化、全连接等操作的神经网络,能够根据输入的灾点情况、服务对象、应急场景、处置阶段以及区域人数推理出通信、定位需求等级。In step 1042, illustratively, as shown in FIG18, first, the data is entered into the disaster classification expert knowledge base based on the index data collected in the historical emergency rescue and the expert classification judgment, and the model is trained through a one-dimensional convolutional neural network. A one-dimensional convolutional network refers to a neural network that performs convolution, pooling, full connection and other operations on a one-dimensional input vector, and can infer the communication and positioning demand level based on the input disaster point situation, service object, emergency scenario, disposal stage and number of people in the area.
在进行前向传递之前,需要对输入数据进行One-hot编码。输入数据可以被表示为一个大小为L×F的矩阵,其中每行代表一个样本,每列代表一个特征。接下来,将此矩阵输入卷积层,从而进行特征提取和降维。一维卷积层可以将输入数据序列通过一组可学习的卷积核进行滤波,以提取数据中的重要特征。在卷积操作后,一维卷积层需采用激活函数对结果进行非线性变换。其中,ReLU激活函数是一种常用的非线性激活函数,在应对梯度消失问题方面具有优异的性能,因此采用ReLU激活函数进行处理,其定义为f(x)=max(0,x)。接下来,通过池化层减小输出的维度,最终通过全连接层输出通信定位需求分级的结果。Before forward pass, the input data needs to be One-hot encoded. The input data can be represented as a matrix of size L×F, where each row represents a sample and each column represents a feature. Next, this matrix is input into the convolution layer for feature extraction and dimensionality reduction. The one-dimensional convolution layer can filter the input data sequence through a set of learnable convolution kernels to extract important features from the data. After the convolution operation, the one-dimensional convolution layer needs to use an activation function to perform a nonlinear transformation on the result. Among them, the ReLU activation function is a commonly used nonlinear activation function that has excellent performance in dealing with the problem of gradient disappearance. Therefore, the ReLU activation function is used for processing, which is defined as f(x)=max(0,x). Next, the output dimension is reduced through the pooling layer, and finally the result of communication positioning demand classification is output through the fully connected layer.
在训练过程中,需要对网络输出和真实标签之间的差异进行度量,以便调整模型参数以提高性能。Categorical_crossentropy损失函数能够量化网络输出和真实标签之间的差异,并使用反向传播算法计算每个参数的梯度,以便沿着梯度的方向更新参数以减小交叉熵损失函数,交叉熵损失函数J()可表示为:During the training process, the difference between the network output and the true label needs to be measured in order to adjust the model parameters to improve performance. The categorical_crossentropy loss function can quantify the difference between the network output and the true label, and use the back propagation algorithm to calculate the gradient of each parameter so that the parameters can be updated along the direction of the gradient to reduce the cross entropy loss function. The cross entropy loss function J() can be expressed as:
其中,C是类别的数量,y是一个C维向量,表示数据样本的真实类别标签。是一个C维向量,表示模型的输出结果,每个元素的值为0到1之间的实数,且所有元素之和为1。yi表示真实类别为i的概率,而则表示模型预测为类别i的概率,其值通常使用Softmax函数来计算。通过将输入Softmax函数,可以将其转换为一个概率向量,满足所有元素之和为1。Softmax函数σ()可表示为:Where C is the number of categories and y is a C-dimensional vector representing the true category label of the data sample. is a C-dimensional vector representing the output of the model. The value of each element is a real number between 0 and 1, and the sum of all elements is 1. yi represents the probability that the true category is i, and It represents the probability of the model predicting that it is category i, and its value is usually calculated using the Softmax function. Input the Softmax function and convert it into a probability vector so that the sum of all elements is 1. The Softmax function σ() can be expressed as:
将Softmax函数的计算结果代入Categorical_crossentropy损失函数公式中,可以得到模型预测结果和真实标签y之间的交叉熵损失,用于衡量模型预测的准确度。Substituting the calculation result of the Softmax function into the Categorical_crossentropy loss function formula, we can get the model prediction result The cross entropy loss between y and the true label y is used to measure the accuracy of the model prediction.
同时,为了提高训练效率和准确性,本实施例采用Adam优化器对神经网络中的每个权重参数进行更新优化。具体地,Adam优化器会计算每个权重参数θt的一阶矩估计mt和二阶矩估计vt的梯度,并对这些估计值进行偏差修正,得到和然后根据这些估计值计算出一个自适应的学习率,并用该学习率来更新参数θt的值。Adam优化器的更新规则可以表示为:At the same time, in order to improve training efficiency and accuracy, this embodiment uses the Adam optimizer to update and optimize each weight parameter in the neural network. Specifically, the Adam optimizer calculates the gradient of the first-order moment estimate m t and the second-order moment estimate v t of each weight parameter θ t , and performs deviation correction on these estimates to obtain and Then, an adaptive learning rate is calculated based on these estimates, and the learning rate is used to update the value of the parameter θ t . The update rule of the Adam optimizer can be expressed as:
其中,mt-1是更新前的一阶矩估计,vt-1是更新前的二阶矩估计,θt-1是更新前的权重参数,J(θt-1)是损失函数,是损失函数关于参数θ的梯度,θt是更新后的参数值,η是学习率,∈是一个很小的常数,防止分母为零,β1和β2分别是一阶和二阶矩值计的衰减率,旨在平滑估计值。Among them, m t-1 is the first-order moment estimate before updating, v t-1 is the second-order moment estimate before updating, θ t-1 is the weight parameter before updating, J(θ t-1 ) is the loss function, is the gradient of the loss function with respect to the parameter θ, θt is the updated parameter value, η is the learning rate, ∈ is a small constant to prevent the denominator from being zero, and β1 and β2 are the decay rates of the first-order and second-order moment meters, respectively, which aim to smooth the estimated value.
在步骤1044中,根据输出结果以及区域分级的粒度信息将通信定位需求分级标定结果和灾点标定结果在实景态势图中标定,以协助指挥人员更高效地做出更合理的通信定位需求分级决策。并且在每次实地救援中使用实地数据来微调网络模型,从而对模型进行调整和改进,以进一步提升模型在实际应急救援中的表现。In step 1044, the communication positioning demand classification calibration results and disaster point calibration results are calibrated in the real-life situation map based on the output results and the granularity information of the regional classification, so as to assist the commanders in making more efficient and reasonable communication positioning demand classification decisions. In addition, the network model is fine-tuned using field data in each field rescue, so as to adjust and improve the model to further improve the performance of the model in actual emergency rescue.
在本实施例中,通过实时灾情感知态势图可以进行灾点标定,通过通信定位需求分级模型可以进行通信定位需求分级标定,从而得到基于实时灾情感知态势图和通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图,可以提升通信定位需求分级的准确性。In this embodiment, disaster points can be calibrated through the real-time disaster awareness situation map, and communication positioning demand grading calibration can be performed through the communication positioning demand grading model, so as to obtain a real-time disaster awareness situation map and the communication positioning demand grading model, and construct a real-life map based on the communication positioning demand grading calibration and disaster point calibration, which can improve the accuracy of communication positioning demand grading.
下面通过实验结果对本发明实施例的应急场景下的实景地图构建方法的优点进行详细说明。The advantages of the method for constructing a real scene map in an emergency scenario according to an embodiment of the present invention are described in detail below through experimental results.
首先,搭建实验仿真平台,实验仿真平台为Python平台,Python平台采用PyTorch框架实现。PyTorch框架是由Facebook开发的深度学习框架,其优势包括动态计算图、GPU加速、丰富的工具库等。仿真所用GPU为RTX3090。表1示出了主要仿真参数:First, build an experimental simulation platform. The experimental simulation platform is a Python platform, which is implemented using the PyTorch framework. The PyTorch framework is a deep learning framework developed by Facebook. Its advantages include dynamic computational graphs, GPU acceleration, and a rich tool library. The GPU used for simulation is RTX3090. Table 1 shows the main simulation parameters:
表1主要仿真参数表Table 1 Main simulation parameters
如图19所示,是无人机采集的部分灾点实景信息。针对灾情区域全局采样的实景地图生成情况,其中包括了无人机航拍采集的全局图片信息,并融合了Unity3D制造的火焰信息,经过图片坐标转换后,将采集图片填入地图。As shown in Figure 19, some of the real scene information of the disaster site collected by the drone is shown. The real scene map of the global sampling of the disaster area is generated, which includes the global image information collected by the drone aerial photography and integrates the flame information created by Unity3D. After the image coordinates are converted, the collected images are filled into the map.
在救援过程中,获取受灾区域全局的实景照片及其灾点细节对于有效应对灾情至关重要。然而,若灾情区域扩大,灾情态势蔓延快。一方面,全局采样时延很高,难以支持应急救援场景下实景态势图的新鲜度要求。另一方面,仅进行拼接操作需要前后两帧图像有着关联关系,意味着需要采集更多具有重复部分的图片,十分消耗应急救援场景下的带宽与算力资源。而本实施例进行部分采样操作,将采集到的图片根据坐标关系拼入地图中,若重复则进行拼接操作。完成上述操作后,将空缺区域进行图像补全,能够极大降低时延与带宽、算力消耗,并且支持无人机实时采集图片更新实景地图,能够保持较高的新鲜度。根据上述无人机采集的不同特征和拍摄位置的图片,本实施例基于坐标转换模块进行了图像拼接。经过对这些局部图片的细致处理和合理组合,成功地将它们拼接成了一张如图20所示的完整的、高分辨率的全景图像。During the rescue process, it is crucial to obtain global real-life photos of the affected area and the details of the disaster points for effectively responding to the disaster. However, if the disaster area expands, the disaster situation will spread quickly. On the one hand, the global sampling delay is very high, and it is difficult to support the freshness requirements of the real-life situation map in the emergency rescue scenario. On the other hand, only the splicing operation requires that the two frames before and after the image have a correlation, which means that more pictures with repeated parts need to be collected, which consumes bandwidth and computing power resources in the emergency rescue scenario. In this embodiment, partial sampling operations are performed, and the collected pictures are stitched into the map according to the coordinate relationship. If they are repeated, the splicing operation is performed. After completing the above operations, the vacant area is image-filled, which can greatly reduce the latency and bandwidth and computing power consumption, and support the drone to collect pictures in real time to update the real-life map, which can maintain a high degree of freshness. According to the pictures of different features and shooting positions collected by the above-mentioned drone, this embodiment performs image splicing based on the coordinate conversion module. After careful processing and reasonable combination of these local pictures, they are successfully spliced into a complete, high-resolution panoramic image as shown in Figure 20.
如图20所示,图像拼接后可能存在空缺区域,这种现象通常由透视变换导致的变形、图像色彩和亮度的变化、图像边缘处理等因素引起。如果缺失区域位于实景地图边缘,如上图黑色区域所示,这些区域往往信息较少,可以直接进行裁剪对齐处理,裁剪后结果如图21所示。As shown in Figure 20, there may be missing areas after the images are stitched. This phenomenon is usually caused by factors such as deformation caused by perspective transformation, changes in image color and brightness, and image edge processing. If the missing area is located at the edge of the real map, as shown in the black area in the figure above, these areas often have less information and can be directly cropped and aligned. The cropped result is shown in Figure 21.
如果缺失区域在实景地图区域内,如图21中的白色区域,这些信息比较重要,因此本实施例采用图像补全方法进行空缺预测和填补。对于图像补全工作,本实施例在训练过程中基于Place数据集的模型进行预训练,接着本发明针对林区、火灾等特定场景,进行了进一步的训练,使得Transformer模型能生成更好的视觉先验。下面将以上述地图的某灾情区域为例,如图22所示,展示缺失区域图像补全的效果。If the missing area is within the real-life map area, such as the white area in Figure 21, this information is relatively important, so this embodiment uses an image completion method to predict and fill in the gaps. For image completion, this embodiment performs pre-training based on a model of the Place data set during the training process, and then the present invention performs further training for specific scenes such as forest areas and fires, so that the Transformer model can generate better visual priors. Below, we will take a disaster area in the above map as an example, as shown in Figure 22, to demonstrate the effect of image completion in the missing area.
在每组结果中,首行左图白洞部分为缺失区域,右图展示了基于Transformer的视觉先验低分辨率修复的多元生成结果。可以看出,通过低分辨率先验修复,修复结果能够基本还原原图轮廓,并且能够生成多样化的修复效果。每组第二行的左图展示了原图,右图展示了基于多元视觉先验的CNN的高分辨率修复结果。如图所示,在输入掩码图片分辨率指引下,模型能够还原出较好的结果。此外,随着应急救援实景数据的不断积累,模型能够进一步提升还原质量。在实景态势图的构建中,支持根据指挥人员的主观评判以及灾点检测置信度、图像生成质量等客观指标的评判,从而确定选择最优的还原效果。In each set of results, the white hole part of the first row on the left is the missing area, and the right picture shows the multivariate generation result of low-resolution restoration based on Transformer's visual prior. It can be seen that through low-resolution prior restoration, the restoration result can basically restore the outline of the original image and can generate a variety of restoration effects. The left picture of the second row of each group shows the original image, and the right picture shows the high-resolution restoration result of CNN based on multivariate visual prior. As shown in the figure, under the guidance of the input mask image resolution, the model can restore better results. In addition, with the continuous accumulation of real-life emergency rescue data, the model can further improve the restoration quality. In the construction of the real-life situation map, it supports the subjective judgment of the commander and the judgment of objective indicators such as disaster point detection confidence and image generation quality, so as to determine the selection of the optimal restoration effect.
在应急救援过程中,尽管局部采样可以节省算力、带宽等资源。但是,如果缺失实景照片的面积过大,实景地图的还原质量可能无法达到要求。因此,本实施例进行了采样密度与带宽的分析,分析缺失区域对图像生成质量的影响。本实施例评估了距离得分(Frechet Inception Distance score,FID)、结构相似性(Structural Similarity,SSIM)、归一化均方根误差(Normalized Root Mean Square Error,NRMSE)和峰值信噪比(Peak Signal to Noise Ratio,PSNR)等图像评估指标的变化情况,以指导无人机采取更合理的采样策略。In the process of emergency rescue, although local sampling can save computing power, bandwidth and other resources. However, if the area of the missing real-life photos is too large, the restoration quality of the real-life map may not meet the requirements. Therefore, this embodiment analyzes the sampling density and bandwidth to analyze the impact of the missing area on the image generation quality. This embodiment evaluates the changes in image evaluation indicators such as the Frechet Inception Distance score (FID), structural similarity (SSIM), normalized root mean square error (NRMSE), and peak signal-to-noise ratio (PSNR) to guide the drone to adopt a more reasonable sampling strategy.
如图23所示,根据图像的指标情况,可以发现实景区域缺失面积对图像生成质量有着显著的影响。因此,在实际救援过程中,可以采用部分地图原图区域与生成区域进行指标测试的方法,以更加全面、准确地评估图像生成质量。除了图像指标情况外,还有一些其他因素也需要考虑,比如算力和带宽等资源的限制,以及受灾面积的大小和复杂程度等。以带宽为例,假设10M带宽在T时间内,能够采样全局数据,在可用带宽不断减少时,高分辨率图片需要进行降采样回传,此时,生成地图质量并不会随着采集区域的增大而出现显著提升的趋势。在综合考虑这些因素的基础上,可以确定适当的采样密度,从而制定出更加合理、高效的无人机灾情采样策略。这种策略可以在最大程度上保证图像生成的质量,同时也能够节约资源和提高救援效率,有助于在应急救援中取得更好的效果。As shown in Figure 23, according to the image indicators, it can be found that the missing area of the real scene area has a significant impact on the image generation quality. Therefore, in the actual rescue process, the method of using some of the original map areas and the generated area for indicator testing can be used to more comprehensively and accurately evaluate the image generation quality. In addition to the image indicators, there are some other factors that need to be considered, such as the limitations of resources such as computing power and bandwidth, as well as the size and complexity of the disaster area. Taking bandwidth as an example, assuming that 10M bandwidth can sample global data within T time, when the available bandwidth continues to decrease, high-resolution images need to be downsampled and transmitted back. At this time, the quality of the generated map will not show a significant improvement trend as the acquisition area increases. Based on comprehensive consideration of these factors, the appropriate sampling density can be determined, thereby formulating a more reasonable and efficient drone disaster sampling strategy. This strategy can guarantee the quality of image generation to the greatest extent, while also saving resources and improving rescue efficiency, which helps to achieve better results in emergency rescue.
如图24所示,将根据生成的实景地图进行基于Yolov5的灾情检测,为了提高灾点检测的精度,本实施例使用了Coco数据集进行预训练,并使用火焰、烟雾等灾情数据集进行模型微调。从而得到在灾情检测方面表现较好的模型。在实景地图中,本研究点将此模型应用于灾点检测,从而确定实际灾点于实景地图中位置。此外,通过坐标转换模块,本实施例还可以获取灾点的真实地理坐标。实景地图中灾点检测效果标定如图24所示。As shown in Figure 24, the Yolov5-based disaster detection will be performed based on the generated real-life map. In order to improve the accuracy of disaster point detection, this embodiment uses the Coco dataset for pre-training, and uses disaster datasets such as flames and smoke to fine-tune the model. Thus, a model with good performance in disaster detection is obtained. In the real-life map, this research point applies this model to disaster point detection to determine the actual location of the disaster point in the real-life map. In addition, through the coordinate conversion module, this embodiment can also obtain the real geographic coordinates of the disaster point. The calibration of the disaster point detection effect in the real-life map is shown in Figure 24.
下面将根据灾点检测情况、服务对象、处置阶段、区域人数等指标对不同区域的通信、定位需求等级进行划分,本实施例采用混淆矩阵评估基于历史救援分级数据的数据集所训练的模型性能。混淆矩阵是一种有效的评估模型性能的工具,可以根据真实标签和预测标签的情况,计算出分类精确率、准确率、召回率、F1值等性能指标,以评估模型的分级效果。由于通信定位需求分级问题为多分类问题,因此本实施例采用Macro-average方法进行指标计算,如下公式(25)-(28)所示。The following will divide the communication and positioning demand levels of different regions according to indicators such as disaster point detection, service objects, disposal stage, and regional population. This embodiment uses a confusion matrix to evaluate the performance of the model trained based on a data set of historical rescue classification data. The confusion matrix is an effective tool for evaluating model performance. It can calculate performance indicators such as classification accuracy, precision, recall rate, F1 value, etc. based on the real labels and predicted labels to evaluate the classification effect of the model. Since the communication and positioning demand classification problem is a multi-classification problem, this embodiment uses the Macro-average method to calculate the indicators, as shown in the following formulas (25)-(28).
其中,Accuracymacro表示精确率,Precisionmacro表示准确率,Recallmacro表示召回率,Fmacro表示F1值,TP(True Positive)表示真正例,即预测为等级N且实际值为等级N的样本数;FP(False Positive)表示假正例,即预测为等级N但实际值为非等级N的样本数;TN(True Negative)表示真反例,即预测为非等级N且实际值为非等级N的样本数;FN(FalseNegative)表示假反例,即预测为非等级N但实际值为等级N的样本数。Among them, Accuracy macro represents precision, Precision macro represents accuracy, Recall macro represents recall, F macro represents F1 value, TP (True Positive) represents true positive examples, that is, the number of samples predicted to be level N and whose actual value is level N; FP (False Positive) represents false positive examples, that is, the number of samples predicted to be level N but whose actual value is non-level N; TN (True Negative) represents true negative examples, that is, the number of samples predicted to be non-level N and whose actual value is non-level N; FN (False Negative) represents false negative examples, that is, the number of samples predicted to be non-level N but whose actual value is level N.
图25a和图25b分别展示了根据通信和定位需求分级数据集训练出的模型的预测混淆矩阵。根据混淆矩阵及其评估指标分析,可有效地评估模型性能,如表2所示。Figure 25a and Figure 25b show the prediction confusion matrices of the models trained based on the communication and positioning requirements classification datasets, respectively. Based on the confusion matrix and its evaluation index analysis, the model performance can be effectively evaluated, as shown in Table 2.
表2需求分级模型性能指标表Table 2 Performance index table of demand grading model
表2展示了在通信和定位分级数据集下训练的一维卷积模型的分级性能。在通信和定位需求分级推理方面,模型的各项指标均达到了95%以上,表现出优异的性能。此外,分级模型还支持根据实际救援中专家的指示进行微调,能够高效地协助应急救援工作。根据不同区域的受困群众人数、服务对象等指标,灾情感知态势图构建与通信定位需求分级方法结果示意图可构建如图26所示。Table 2 shows the classification performance of the one-dimensional convolutional model trained on the communication and positioning classification dataset. In terms of communication and positioning demand classification reasoning, all indicators of the model reached more than 95%, showing excellent performance. In addition, the classification model also supports fine-tuning according to the instructions of experts in actual rescue, which can efficiently assist emergency rescue work. According to indicators such as the number of trapped people and service objects in different areas, the schematic diagram of the results of the disaster awareness situation map construction and communication positioning demand classification method can be constructed as shown in Figure 26.
图26中,受灾区域通信定位需求分级等级差异化,需求分级等级由L1至L5递增,可以作为后续章节中资源适配研究的重要参考依据。在资源适配研究中,需求分级等级用于设定带宽分配、通信速率阈值以及定位精度阈值等参数。因此,这些等级的划分对于后续的资源适配研究具有重要的指导作用。In Figure 26, the communication and positioning requirements in the disaster-stricken areas are differentiated, and the requirements are graded from L1 to L5, which can serve as an important reference for the resource adaptation research in the subsequent chapters. In the resource adaptation research, the requirements are used to set parameters such as bandwidth allocation, communication rate threshold, and positioning accuracy threshold. Therefore, the division of these levels has an important guiding role in the subsequent resource adaptation research.
下面对本发明提供的应急场景下的实景地图构建装置进行描述,下文描述的应急场景下的实景地图构建装置与上文描述的应急场景下的实景地图构建方法可相互对应参照。The following is a description of a real scene map construction device for an emergency scenario provided by the present invention. The real scene map construction device for an emergency scenario described below and the real scene map construction method for an emergency scenario described above can be referenced to each other.
请参照图27,图27是本发明实施例提供的应急场景下的实景地图构建装置的结构示意图。如图27所示,该装置可以包括:Please refer to Figure 27, which is a schematic diagram of the structure of a real scene map construction device in an emergency scenario provided by an embodiment of the present invention. As shown in Figure 27, the device may include:
第一获取模块10,用于获取多无人机拍摄的受灾区域的实景图像;A first acquisition module 10 is used to acquire real-life images of the disaster-stricken area taken by multiple drones;
第一构建模块20,用于基于各实景图像,构建实时灾情感知态势图;The first construction module 20 is used to construct a real-time disaster awareness situation map based on each real scene image;
第二获取模块30,用于获取通信定位需求分级模型,通信定位需求分级模型是基于应急场景下的历史分级数据训练得到的;A second acquisition module 30 is used to acquire a communication positioning demand classification model, where the communication positioning demand classification model is trained based on historical classification data in emergency scenarios;
第二构建模块40,用于基于实时灾情感知态势图和通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图。The second construction module 40 is used to construct a real-life map based on communication positioning demand classification calibration and disaster point calibration based on the real-time disaster awareness situation map and the communication positioning demand classification model.
在一种示例实施例中,第一构建模块20可以包括:In an exemplary embodiment, the first building block 20 may include:
转换子模块,用于将实景图像从像素坐标系转换到世界坐标系,得到目标图像;A conversion submodule is used to convert the real scene image from the pixel coordinate system to the world coordinate system to obtain the target image;
拼接子模块,用于基于各目标图像的世界坐标,将各目标图像进行拼接处理,得到拼接图像;A stitching submodule, used for stitching the target images based on the world coordinates of the target images to obtain a stitched image;
补全子模块,用于将拼接图像中的空缺区域进行补全处理,得到实时灾情感知态势图。The completion submodule is used to complete the missing areas in the spliced image to obtain a real-time disaster awareness situation map.
在一种示例实施例中,拼接子模块具体用于:In an exemplary embodiment, the splicing submodule is specifically used for:
针对任意两张目标图像,根据两张目标图像的世界坐标判断两张目标图像是否存在关联区域;For any two target images, determine whether there is a related area between the two target images according to the world coordinates of the two target images;
在两张目标图像存在关联区域的情况下,采用尺度不变的特征点检测算法对两张目标图像分别进行特征检测,得到特征点集;When there are related areas between the two target images, a scale-invariant feature point detection algorithm is used to perform feature detection on the two target images respectively to obtain a feature point set;
计算特征点集中每个特征点的描述符,得到描述符集;Calculate the descriptor of each feature point in the feature point set to obtain a descriptor set;
基于两张目标图像对应的描述符集,对两张目标图像对应的特征点集进行相似特征点的匹配处理,得到多个匹配点;Based on the descriptor sets corresponding to the two target images, matching processing is performed on similar feature points of the feature point sets corresponding to the two target images to obtain multiple matching points;
将多个匹配点中的异常匹配点进行过滤,以对多个匹配点进行矫正;Filtering abnormal matching points among the multiple matching points to correct the multiple matching points;
将两张目标图像分别进行透视变换;Perform perspective transformation on the two target images respectively;
基于矫正后的多个匹配点,对透视变换后的两张目标图像进行图像融合,得到拼接图像;其中,在图像融合的过程中,采用羽化技术进行平滑的图像转换,以及通过加权平均颜色值对重叠的像素进行融合。Based on the corrected multiple matching points, the two target images after perspective transformation are fused to obtain a stitched image. In the process of image fusion, feathering technology is used to perform smooth image conversion, and the overlapping pixels are fused by weighted average color value.
在一种示例实施例中,补全子模块包括:In an exemplary embodiment, the completion submodule includes:
重建单元,用于对拼接图像中的空缺区域进行低维视觉先验信息的重建;A reconstruction unit, used to reconstruct low-dimensional visual prior information of the vacant area in the spliced image;
补充单元,用于基于低维视觉先验信息,向重建后的拼接图像中补充纹理细节信息,得到实时灾情感知态势图。The supplement unit is used to supplement texture detail information into the reconstructed spliced image based on low-dimensional visual prior information to obtain a real-time disaster awareness situation map.
在一种示例实施例中,重建单元具体用于:In an exemplary embodiment, the reconstruction unit is specifically configured to:
将拼接图像降采样为低维图像;Downsample the spliced image to a low-dimensional image;
将低维图像确定为拼接图像的视觉先验图像;Determine the low-dimensional image as a visual prior image for the stitched image;
采用聚类算法在预设图像数据集的颜色空间中生成颜色字典;A clustering algorithm is used to generate a color dictionary in the color space of a preset image data set;
针对视觉先验图像中的每个像素,在颜色字典中查找与像素最接近的元素索引,得到像素的离散表示;For each pixel in the visual prior image, find the element index closest to the pixel in the color dictionary to obtain a discrete representation of the pixel;
将视觉先验图像中各像素的离散表示输入Transformer模型中进行迭代计算,将拼接图像中的空缺区域中各像素的离散表示中的元素替换为吉布斯采样标记,得到离散序列;The discrete representation of each pixel in the visual prior image is input into the Transformer model for iterative calculation, and the elements in the discrete representation of each pixel in the vacant area of the spliced image are replaced with Gibbs sampling marks to obtain a discrete sequence;
针对每个离散序列,通过查询颜色字典来重建低维视觉先验信息。For each discrete sequence, the low-dimensional visual prior information is reconstructed by querying the color dictionary.
在一种示例实施例中,补充单元具体用于:In an exemplary embodiment, the supplement unit is specifically configured to:
将重建后的视觉先验图像的图像矩阵和图像矩阵的双线性插值结果输入到引导上采样网络中进行处理,得到预测值;The image matrix of the reconstructed visual prior image and the bilinear interpolation result of the image matrix are input into the guided upsampling network for processing to obtain the predicted value;
基于预测值和真实值之间的损失函数来调整引导上采样网络的模型参数;损失函数为L1损失函数和对抗性损失函数的加权和;对抗性损失函数与判别器对预测值的判别值相关;The model parameters of the guided upsampling network are adjusted based on the loss function between the predicted value and the true value; the loss function is the weighted sum of the L1 loss function and the adversarial loss function; the adversarial loss function is related to the discriminant value of the discriminator on the predicted value;
联合训练引导上采样网络和判别器,得到多元化还原结果;Jointly train the guided upsampling network and the discriminator to obtain diversified restoration results;
从多元化还原结果中确定鲁棒性最高的结果;Determine the most robust result from the diversified reduction results;
基于鲁棒性最高的结果向重建后的拼接图像中补充纹理细节信息,得到实时灾情感知态势图。Based on the most robust result, texture detail information is added to the reconstructed spliced image to obtain a real-time disaster awareness situation map.
在一种示例实施例中,第二获取模块30具体用于:In an exemplary embodiment, the second acquisition module 30 is specifically configured to:
获取应急场景下的历史分级数据;历史分级数据包括分级指标和针对每个分级指标的专家评语标签;Obtain historical classification data under emergency scenarios; historical classification data includes classification indicators and expert comment labels for each classification indicator;
将历史分级数据划分为训练集和测试集;Divide the historical classification data into training set and test set;
将训练集输入通信定位需求分级模型中进行训练;The training set is input into the communication positioning requirement classification model for training;
将测试集输入训练好的通信定位需求分级模型中进行测试;Input the test set into the trained communication positioning requirement classification model for testing;
在通信定位需求分级模型的测试结果未通过的情况下,对通信定位需求分级模型的模型参数进行微调,直至通信定位需求分级模型的测试结果通过。When the test result of the communication positioning demand grading model fails, the model parameters of the communication positioning demand grading model are fine-tuned until the test result of the communication positioning demand grading model passes.
在一种示例实施例中,第二构建模块40具体用于:In an exemplary embodiment, the second building module 40 is specifically configured to:
确定通信定位需求分级模型的输入指标;输入指标包括服务对象、应急场景、处置阶段和事件客观因素的细化类型以及通信定位需求的各个等级;Determine the input indicators of the communication positioning demand classification model; the input indicators include service objects, emergency scenarios, handling stages, detailed types of objective factors of events, and various levels of communication positioning demand;
将输入指标输入通信定位需求分级模型中进行处理,输出通信定位需求分级标定结果;Inputting the input indicators into the communication positioning demand grading model for processing, and outputting the communication positioning demand grading calibration results;
基于实时灾情感知态势图确定灾点标定结果;Determine the disaster point calibration results based on the real-time disaster awareness situation map;
将通信定位需求分级标定结果和灾点标定结果添加至实时灾情感知态势图中,得到实景地图。The communication positioning demand classification calibration results and disaster point calibration results are added to the real-time disaster awareness situation map to obtain a real-scene map.
图28示例了一种电子设备的实体结构示意图,如图28所示,该电子设备可以包括:处理器(processor)810、通信接口(Communications Interface)820、存储器(memory)830和通信总线840,其中,处理器810,通信接口820,存储器830通过通信总线840完成相互间的通信。处理器810可以调用存储器830中的逻辑指令,以执行应急场景下的实景地图构建方法,该方法包括:获取多无人机拍摄的多个受灾区域的实景图像;基于各实景图像,构建实时灾情感知态势图;获取通信定位需求分级模型,通信定位需求分级模型是基于应急场景下的历史分级数据训练得到的;基于实时灾情感知态势图和通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图。FIG28 illustrates a schematic diagram of the physical structure of an electronic device. As shown in FIG28 , the electronic device may include: a processor 810, a communications interface 820, a memory 830, and a communication bus 840, wherein the processor 810, the communications interface 820, and the memory 830 communicate with each other through the communication bus 840. The processor 810 may call the logic instructions in the memory 830 to execute a method for constructing a real-life map in an emergency scenario, the method comprising: obtaining real-life images of multiple disaster-stricken areas taken by multiple drones; constructing a real-time disaster awareness situation map based on each real-life image; obtaining a communication positioning demand grading model, the communication positioning demand grading model is obtained by training based on historical grading data in an emergency scenario; and constructing a real-life map based on communication positioning demand grading calibration and disaster point calibration based on the real-time disaster awareness situation map and the communication positioning demand grading model.
此外,上述的存储器830中的逻辑指令可以通过软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本发明各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质。In addition, the logic instructions in the above-mentioned memory 830 can be implemented in the form of a software functional unit and can be stored in a computer-readable storage medium when it is sold or used as an independent product. Based on such an understanding, the technical solution of the present invention, in essence, or the part that contributes to the prior art or the part of the technical solution, can be embodied in the form of a software product, and the computer software product is stored in a storage medium, including a number of instructions for a computer device (which can be a personal computer, a server, or a network device, etc.) to perform all or part of the steps of the method described in each embodiment of the present invention. The aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), disk or optical disk and other media that can store program codes.
另一方面,本发明还提供一种计算机程序产品,所述计算机程序产品包括存储在非暂态计算机可读存储介质上的计算机程序,所述计算机程序包括程序指令,当程序指令被计算机执行时,计算机能够执行上述各方法所提供的应急场景下的实景地图构建方法,该方法包括:获取多无人机拍摄的多个受灾区域的实景图像;基于各实景图像,构建实时灾情感知态势图;获取通信定位需求分级模型,通信定位需求分级模型是基于应急场景下的历史分级数据训练得到的;基于实时灾情感知态势图和通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图。On the other hand, the present invention also provides a computer program product, which includes a computer program stored on a non-transitory computer-readable storage medium, and the computer program includes program instructions. When the program instructions are executed by a computer, the computer can execute the real-life map construction method in emergency scenarios provided by the above-mentioned methods, and the method includes: obtaining real-life images of multiple disaster-stricken areas taken by multiple drones; based on each real-life image, constructing a real-time disaster awareness situation map; obtaining a communication positioning demand grading model, which is obtained by training based on historical grading data in emergency scenarios; based on the real-time disaster awareness situation map and the communication positioning demand grading model, constructing a real-life map based on communication positioning demand grading calibration and disaster point calibration.
又一方面,本发明还提供一种非暂态计算机可读存储介质,其上存储有计算机程序,该计算机程序被处理器执行时实现以执行上述各提供的应急场景下的实景地图构建方法,该方法包括:获取多无人机拍摄的多个受灾区域的实景图像;基于各实景图像,构建实时灾情感知态势图;获取通信定位需求分级模型,通信定位需求分级模型是基于应急场景下的历史分级数据训练得到的;基于实时灾情感知态势图和通信定位需求分级模型,构建基于通信定位需求分级标定与灾点标定的实景地图。On the other hand, the present invention also provides a non-transitory computer-readable storage medium having a computer program stored thereon, which, when executed by a processor, is implemented to execute the above-mentioned methods for constructing real-life maps in emergency scenarios, the method comprising: obtaining real-life images of multiple disaster-stricken areas taken by multiple drones; constructing a real-time disaster awareness situation map based on each real-life image; obtaining a communication positioning demand grading model, the communication positioning demand grading model being trained based on historical grading data in emergency scenarios; and constructing a real-life map based on communication positioning demand grading calibration and disaster point calibration based on the real-time disaster awareness situation map and the communication positioning demand grading model.
以上所描述的装置实施例仅仅是示意性的,其中所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。本领域普通技术人员在不付出创造性的劳动的情况下,即可以理解并实施。The device embodiments described above are merely illustrative, wherein the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the scheme of this embodiment. Those of ordinary skill in the art may understand and implement it without creative work.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到各实施方式可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件。基于这样的理解,上述技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品可以存储在计算机可读存储介质中,如ROM/RAM、磁碟、光盘等,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行各个实施例或者实施例的某些部分所述的方法。Through the description of the above implementation methods, those skilled in the art can clearly understand that each implementation method can be implemented by means of software plus a necessary general hardware platform, and of course, it can also be implemented by hardware. Based on this understanding, the above technical solution is essentially or the part that contributes to the prior art can be embodied in the form of a software product, and the computer software product can be stored in a computer-readable storage medium, such as ROM/RAM, a disk, an optical disk, etc., including a number of instructions for a computer device (which can be a personal computer, a server, or a network device, etc.) to execute the methods described in each embodiment or some parts of the embodiments.
最后应说明的是:以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。Finally, it should be noted that the above embodiments are only used to illustrate the technical solutions of the present invention, rather than to limit it. Although the present invention has been described in detail with reference to the aforementioned embodiments, those skilled in the art should understand that they can still modify the technical solutions described in the aforementioned embodiments, or make equivalent replacements for some of the technical features therein. However, these modifications or replacements do not deviate the essence of the corresponding technical solutions from the spirit and scope of the technical solutions of the embodiments of the present invention.
Claims (11)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310719078.8A CN116993933A (en) | 2023-06-16 | 2023-06-16 | Live-action map construction method, device and equipment under emergency scene and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310719078.8A CN116993933A (en) | 2023-06-16 | 2023-06-16 | Live-action map construction method, device and equipment under emergency scene and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116993933A true CN116993933A (en) | 2023-11-03 |
Family
ID=88532888
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310719078.8A Pending CN116993933A (en) | 2023-06-16 | 2023-06-16 | Live-action map construction method, device and equipment under emergency scene and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116993933A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117555979A (en) * | 2024-01-11 | 2024-02-13 | 人民中科(北京)智能技术有限公司 | Efficient bottom-up map position missing identification method |
-
2023
- 2023-06-16 CN CN202310719078.8A patent/CN116993933A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117555979A (en) * | 2024-01-11 | 2024-02-13 | 人民中科(北京)智能技术有限公司 | Efficient bottom-up map position missing identification method |
CN117555979B (en) * | 2024-01-11 | 2024-04-19 | 人民中科(北京)智能技术有限公司 | Efficient bottom-up map position missing identification method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106157307B (en) | A kind of monocular image depth estimation method based on multiple dimensioned CNN and continuous CRF | |
CN110222787B (en) | Multi-scale target detection method, device, computer equipment and storage medium | |
CN113936256A (en) | Image target detection method, device, equipment and storage medium | |
CN113870335A (en) | Monocular depth estimation method based on multi-scale feature fusion | |
CN112597995B (en) | License plate detection model training method, device, equipment and medium | |
CN110942484B (en) | Camera self-motion estimation method based on occlusion perception and feature pyramid matching | |
CN113011308A (en) | Pedestrian detection method introducing attention mechanism | |
CN114782798A (en) | Underwater target detection method based on attention fusion | |
CN113724379A (en) | Three-dimensional reconstruction method, device, equipment and storage medium | |
CN115170746A (en) | Multi-view three-dimensional reconstruction method, system and equipment based on deep learning | |
CN114169425A (en) | Training target tracking model and target tracking method and device | |
US7486815B2 (en) | Method and apparatus for scene learning and three-dimensional tracking using stereo video cameras | |
Huang et al. | ES-Net: An efficient stereo matching network | |
CN116993933A (en) | Live-action map construction method, device and equipment under emergency scene and storage medium | |
WO2023222643A1 (en) | Method for image segmentation matching | |
CN113191301B (en) | Video dense crowd counting method and system integrating time sequence and spatial information | |
CN110335228B (en) | Method, device and system for determining image parallax | |
CN111986233B (en) | Remote sensing video tracking method for extremely small targets in large scenes based on feature self-learning | |
CN116912488B (en) | Three-dimensional panoramic segmentation method and device based on multi-camera | |
CN116258756B (en) | A self-supervised monocular depth estimation method and system | |
CN115909255B (en) | Image generation and image segmentation methods, devices, equipment, vehicle-mounted terminal and medium | |
CN114842287B (en) | Monocular three-dimensional target detection model training method and device of depth-guided deformer | |
CN111126177A (en) | People counting method and device | |
CN114091519A (en) | An occluded pedestrian re-identification method based on multi-granularity occlusion perception | |
CN116883770A (en) | Training method and device of depth estimation model, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |