[go: up one dir, main page]

CN112541912A - Method and device for rapidly detecting saliency target in mine sudden disaster scene - Google Patents

Method and device for rapidly detecting saliency target in mine sudden disaster scene Download PDF

Info

Publication number
CN112541912A
CN112541912A CN202011541385.4A CN202011541385A CN112541912A CN 112541912 A CN112541912 A CN 112541912A CN 202011541385 A CN202011541385 A CN 202011541385A CN 112541912 A CN112541912 A CN 112541912A
Authority
CN
China
Prior art keywords
node
input image
network
target
segmentation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011541385.4A
Other languages
Chinese (zh)
Other versions
CN112541912B (en
Inventor
程德强
刘瑞航
李佳函
寇旗旗
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Huatu Mining Technology Co ltd
China University of Mining and Technology CUMT
Original Assignee
Jiangsu Huatu Mining Technology Co ltd
China University of Mining and Technology CUMT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu Huatu Mining Technology Co ltd, China University of Mining and Technology CUMT filed Critical Jiangsu Huatu Mining Technology Co ltd
Priority to CN202011541385.4A priority Critical patent/CN112541912B/en
Publication of CN112541912A publication Critical patent/CN112541912A/en
Application granted granted Critical
Publication of CN112541912B publication Critical patent/CN112541912B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details
    • G06T2207/20192Edge enhancement; Edge preservation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Quality & Reliability (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

本发明涉及一种矿井突发灾害场景中显著性目标的快速检测方法及装置,属于计算机视觉技术领域,解决了现有的显著性目标检测方法的检测精度较差和速度较慢的问题。方法包括:获取关于矿井突发灾害场景的输入图像;构造Unet网络,将输入图像输入Unet网络,得到输入图像对应的分割映射图像;构造全卷积FCN网络,将输入图像输入全卷积FCN网络,得到输入图像对应的显著性映射图像;将输入图像对应的分割映射图像和输入图像对应的显著性映射图像融合,得到融合映射图像,并基于融合映射图像得到局部图;基于局部图获得最终的显著性目标。实现了矿井突发灾害场景中显著性目标的快速检测,提高了显著性目标检测的精度和速度。

Figure 202011541385

The invention relates to a rapid detection method and device for a salient target in a mine sudden disaster scene, belonging to the technical field of computer vision, and solves the problems of poor detection accuracy and slow speed of the existing salient target detection method. The method includes: acquiring an input image about a mine sudden disaster scene; constructing a Unet network, inputting the input image into the Unet network, and obtaining a segmentation map image corresponding to the input image; constructing a fully convolutional FCN network, and inputting the input image into the fully convolutional FCN network , obtain the saliency map image corresponding to the input image; fuse the segmentation map image corresponding to the input image and the saliency map image corresponding to the input image to obtain the fusion map image, and obtain the local map based on the fusion map image; obtain the final image based on the local map salient goals. The rapid detection of salient targets in the mine sudden disaster scene is realized, and the accuracy and speed of salient target detection are improved.

Figure 202011541385

Description

矿井突发灾害场景中显著性目标的快速检测方法及装置Rapid detection method and device for salient objects in mine sudden disaster scene

技术领域technical field

本发明涉及计算机视觉技术领域,尤其涉及一种矿井突发灾害场景中显著性目标的快速检测方法及装置。The invention relates to the technical field of computer vision, in particular to a method and device for rapid detection of salient objects in a mine sudden disaster scene.

背景技术Background technique

煤矿井下视频图像在特殊的工矿环境下,采集到的图像照度低、变化大,背景噪声大且噪声分布不均匀,此外,部分煤矿井下视频既包含固定摄像机拍摄的视频图像也包括车载摄像机摄录的视频图像,煤矿井下全天候人工照明环境以及粉尘和潮湿等因素影响使得图像采集精度较差,都极大地影响了现场情景的呈现精度和远程主动预警联动效果,限制了矿井视频高精度采集和目标识别应用,Under the special industrial and mining environment, the video images collected in coal mines have low illuminance, large changes, large background noise and uneven noise distribution. In addition, some underground videos in coal mines include both video images captured by fixed cameras and vehicle-mounted cameras. The influence of the all-weather artificial lighting environment in the coal mine and factors such as dust and humidity make the image acquisition accuracy poor, which greatly affects the presentation accuracy of the scene scene and the remote active early warning linkage effect, and limits the high-precision acquisition and target of mine video. identify applications,

传统的煤矿视频监控系统只能对监控场景进行记录,这样就需要工作人员认真、连续的观察,不仅对其注意力要求高度集中,且在事故发生时不能快速的对事故进行报警及联动处理。煤矿井下视频的特殊性给人员自动检测带来很大困难,也使得目前的方法在井下突发灾害场景中显著性目标的快速检测及响应联动中受到限制。The traditional coal mine video monitoring system can only record the monitoring scene, which requires the staff to observe carefully and continuously, not only requires a high degree of attention, but also cannot quickly alarm and deal with the accident when the accident occurs. The particularity of underground video in coal mines brings great difficulties to the automatic detection of personnel, and also restricts the current methods in the rapid detection and response linkage of salient targets in sudden underground disaster scenarios.

目前的显著性目标检测方法大多是基于深度学习神经网络获取深度显著性特征的检测方法,其存在的问题包括:第一,目标检测大多是关注目标中心的位置,对于边缘的处理和关注存在一定缺陷;第二,对于有多个显著性目标的图像,各个目标之间的关联性没有用到显著性的推理当中,使得现有的显著性目标检测方法检测精度较差和速度较慢。Most of the current saliency target detection methods are detection methods based on deep learning neural networks to obtain deep salient features. The existing problems include: First, target detection mostly focuses on the position of the target center, and there is a certain degree of edge processing and attention. Defects; second, for images with multiple saliency targets, the correlation between each target is not used in saliency reasoning, which makes the existing saliency target detection methods have poor detection accuracy and slow speed.

发明内容SUMMARY OF THE INVENTION

鉴于上述的分析,本发明实施例旨在提供一种矿井突发灾害场景中显著性目标的快速检测方法及装置,用以解决现有的显著性目标检测方法的检测精度较差和速度较慢的问题。In view of the above analysis, the embodiments of the present invention aim to provide a rapid detection method and device for a salient target in a mine sudden disaster scene, so as to solve the problem of poor detection accuracy and slow speed of the existing salient target detection method The problem.

一方面,本发明实施例提供了一种矿井突发灾害场景中显著性目标的快速检测方法,包括下述步骤:On the one hand, an embodiment of the present invention provides a rapid detection method for a salient target in a mine sudden disaster scene, including the following steps:

获取关于矿井突发灾害场景的输入图像;Obtain input images of mine sudden disaster scenarios;

构造Unet网络,将所述输入图像输入Unet网络,得到输入图像对应的分割映射图像,所述分割映射图像包含多个分割目标;Constructing a Unet network, inputting the input image into the Unet network, to obtain a segmentation map image corresponding to the input image, and the segmentation map image includes a plurality of segmentation targets;

构造全卷积FCN网络,将所述输入图像输入全卷积FCN网络,得到输入图像对应的显著性映射图像;Constructing a fully convolutional FCN network, and inputting the input image into a fully convolutional FCN network to obtain a saliency map image corresponding to the input image;

将所述输入图像对应的分割映射图像和输入图像对应的显著性映射图像融合,得到融合映射图像,并基于所述融合映射图像得到局部图;fusing the segmentation map image corresponding to the input image and the saliency map image corresponding to the input image to obtain a fusion map image, and obtain a local map based on the fusion map image;

基于所述局部图获得最终的显著性目标。The final saliency target is obtained based on the local map.

进一步,所述Unet网络包括编码器和解码器;Further, the Unet network includes an encoder and a decoder;

所述编码器包括依次连接的两个第一网络和四个第二网络,其中,所述第一网络包括依次连接的卷积层、批标准化层和激活层;所述第二网络包括依次连接的最大池化层和两个第一网络;The encoder includes two first networks and four second networks connected in sequence, wherein the first network includes a convolution layer, a batch normalization layer and an activation layer connected in sequence; the second network includes a sequential connection The max pooling layer and the two first networks;

所述解码器包括依次连接的四个第三网络和一个卷积层,其中,所述第三网络包括依次连接的转置卷积层和两个第一网络。The decoder includes four third networks and one convolutional layer connected in sequence, wherein the third network includes a transposed convolutional layer and two first networks connected in sequence.

进一步,所述全卷积FCN网络包括三部分;其中,Further, the fully convolutional FCN network includes three parts; wherein,

第一部分包含五个卷积层,每个卷积层后均连接一个池化层;The first part contains five convolutional layers, each of which is followed by a pooling layer;

第二部分包含依次连接的三个卷积层;The second part contains three convolutional layers connected in sequence;

第三部分用于对第二部分的输出进行上采样操作,所述上采样操作基于池化层和反卷积层实现。The third part is used to perform an upsampling operation on the output of the second part, and the upsampling operation is implemented based on a pooling layer and a deconvolution layer.

进一步,将融合映射图像中的每一分割目标作为一个节点,并采用KD最近邻算法计算得到任意两个节点之间的距离和相对位置;Further, take each segmentation target in the fusion map image as a node, and use the KD nearest neighbor algorithm to calculate the distance and relative position between any two nodes;

将每一个节点作为中心节点,获得距离中心节点最短和次短的两个节点作为中心节点的一跳节点,并将距离每一个所述一跳节点的最短和次短的两个节点作为中心节点的二跳节点;Take each node as the central node, obtain the two nodes with the shortest and the second shortest distance from the central node as the one-hop node of the central node, and take the two nodes with the shortest and the second shortest distance from each of the one-hop nodes as the central node the two-hop node;

基于所述中心节点的一跳节点和二跳节点得到节点集合,并获得所述节点集合中每一个节点的近邻节点,并将节点集合中每一个节点与其在节点集合中的所述近邻节点连接,得到每一分割目标对应的局部图。A node set is obtained based on the one-hop node and two-hop node of the central node, the neighbor nodes of each node in the node set are obtained, and each node in the node set is connected with the neighbor node in the node set , and obtain the local map corresponding to each segmentation target.

进一步,基于所述局部图获得最终的显著性目标,包括下述步骤:Further, obtaining the final saliency target based on the local map includes the following steps:

基于图卷积推理网络GCN将所有的局部图组合,得到组合图,并获取所述组合图中的每一节点与其对应的一跳节点的边权;Based on the graph convolutional reasoning network GCN, combine all the local graphs to obtain a combined graph, and obtain the edge weight of each node in the combined graph and its corresponding one-hop node;

基于所述组合图中的每一节点与其对应的一跳节点的边权得到最终的显著性目标。The final saliency target is obtained based on the edge weight of each node in the combined graph and its corresponding one-hop node.

进一步,所述图卷积推理网络GCN包括依次连接的四个卷积层和两个全连接层。Further, the graph convolutional reasoning network GCN includes four convolutional layers and two fully connected layers connected in sequence.

进一步,基于所述组合图中的每一节点与其对应的一跳节点的边权得到最终的显著性目标,包括下述步骤:Further, the final saliency target is obtained based on the edge weight of each node in the combined graph and its corresponding one-hop node, including the following steps:

判断每一节点与其对应的一跳节点的边权是否大于权重阈值,若是,合并所述节点与其对应的一跳节点,若否,不合并所述节点与其对应的一跳节点;Determine whether the edge weight of each node and its corresponding one-hop node is greater than the weight threshold, if so, merge the node and its corresponding one-hop node, if not, do not merge the node and its corresponding one-hop node;

遍历所述组合图中的所有节点,得到最终的显著性目标。Traverse all nodes in the combined graph to get the final saliency target.

另一方面,本发明实施例提供了一种矿井突发灾害场景中显著性目标的快速检测装置,包括:On the other hand, an embodiment of the present invention provides a rapid detection device for a salient target in a mine sudden disaster scene, including:

数据采集模块,用于获取关于矿井突发灾害场景的输入图像;The data acquisition module is used to acquire input images about the mine sudden disaster scene;

Unet网络构建模块,用于构造Unet网络,将所述输入图像输入Unet网络,得到输入图像对应的分割映射图像,所述分割映射图像包含多个分割目标;The Unet network building module is used to construct the Unet network, and the input image is input into the Unet network to obtain a segmentation map image corresponding to the input image, and the segmentation map image includes a plurality of segmentation targets;

全卷积FCN网络构建模块,用于构造全卷积FCN网络,将所述输入图像输入全卷积FCN网络,得到输入图像对应的显著性映射图像;A fully convolutional FCN network building module is used to construct a fully convolutional FCN network, and the input image is input into the fully convolutional FCN network to obtain a saliency map image corresponding to the input image;

局部图获得模块,用于将所述输入图像对应的分割映射图像和输入图像对应的显著性映射图像融合,得到融合映射图像,并基于所述融合映射图像得到局部图;a local map obtaining module, configured to fuse the segmentation map image corresponding to the input image and the saliency map image corresponding to the input image to obtain a fusion map image, and obtain a local map based on the fusion map image;

显著性目标获得模块,基于所述局部图获得最终的显著性目标。The saliency target obtaining module obtains the final saliency target based on the local map.

进一步,所述Unet网络包括编码器和解码器;Further, the Unet network includes an encoder and a decoder;

所述编码器包括依次连接的两个第一网络和四个第二网络,其中,所述第一网络包括依次连接的卷积层、批标准化层和激活层;所述第二网络包括依次连接的最大池化层和两个第一网络;The encoder includes two first networks and four second networks connected in sequence, wherein the first network includes a convolution layer, a batch normalization layer and an activation layer connected in sequence; the second network includes a sequential connection The max pooling layer and the two first networks;

所述解码器包括依次连接的四个第三网络和一个卷积层,其中,所述第三网络包括依次连接的转置卷积层和两个第一网络。The decoder includes four third networks and one convolutional layer connected in sequence, wherein the third network includes a transposed convolutional layer and two first networks connected in sequence.

进一步,所述全卷积FCN网络包括三部分;其中,Further, the fully convolutional FCN network includes three parts; wherein,

第一部分包含五个卷积层,每个卷积层后均连接一个池化层;The first part contains five convolutional layers, each of which is followed by a pooling layer;

第二部分包含依次连接的三个卷积层;The second part contains three convolutional layers connected in sequence;

第三部分用于对第二部分的输出进行上采样操作,所述上采样操作基于池化层和反卷积层实现。The third part is used to perform an upsampling operation on the output of the second part, and the upsampling operation is implemented based on a pooling layer and a deconvolution layer.

与现有技术相比,本发明至少可实现如下有益效果之一:Compared with the prior art, the present invention can achieve at least one of the following beneficial effects:

1、一种矿井突发灾害场景中显著性目标的快速检测方法,将Unet网络得到的分割映射图像和全卷积FCN网络得到的显著性映射图像融合,得到融合映射图像,接着基于该融合映射图像得到多张局部图,最后通过图卷积推理网络GCN对多张局部图进行处理,得到最终的显著性目标。该方法简单易行,易于实施,考虑了每一个分割目标的关联性,提高了获得的显著性目标的精度,提高了检测速度,具有较高的实用价值。1. A rapid detection method for salient objects in mine sudden disaster scenes, which fuses the segmentation map image obtained by Unet network and the saliency map image obtained by fully convolutional FCN network to obtain a fusion map image, and then based on the fusion map The image obtains multiple local maps, and finally processes the multiple local maps through the graph convolution inference network GCN to obtain the final saliency target. The method is simple, easy to implement, considers the relevance of each segmentation target, improves the accuracy of the obtained saliency target, improves the detection speed, and has high practical value.

2、通过Unet网络获取输入图像对应的分割映射图像,为后期进行图像融合及显著性目标检测提供了技术支撑和依据。同时,分割映射图像包含多个分割目标,在后期融合图像的基础上对分割目标处理以得到局部图,考虑了各个分割目标之间的联系,有利于提高获得的显著性目标的精度。2. The segmentation map image corresponding to the input image is obtained through the Unet network, which provides technical support and basis for image fusion and salient target detection in the later stage. At the same time, the segmentation map image contains multiple segmentation targets, and the segmentation targets are processed on the basis of the later fusion image to obtain a local map, considering the relationship between the segmentation targets, which is beneficial to improve the accuracy of the obtained saliency targets.

3、将融合映射图像中的每一分割目标作为一个中心节点,以获得每一个中心节点对应的局部图,考虑了每一个分割目标之间的联系,有利于增强边缘的效果,使得最终获得的显著性目标精度更高。3. Take each segmentation target in the fusion map image as a central node to obtain the local graph corresponding to each central node, and consider the connection between each segmentation target, which is conducive to enhancing the effect of the edge, so that the final obtained The saliency target is more accurate.

本发明中,上述各技术方案之间还可以相互组合,以实现更多的优选组合方案。本发明的其他特征和优点将在随后的说明书中阐述,并且,部分优点可从说明书中变得显而易见,或者通过实施本发明而了解。本发明的目的和其他优点可通过说明书以及附图中所特别指出的内容中来实现和获得。In the present invention, the above technical solutions can also be combined with each other to achieve more preferred combination solutions. Additional features and advantages of the invention will be set forth in the description which follows, and some of the advantages may become apparent from the description, or may be learned by practice of the invention. The objectives and other advantages of the invention may be realized and attained by means of particularly pointed out in the description and drawings.

附图说明Description of drawings

附图仅用于示出具体实施例的目的,而并不认为是对本发明的限制,在整个附图中,相同的参考符号表示相同的部件。The drawings are for the purpose of illustrating specific embodiments only and are not to be considered limiting of the invention, and like reference numerals refer to like parts throughout the drawings.

图1为一个实施例中矿井突发灾害场景中显著性目标的快速检测方法框架图;1 is a framework diagram of a method for rapid detection of salient objects in a mine sudden disaster scene in one embodiment;

图2为一个实施例中矿井突发灾害场景中显著性目标的快速检测方法流程图;2 is a flowchart of a method for rapid detection of salient objects in a mine sudden disaster scene in one embodiment;

图3为一个实施例中Unet网络结构图;Fig. 3 is Unet network structure diagram in one embodiment;

图4为一个实施例中全卷积FCN网络结构图;Fig. 4 is a fully convolutional FCN network structure diagram in one embodiment;

图5为一个实施例中矿井突发灾害场景中显著性目标的快速检测装置结构图;5 is a structural diagram of a rapid detection device for salient objects in a mine sudden disaster scene in one embodiment;

图6为一个实施例中执行本申请发明实施例提供的矿井突发灾害场景中显著性目标的快速检测方法的电子设备的硬件结构示意图。FIG. 6 is a schematic diagram of a hardware structure of an electronic device for implementing the method for rapid detection of a salient target in a mine sudden disaster scene provided by an embodiment of the present invention in one embodiment.

附图标记:Reference number:

100-数据采集模块,200-Unet网络构建模块,300-全卷积FCN网络构建模块,400-局部图获得模块,500-显著性目标获得模块,610-处理器,620-存储器,630-输入装置,640-输出装置。100-Data acquisition module, 200-Unet network building module, 300-Fully convolutional FCN network building module, 400-Local graph acquisition module, 500-Saliency target acquisition module, 610-Processor, 620-Memory, 630-Input Device, 640 - Output Device.

具体实施方式Detailed ways

下面结合附图来具体描述本发明的优选实施例,其中,附图构成本申请一部分,并与本发明的实施例一起用于阐释本发明的原理,并非用于限定本发明的范围。The preferred embodiments of the present invention are specifically described below with reference to the accompanying drawings, wherein the accompanying drawings constitute a part of the present application, and together with the embodiments of the present invention, are used to explain the principles of the present invention, but are not used to limit the scope of the present invention.

目前的显著性目标检测方法大多是基于深度学习神经网络获取深度显著性特征的检测方法,其存在的问题包括:第一,目标检测大多是关注目标中心的位置,对于边缘的处理和关注存在一定缺陷;第二,对于有多个显著性目标的图像,各个目标之间的关联性没有用到显著性的推理当中。为此,本申请提出了一种矿井突发灾害场景中显著性目标的快速检测方法及装置,如图1所示,通过Unet网络获得输入图像对应的分割映射图像,通过全卷积FCN网络获得输入图像对应的显著性映射图像,并将分割映射图像和显著性映射图像融合得到融合映射图像,接着基于该融合映射图像得到多张局部图,最后通过图卷积推理网络GCN对多张局部图进行处理,获得各个目标之间的联系,根据目标之间的联系得到最终的显著性目标。该方法简单易行,易于实施,考虑了每一个分割目标的关联性,提高了获得的显著性目标的精度,提高了检测速度,具有较高的实用价值。Most of the current saliency target detection methods are detection methods based on deep learning neural networks to obtain deep salient features. The existing problems include: First, target detection mostly focuses on the position of the target center, and there is a certain degree of edge processing and attention. Defects; second, for images with multiple saliency targets, the correlation between each target is not used in saliency reasoning. To this end, the present application proposes a method and device for rapid detection of salient objects in a mine sudden disaster scene. As shown in Figure 1, the segmentation map image corresponding to the input image is obtained through the Unet network, and obtained through the fully convolutional FCN network. Input the saliency map image corresponding to the image, and fuse the segmentation map image and the saliency map image to obtain a fusion map image, then obtain multiple local maps based on the fusion map image, and finally pass the graph convolution inference network GCN to the multiple local maps. After processing, the connection between each target is obtained, and the final saliency target is obtained according to the connection between the targets. The method is simple, easy to implement, considers the relevance of each segmentation target, improves the accuracy of the obtained saliency target, improves the detection speed, and has high practical value.

本发明的一个具体实施例,公开了一种矿井突发灾害场景中显著性目标的快速检测方法,如图2所示,包括下述步骤S1~S5。A specific embodiment of the present invention discloses a rapid detection method for a salient target in a mine sudden disaster scene, as shown in FIG. 2 , including the following steps S1 to S5.

步骤S1、获取关于矿井突发灾害场景的输入图像。具体来说,矿井突发灾害场景的输入图像可以从煤矿井下的固定摄像机拍摄的视频图像中获取,也可以从煤矿井下的车载摄像机摄录的视频图像中获取。Step S1, acquiring an input image about a mine sudden disaster scene. Specifically, the input image of the mine sudden disaster scene can be obtained from the video image captured by the fixed camera in the coal mine, and can also be obtained from the video image captured by the vehicle-mounted camera in the coal mine.

步骤S2、构造Unet网络,将输入图像输入Unet网络,得到输入图像对应的分割映射图像,分割映射图像包含多个分割目标。具体来说,如图3所示,Unet网络包括编码器和解码器;其中,Step S2, constructing a Unet network, inputting the input image into the Unet network, and obtaining a segmentation map image corresponding to the input image, and the segmentation map image includes multiple segmentation targets. Specifically, as shown in Figure 3, the Unet network includes an encoder and a decoder; where,

编码器包括依次连接的两个第一网络和四个第二网络,其中,第一网络包括依次连接的卷积层Conv、批标准化层BN和激活层Relu;第二网络包括依次连接的最大池化层Maxpool和两个第一网络。详细地,输入图像通过两个第一网络生成64维特征图谱,64维特征图谱通过四个第二网络生成1024维特征图。The encoder includes two first networks and four second networks connected in sequence, wherein the first network includes a convolutional layer Conv, a batch normalization layer BN and an activation layer Relu connected in sequence; the second network includes a max pool connected in sequence Layer Maxpool and two first networks. In detail, the input image generates 64-dimensional feature maps through two first networks, and the 64-dimensional feature maps generate 1024-dimensional feature maps through four second networks.

解码器包括依次连接的四个第三网络和一个卷积层,其中,第三网络包括依次连接的转置卷积层UpConv和两个第一网络。详细地,通过四个第三网络将1024维特征图采样为原图大小的64维特征图谱,并在利用跳跃连接将编码器生成的中间特征图谱拼接到解码器生成的特征图谱上;最后通过一个卷积层将拼接后的64维特征图谱生成分割图像。The decoder includes four third networks connected in sequence and one convolutional layer, wherein the third network includes a transposed convolutional layer UpConv connected in sequence and two first networks. In detail, the 1024-dimensional feature map is sampled into a 64-dimensional feature map of the original image size through four third networks, and the intermediate feature map generated by the encoder is spliced to the feature map generated by the decoder using skip connections; A convolutional layer generates segmented images from the concatenated 64-dimensional feature maps.

通过Unet网络获取输入图像对应的分割映射图像,为后期进行图像融合及显著性目标检测提供了技术支撑和依据。同时,分割映射图像包含多个分割目标,在后期融合图像的基础上对分割目标处理,以得到局部图,考虑了各个分割目标之间的联系,有利于提高获得的显著性目标的精度。The segmentation map image corresponding to the input image is obtained through the Unet network, which provides technical support and basis for image fusion and saliency target detection in the later stage. At the same time, the segmentation map image contains multiple segmentation targets, and the segmentation targets are processed on the basis of the later fusion image to obtain a local map. Considering the relationship between each segmentation target, it is beneficial to improve the accuracy of the obtained saliency target.

步骤S3、构造全卷积FCN网络,将输入图像输入全卷积FCN网络,得到输入图像对应的显著性映射图像。全卷积FCN网络包括三部分;如图4所示,第一部分包含五个卷积层,每个卷积层后均连接一个池化层;其中,每个卷积层之后连接的池化层用于将图像的尺寸变为上一层尺寸的1/2。Step S3, constructing a fully convolutional FCN network, and inputting the input image into the fully convolutional FCN network to obtain a saliency map image corresponding to the input image. The fully convolutional FCN network consists of three parts; as shown in Figure 4, the first part contains five convolutional layers, each convolutional layer is connected with a pooling layer; among them, the pooling layer connected after each convolutional layer Used to change the size of the image to 1/2 the size of the previous layer.

第二部分包含依次连接的三个卷积层,分别将图像的维度变为4096、4096、1000维。The second part consists of three convolutional layers connected in sequence, changing the dimensions of the image to 4096, 4096, and 1000 dimensions, respectively.

第三部分用于对第二部分的输出进行上采样操作,上采样操作基于池化层和反卷积层实现,具体表现为进行上池化和反卷积操作。The third part is used to perform an up-sampling operation on the output of the second part. The up-sampling operation is implemented based on the pooling layer and the deconvolution layer, which is embodied in the up-pooling and deconvolution operations.

通过全卷积FCN网络获取输入图像对应的显著性映射图像,为后期进行图像融合及显著性目标检测提供了技术支撑和依据。通过Unet网络和全卷积FCN网络的配合,有利于提高获得的显著性目标的精度和检测速度。The saliency map image corresponding to the input image is obtained through the fully convolutional FCN network, which provides technical support and basis for the later image fusion and saliency target detection. Through the cooperation of the Unet network and the fully convolutional FCN network, it is beneficial to improve the accuracy and detection speed of the obtained salient objects.

步骤S4、将输入图像对应的分割映射图像和输入图像对应的显著性映射图像融合,得到融合映射图像,并基于融合映射图像得到局部图。具体来说,将分割映射图像与显著性映射图像融合,可得到一张融合映射图像,对该融合映射图像进行处理,可得到多张局部图。Step S4, fuse the segmentation map image corresponding to the input image and the saliency map image corresponding to the input image to obtain a fusion map image, and obtain a local map based on the fusion map image. Specifically, by fusing the segmentation map image with the saliency map image, a fused map image can be obtained, and by processing the fused map image, multiple local maps can be obtained.

优选地,基于融合映射图像得到局部图,包括下述步骤:Preferably, obtaining a local map based on the fusion map image includes the following steps:

步骤S401、将融合映射图像中的每一分割目标作为一个节点,并采用KD最近邻算法计算得到任意两个节点之间的距离和相对位置。具体来说,将输入图像输入Unet网络可得到分割映射图像,该分割映射图像包含多个分割目标,故融合映射图像中也包含多个分割目标,将每一分割目标作为一个节点,利用KD最近邻算法计算得到任意两个节点之间的距离和相对位置。详细地,KD最近邻算法的原理是:首先找到包含目标点的叶节点;然后从该叶节点出发,一次退回到父节点,不断查找与目标点最近的节点,当确定不可能存在更近的节点时停止,即可得到任意两个节点之间的距离和相对位置。Step S401 , take each segmentation target in the fusion map image as a node, and use the KD nearest neighbor algorithm to calculate the distance and relative position between any two nodes. Specifically, inputting the input image into the Unet network can obtain a segmentation map image. The segmentation map image contains multiple segmentation targets. Therefore, the fusion map image also contains multiple segmentation targets. Each segmentation target is regarded as a node. The neighbor algorithm calculates the distance and relative position between any two nodes. In detail, the principle of the KD nearest neighbor algorithm is: first find the leaf node containing the target point; then start from the leaf node, return to the parent node at a time, and continuously find the node closest to the target point, when it is determined that there is no closer node. When the node stops, the distance and relative position between any two nodes can be obtained.

步骤S402、将每一个节点作为中心节点,获得距离中心节点最短和次短的两个节点作为中心节点的一跳节点,并将距离每一个一跳节点的最短和次短的两个节点作为中心节点的二跳节点。Step S402, take each node as the central node, obtain two nodes with the shortest and second shortest distances from the central node as the one-hop nodes of the central node, and take the shortest and second shortest two nodes from each one-hop node as the center The node's two-hop node.

步骤S403、基于中心节点的一跳节点和二跳节点得到一个节点集合N,并获得节点集合中每一个节点的三个近邻节点Ne,并将节点集合中每一个节点与其在节点集合中的近邻节点连接,得到每一分割目标对应的局部图。具体来说,得到节点集合中每一个节点的三个近邻节点Ne后,判断每一个近邻节点Ne是否在节点集合N中,若近邻节点Ne在节点集合N中,将节点与其在节点集合中的近邻节点连接,可得到分割目标对应的局部图。遍历每一个中心节点,可得到每一个分割目标对应的局部图。In step S403, a node set N is obtained based on the one-hop node and the two-hop node of the central node, and the three neighbor nodes Ne of each node in the node set are obtained, and each node in the node set is compared with its neighbors in the node set. The nodes are connected to obtain the local graph corresponding to each segmentation target. Specifically, after obtaining the three neighbor nodes Ne of each node in the node set, it is judged whether each neighbor node Ne is in the node set N, if the neighbor node Ne is in the node set N, the node and its in the node set The adjacent nodes are connected to obtain the local graph corresponding to the segmentation target. By traversing each central node, the local graph corresponding to each segmentation target can be obtained.

将融合映射图像中的每一分割目标作为一个中心节点,以获得每一个分割目标对应的局部图,考虑了每一个分割目标之间的联系,有利于增强边缘的效果,使得最终获得的显著性目标精度更高。Taking each segmentation target in the fusion map image as a central node to obtain the local graph corresponding to each segmentation target, considering the connection between each segmentation target, it is beneficial to enhance the effect of the edge and make the final obtained saliency Target accuracy is higher.

步骤S5、基于局部图获得最终的显著性目标,包括下述步骤:Step S5, obtaining the final saliency target based on the local graph, including the following steps:

步骤S501、基于图卷积推理网络GCN将所有的局部图组合,得到组合图,并获取组合图中的每一节点与其对应的一跳节点的边权。具体来说,图卷积推理网络GCN包括依次连接的四个卷积层和两个全连接层,且在GCN网络中卷积层用到的激活函数是PReLU。详细地,图卷积推理网络GCN得到组合图中的每一节点与其对应的一跳节点的边权计算公式如下:Step S501 , combine all local graphs based on the graph convolutional reasoning network GCN to obtain a combined graph, and obtain the edge weight of each node in the combined graph and its corresponding one-hop node. Specifically, the graph convolutional reasoning network GCN includes four convolutional layers and two fully connected layers connected in sequence, and the activation function used in the convolutional layer in the GCN network is PReLU. In detail, the graph convolutional reasoning network GCN obtains the edge weight calculation formula of each node in the combined graph and its corresponding one-hop node as follows:

Y=σ[(X||GX)W]Y=σ[(X||GX)W]

上式中,Y为组合图中的每一节点与其对应的一跳节点的边权,X局部图的特征矩阵,G为聚合矩阵,W为图卷积推理网络GCN的权重矩阵,σ()表示非线性激活函数,其中,G=Λ-1/2-1/2,Λ表示对角矩阵,A表示邻接矩阵。In the above formula, Y is the edge weight of each node in the combined graph and its corresponding one-hop node, X is the feature matrix of the local graph, G is the aggregation matrix, W is the weight matrix of the graph convolution inference network GCN, σ() represents a nonlinear activation function, where G=Λ -1/2-1/2 , Λ represents a diagonal matrix, and A represents an adjacency matrix.

步骤S502、基于组合图中的每一节点与其对应的一跳节点的边权得到最终的显著性目标,包括下述步骤:Step S502, obtaining the final saliency target based on the edge weight of each node in the combined graph and its corresponding one-hop node, including the following steps:

判断每一节点与其对应的一跳节点的边权是否大于权重阈值,若是,合并所述节点与其对应的一跳节点,若否,不合并所述节点与其对应的一跳节点;遍历组合图中的所有节点,得到最终的显著性目标。Determine whether the edge weight of each node and its corresponding one-hop node is greater than the weight threshold, if so, merge the node and its corresponding one-hop node, if not, do not merge the node and its corresponding one-hop node; traverse the combined graph All nodes of , get the final saliency target.

具体来说,基于上述图卷积推理网络GCN能够将所有的局部图组合得到组合图,并得到组合图中的每一节点与其对应的一跳节点的边权。接着判断每一节点与其对应的一跳节点的边权是否大于权重阈值,其中,权重阈值在实际情况中基于人为设定,当每一节点与其对应的一跳节点的边权大于权重阈值时,将组合图中的每一节点与其对应的一跳节点的边权合并,遍历组合图中的所有节点,即可得到最终的显著性目标。Specifically, based on the above graph convolutional reasoning network GCN, all local graphs can be combined to obtain a combined graph, and the edge weight of each node in the combined graph and its corresponding one-hop node can be obtained. Next, it is judged whether the edge weight of each node and its corresponding one-hop node is greater than the weight threshold, wherein the weight threshold is manually set in the actual situation. The final saliency target can be obtained by merging the edge weights of each node in the combined graph and its corresponding one-hop node, and traversing all the nodes in the combined graph.

与现有技术相比,本实施例提供的矿井突发灾害场景中显著性目标的快速检测方法及装置,通过Unet网络得到的分割映射图像和全卷积FCN网络得到的显著性映射图像融合得到融合映射图像,接着基于该融合映射图像得到多张局部图,最后通过图卷积推理网络GCN对多张局部图进行处理,获得各个目标之间的联系,根据目标之间的联系得到最终的显著性目标。该方法简单易行,易于实施,考虑了每一个分割目标的关联性,提高了获得的显著性目标的精度,提高了检测速度,具有较高的实用价值。Compared with the prior art, the method and device for rapid detection of salient targets in a mine sudden disaster scene provided by this embodiment are obtained by fusing the segmentation map image obtained by the Unet network and the saliency map image obtained by the fully convolutional FCN network. Fuse the map image, and then obtain multiple local maps based on the fusion map image, and finally process the multiple local maps through the graph convolution inference network GCN to obtain the connection between each target. sexual goals. The method is simple, easy to implement, considers the relevance of each segmentation target, improves the accuracy of the obtained saliency target, improves the detection speed, and has high practical value.

本发明的另一个具体实施例,公开了一种矿井突发灾害场景中显著性目标的快速检测装置,如图5所示,包括:Another specific embodiment of the present invention discloses a rapid detection device for salient objects in a mine sudden disaster scene, as shown in FIG. 5 , including:

数据采集模块100,用于获取关于矿井突发灾害场景的输入图像;The data acquisition module 100 is used for acquiring the input image about the mine sudden disaster scene;

Unet网络构建模块200,用于构造Unet网络,将输入图像输入Unet网络,得到输入图像对应的分割映射图像,所述分割映射图像包含多个分割目标;The Unet network construction module 200 is used to construct the Unet network, and the input image is input into the Unet network to obtain a segmentation map image corresponding to the input image, and the segmentation map image includes a plurality of segmentation targets;

全卷积FCN网络构建模块300,用于构造全卷积FCN网络,将输入图像输入全卷积FCN网络,得到输入图像对应的显著性映射图像;The fully convolutional FCN network construction module 300 is used to construct a fully convolutional FCN network, input the input image into the fully convolutional FCN network, and obtain a saliency map image corresponding to the input image;

局部图获得模块400,用于将输入图像对应的分割映射图像和输入图像对应的显著性映射图像融合,得到融合映射图像,并基于所述融合映射图像得到局部图;The local map obtaining module 400 is used to fuse the segmentation map image corresponding to the input image and the saliency map image corresponding to the input image to obtain a fusion map image, and obtain a local map based on the fusion map image;

显著性目标获得模块500,基于局部图获得最终的显著性目标。The saliency target obtaining module 500 obtains the final saliency target based on the local map.

由于矿井突发灾害场景中显著性目标的快速检测装置的实现原理与前述矿井突发灾害场景中显著性目标的快速检测方法的实现原理相同,故这里不再赘述。Since the realization principle of the rapid detection device for salient objects in the mine sudden disaster scene is the same as the realization principle of the aforementioned rapid detection method for the salient objects in the mine sudden disaster scene, it will not be repeated here.

参见图6,本发明另一实施例还提供了执行上述实施例中矿井突发灾害场景中显著性目标的快速检测方法的电子设备。该电子设备包括:Referring to FIG. 6 , another embodiment of the present invention further provides an electronic device for implementing the method for rapid detection of a salient target in a mine sudden disaster scene in the above embodiment. The electronic equipment includes:

一个或多个处理器610以及存储器620,图6中以一个处理器610为例。One or more processors 610 and a memory 620, one processor 610 is taken as an example in FIG. 6 .

矿井突发灾害场景中显著性目标的快速检测方法的电子设备还可以包括:输入装置630和输出装置640。The electronic device of the method for rapid detection of a salient target in a mine sudden disaster scene may further include: an input device 630 and an output device 640 .

处理器610、存储器620、输入装置630和输出装置640可以通过总线或者其他方式连接,图6中以通过总线连接为例。The processor 610, the memory 620, the input device 630, and the output device 640 may be connected by a bus or in other manners, and the connection by a bus is taken as an example in FIG. 6 .

存储器620作为一种非易失性计算机可读存储介质,可用于存储非易失性软件程序、非易失性计算机可执行程序以及模块,如本发明的实施例中的矿井突发灾害场景中显著性目标的快速检测方法对应的程序指令/模块(单元)。处理器610通过运行存储在存储器620中的非易失性软件程序、指令以及模块,从而执行服务器的各种功能应用以及数据处理,即实现上述方法实施例图标显示方法。As a non-volatile computer-readable storage medium, the memory 620 can be used to store non-volatile software programs, non-volatile computer-executable programs and modules, such as in the mine sudden disaster scenario in the embodiment of the present invention. Program instructions/modules (units) corresponding to the method for rapid detection of salient objects. The processor 610 executes various functional applications and data processing of the server by running the non-volatile software programs, instructions and modules stored in the memory 620, that is, to implement the icon display method in the above method embodiment.

存储器620可以包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需要的应用程序;存储数据区可存储获取的应用程序的提醒事项的数量信息等。此外,存储器620可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件、闪存器件、或其他非易失性固态存储器件。在一些实施例中,存储器620可选包括相对于处理器610远程设置的存储器,这些远程存储器可以通过网络连接至列表项操作的处理装置。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。The memory 620 may include a storage program area and a storage data area, wherein the storage program area may store an operating system and an application program required by at least one function; the storage data area may store the quantity information of the acquired reminder items of the application program, and the like. Additionally, memory 620 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other non-volatile solid-state storage device. In some embodiments, the memory 620 may optionally include memory located remotely from the processor 610, which remote memory may be connected via a network to the processing device operated by the listing. Examples of such networks include, but are not limited to, the Internet, an intranet, a local area network, a mobile communication network, and combinations thereof.

输入装置630可接收输入的数字或字符信息,以及产生与矿井突发灾害场景中显著性目标的快速检测装置的用户设置以及功能控制有关的键信号输入。输出装置640可包括显示屏等显示设备。The input device 630 may receive the input numerical or character information, and generate key signal input related to the user setting and function control of the rapid detection device of the salient object in the mine sudden disaster scene. The output device 640 may include a display device such as a display screen.

所述一个或者多个模块存储在所述存储器620中,当被所述一个或者多个处理器610执行时,执行上述任意方法实施例中的矿井突发灾害场景中显著性目标的快速检测方法。The one or more modules are stored in the memory 620, and when executed by the one or more processors 610, execute the method for rapid detection of a salient target in a mine sudden disaster scenario in any of the above method embodiments .

上述产品可执行本发明的实施例所提供的方法,具备执行方法相应的功能模块和有益效果。未在本实施例中详尽描述的技术细节,可参见本发明的实施例所提供的方法。The above product can execute the method provided by the embodiments of the present invention, and has corresponding functional modules and beneficial effects for executing the method. For technical details not described in detail in this embodiment, reference may be made to the methods provided by the embodiments of the present invention.

本发明的实施例的电子设备可以以多种形式存在,包括但不限于:Electronic devices of embodiments of the present invention may exist in various forms, including but not limited to:

(1)移动通信设备:这类设备的特点是具备移动通信功能,并且以提供话音、数据通信为主要目标。这类终端包括:智能手机(例如iPhone)、多媒体手机、功能性手机,以及低端手机等。(1) Mobile communication equipment: This type of equipment is characterized by having mobile communication functions, and its main goal is to provide voice and data communication. Such terminals include: smart phones (eg iPhone), multimedia phones, functional phones, and low-end phones.

(2)超移动个人计算机设备:这类设备属于个人计算机的范畴,有计算和处理功能,一般也具备移动上网特性。这类终端包括:PDA、MID和UMPC设备等,例如iPad。(2) Ultra-mobile personal computer equipment: This type of equipment belongs to the category of personal computers, has computing and processing functions, and generally has the characteristics of mobile Internet access. Such terminals include: PDAs, MIDs, and UMPC devices, such as iPads.

(3)便携式娱乐设备:这类设备可以显示和播放多媒体内容。该类设备包括:音频、视频播放器(例如iPod),掌上游戏机,电子书,以及智能玩具和便携式车载导航设备。(3) Portable entertainment equipment: This type of equipment can display and play multimedia content. Such devices include: audio and video players (eg iPod), handheld game consoles, e-books, as well as smart toys and portable car navigation devices.

(4)服务器:提供计算服务的设备,服务器的构成包括处理器、硬盘、内存、系统总线等,服务器和通用的计算机架构类似,但是由于需要提供高可靠的服务,因此在处理能力、稳定性、可靠性、安全性、可扩展性、可管理性等方面要求较高。(4) Server: A device that provides computing services. The composition of the server includes a processor, hard disk, memory, system bus, etc. The server is similar to a general computer architecture, but due to the need to provide highly reliable services, the processing capacity, stability , reliability, security, scalability, manageability and other aspects of high requirements.

(5)其他具有提醒事项记录功能的电子装置。(5) Other electronic devices with reminder record function.

以上所描述的装置实施例仅仅是示意性的,其中所述作为分离部件说明的单元(模块)可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。The device embodiments described above are only illustrative, wherein the units (modules) described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, That is, it can be located in one place, or it can be distributed to multiple network elements. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution in this embodiment.

本发明实施例提供了一种非暂态计算机可读存储介质,所述计算机存储介质存储有计算机可执行指令,其中,当所述计算机可执行指令被电子设备执行时,使所述电子设备上执行上述任意方法实施例中的矿井突发灾害场景中显著性目标的快速检测方法。An embodiment of the present invention provides a non-transitory computer-readable storage medium, where computer-executable instructions are stored in the computer storage medium, wherein when the computer-executable instructions are executed by an electronic device, Execute the method for rapid detection of salient objects in a mine sudden disaster scene in any of the above method embodiments.

本发明实施例提供了一种计算机程序产品,其中,所述计算机程序产品包括存储在非暂态计算机可读存储介质上的计算机程序,所述计算机程序包括程序指令,其中,当所述程序指令被电子设备执行时,使所述电子设备执行上述任意方法实施例中的矿井突发灾害场景中显著性目标的快速检测方法。An embodiment of the present invention provides a computer program product, wherein the computer program product includes a computer program stored on a non-transitory computer-readable storage medium, and the computer program includes program instructions, wherein when the program instructions are When executed by an electronic device, the electronic device is made to execute the method for rapid detection of a salient target in a mine sudden disaster scene in any of the above method embodiments.

通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到各实施例可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件。基于这样的理解,上述技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品可以存储在计算机可读存储介质中,如ROM/RAM、磁碟、光盘等,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行各个实施例或者实施例的某些部分所述的方法。From the description of the above embodiments, those skilled in the art can clearly understand that each embodiment can be implemented by means of software plus a necessary general hardware platform, and certainly can also be implemented by hardware. Based on this understanding, the above-mentioned technical solutions can be embodied in the form of software products in essence or the parts that make contributions to the prior art, and the computer software products can be stored in computer-readable storage media, such as ROM/RAM, magnetic A disc, an optical disc, etc., includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to perform the methods described in various embodiments or some parts of the embodiments.

以上所述,仅为本发明较佳的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到的变化或替换,都应涵盖在本发明的保护范围之内。The above description is only a preferred embodiment of the present invention, but the protection scope of the present invention is not limited to this. Substitutions should be covered within the protection scope of the present invention.

Claims (10)

1. A method for rapidly detecting a significant target in a mine sudden disaster scene is characterized by comprising the following steps:
acquiring an input image about a mine sudden disaster scene;
constructing a Unet network, inputting the input image into the Unet network to obtain a segmentation mapping image corresponding to the input image, wherein the segmentation mapping image comprises a plurality of segmentation targets;
constructing a full convolution FCN network, inputting the input image into the full convolution FCN network, and obtaining a significance mapping image corresponding to the input image;
fusing the segmentation mapping image corresponding to the input image with the saliency mapping image corresponding to the input image to obtain a fused mapping image, and obtaining a local image based on the fused mapping image;
and obtaining a final significance target based on the local graph.
2. The method of claim 1, wherein the Unet network comprises an encoder and a decoder;
the encoder comprises two first networks and four second networks which are connected in sequence, wherein the first networks comprise a convolution layer, a batch normalization layer and an activation layer which are connected in sequence; the second network comprises a maximum pooling layer and two first networks which are connected in sequence;
the decoder comprises four third networks and a convolutional layer which are sequentially connected, wherein the third networks comprise a transposed convolutional layer and two first networks which are sequentially connected.
3. The method for rapidly detecting a salient object in a mine sudden-disaster scenario according to claim 2, wherein the fully-convoluted FCN network comprises three parts; wherein,
the first part comprises five convolution layers, and a pooling layer is connected behind each convolution layer;
the second part comprises three convolution layers which are connected in sequence;
and the third part is used for performing an up-sampling operation on the output of the second part, wherein the up-sampling operation is realized based on the pooling layer and the deconvolution layer.
4. The method for rapidly detecting the saliency target in the mine sudden disaster scene according to claim 1 is characterized in that a local map is obtained based on the fusion mapping image, and the method comprises the following steps:
taking each segmentation target in the fusion mapping image as a node, and calculating by adopting a KD nearest neighbor algorithm to obtain the distance and the relative position between any two nodes;
taking each node as a central node, obtaining two nodes which are shortest and second shortest from the central node as a first-hop node of the central node, and taking the two nodes which are shortest and second shortest from each first-hop node as a second-hop node of the central node;
and obtaining a node set based on the first hop node and the second hop node of the central node, obtaining a neighbor node of each node in the node set, and connecting each node in the node set with the neighbor node in the node set to obtain a local graph corresponding to each segmentation target.
5. The method for rapidly detecting the saliency target in the mine sudden disaster scene according to claim 4 is characterized in that the final saliency target is obtained based on the local map, and the method comprises the following steps:
combining all local graphs based on a graph convolution reasoning network GCN to obtain a combined graph, and acquiring the edge weight of each node in the combined graph and a corresponding one-hop node;
and obtaining a final significance target based on the edge weight of each node in the combined graph and the corresponding one-hop node.
6. The method for rapidly detecting the saliency target in the mine sudden disaster scene as claimed in claim 5, wherein said graph convolution inference network GCN includes four convolution layers and two fully connected layers connected in sequence.
7. The method for rapidly detecting the saliency target in the mine sudden disaster scene according to claim 6, wherein the final saliency target is obtained based on the edge weight of each node in the combined graph and the corresponding one-hop node, and the method comprises the following steps:
judging whether the edge weight of each node and the corresponding one-hop node is greater than a weight threshold value, if so, combining the node and the corresponding one-hop node, and if not, not combining the node and the corresponding one-hop node;
and traversing all nodes in the combined graph to obtain a final saliency target.
8. A device for rapidly detecting a significant target in a mine sudden disaster scene is characterized by comprising:
the data acquisition module is used for acquiring an input image related to a mine sudden disaster scene;
the Unet network construction module is used for constructing the Unet network, inputting the input image into the Unet network, and obtaining a segmentation mapping image corresponding to the input image, wherein the segmentation mapping image comprises a plurality of segmentation targets;
the full-convolution FCN network construction module is used for constructing a full-convolution FCN network, inputting the input image into the full-convolution FCN network and obtaining a significance mapping image corresponding to the input image;
a local map obtaining module, configured to fuse the segmentation mapping image corresponding to the input image and the saliency mapping image corresponding to the input image to obtain a fused mapping image, and obtain a local map based on the fused mapping image;
and the saliency target obtaining module is used for obtaining a final saliency target based on the local map.
9. The apparatus for rapidly detecting a salient object in a mine sudden disaster scene according to claim 8, wherein said Unet network comprises an encoder and a decoder;
the encoder comprises two first networks and four second networks which are connected in sequence, wherein the first networks comprise a convolution layer, a batch normalization layer and an activation layer which are connected in sequence; the second network comprises a maximum pooling layer and two first networks which are connected in sequence;
the decoder comprises four third networks and a convolutional layer which are sequentially connected, wherein the third networks comprise a transposed convolutional layer and two first networks which are sequentially connected.
10. The apparatus for rapidly detecting a salient object in a mine sudden-disaster scenario according to claim 9, wherein the fully-convoluted FCN network comprises three parts; wherein,
the first part comprises five convolution layers, and a pooling layer is connected behind each convolution layer;
the second part comprises three convolution layers which are connected in sequence;
and the third part is used for performing an up-sampling operation on the output of the second part, wherein the up-sampling operation is realized based on the pooling layer and the deconvolution layer.
CN202011541385.4A 2020-12-23 2020-12-23 Rapid detection method and device for salient targets in mine sudden disaster scene Active CN112541912B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011541385.4A CN112541912B (en) 2020-12-23 2020-12-23 Rapid detection method and device for salient targets in mine sudden disaster scene

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011541385.4A CN112541912B (en) 2020-12-23 2020-12-23 Rapid detection method and device for salient targets in mine sudden disaster scene

Publications (2)

Publication Number Publication Date
CN112541912A true CN112541912A (en) 2021-03-23
CN112541912B CN112541912B (en) 2024-03-12

Family

ID=75017755

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011541385.4A Active CN112541912B (en) 2020-12-23 2020-12-23 Rapid detection method and device for salient targets in mine sudden disaster scene

Country Status (1)

Country Link
CN (1) CN112541912B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113192093A (en) * 2021-05-10 2021-07-30 新疆大学 Quick saliency target detection method based on double-flow network structure
CN113762166A (en) * 2021-09-09 2021-12-07 中国矿业大学 Small target detection improvement method and system based on wearable equipment
CN114998074A (en) * 2022-05-18 2022-09-02 成都秦川物联网科技股份有限公司 Smart city post-disaster contact method and system based on Internet of things
CN115273044A (en) * 2022-07-15 2022-11-01 哈尔滨市科佳通用机电股份有限公司 Recognition and detection method of vehicle door damage based on improved graph convolutional network
CN115509406A (en) * 2022-11-23 2022-12-23 煤炭科学研究总院有限公司 Coal mine multi-scene fusion method and device, storage medium and electronic equipment

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120275701A1 (en) * 2011-04-26 2012-11-01 Minwoo Park Identifying high saliency regions in digital images
US8363939B1 (en) * 2006-10-06 2013-01-29 Hrl Laboratories, Llc Visual attention and segmentation system
CN106157319A (en) * 2016-07-28 2016-11-23 哈尔滨工业大学 The significance detection method that region based on convolutional neural networks and Pixel-level merge
CN106909924A (en) * 2017-02-18 2017-06-30 北京工业大学 A kind of remote sensing image method for quickly retrieving based on depth conspicuousness
US20170358273A1 (en) * 2016-06-10 2017-12-14 Yen4Ken, Inc. Systems and methods for resolution adjustment of streamed video imaging
CN107506761A (en) * 2017-08-30 2017-12-22 山东大学 Brain image dividing method and system based on notable inquiry learning convolutional neural networks
CN107680106A (en) * 2017-10-13 2018-02-09 南京航空航天大学 A kind of conspicuousness object detection method based on Faster R CNN
CN108537239A (en) * 2018-04-18 2018-09-14 西南科技大学 A kind of method of saliency target detection
CN109376676A (en) * 2018-11-01 2019-02-22 哈尔滨工业大学 Safety early warning method for construction personnel on highway engineering site based on UAV platform
CN109800629A (en) * 2018-12-05 2019-05-24 天津大学 A kind of Remote Sensing Target detection method based on convolutional neural networks
WO2019104767A1 (en) * 2017-11-28 2019-06-06 河海大学常州校区 Fabric defect detection method based on deep convolutional neural network and visual saliency
CN110633708A (en) * 2019-06-28 2019-12-31 中国人民解放军军事科学院国防科技创新研究院 Deep network significance detection method based on global model and local optimization
CN110866896A (en) * 2019-10-29 2020-03-06 中国地质大学(武汉) Image saliency object detection method based on k-means and level set superpixel segmentation
CN111104943A (en) * 2019-12-17 2020-05-05 西安电子科技大学 Color image region-of-interest extraction method based on decision-level fusion
CN111209915A (en) * 2019-12-25 2020-05-29 上海航天控制技术研究所 Three-dimensional image synchronous identification and segmentation method based on deep learning
US20200184252A1 (en) * 2018-12-10 2020-06-11 International Business Machines Corporation Deep Learning Network for Salient Region Identification in Images

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8363939B1 (en) * 2006-10-06 2013-01-29 Hrl Laboratories, Llc Visual attention and segmentation system
US20120275701A1 (en) * 2011-04-26 2012-11-01 Minwoo Park Identifying high saliency regions in digital images
US20170358273A1 (en) * 2016-06-10 2017-12-14 Yen4Ken, Inc. Systems and methods for resolution adjustment of streamed video imaging
CN106157319A (en) * 2016-07-28 2016-11-23 哈尔滨工业大学 The significance detection method that region based on convolutional neural networks and Pixel-level merge
CN106909924A (en) * 2017-02-18 2017-06-30 北京工业大学 A kind of remote sensing image method for quickly retrieving based on depth conspicuousness
CN107506761A (en) * 2017-08-30 2017-12-22 山东大学 Brain image dividing method and system based on notable inquiry learning convolutional neural networks
CN107680106A (en) * 2017-10-13 2018-02-09 南京航空航天大学 A kind of conspicuousness object detection method based on Faster R CNN
WO2019104767A1 (en) * 2017-11-28 2019-06-06 河海大学常州校区 Fabric defect detection method based on deep convolutional neural network and visual saliency
CN108537239A (en) * 2018-04-18 2018-09-14 西南科技大学 A kind of method of saliency target detection
CN109376676A (en) * 2018-11-01 2019-02-22 哈尔滨工业大学 Safety early warning method for construction personnel on highway engineering site based on UAV platform
CN109800629A (en) * 2018-12-05 2019-05-24 天津大学 A kind of Remote Sensing Target detection method based on convolutional neural networks
US20200184252A1 (en) * 2018-12-10 2020-06-11 International Business Machines Corporation Deep Learning Network for Salient Region Identification in Images
CN110633708A (en) * 2019-06-28 2019-12-31 中国人民解放军军事科学院国防科技创新研究院 Deep network significance detection method based on global model and local optimization
CN110866896A (en) * 2019-10-29 2020-03-06 中国地质大学(武汉) Image saliency object detection method based on k-means and level set superpixel segmentation
CN111104943A (en) * 2019-12-17 2020-05-05 西安电子科技大学 Color image region-of-interest extraction method based on decision-level fusion
CN111209915A (en) * 2019-12-25 2020-05-29 上海航天控制技术研究所 Three-dimensional image synchronous identification and segmentation method based on deep learning

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
JUN WANG; GUOQING YE; JIAHAN LI; QIQI KOU: "Improved Object-Tracking Algorithm for an Underground Mining Environment", 《2019 6TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING & MACHINE INTELLIGENCE (ISCMI)》 *
WEITING LIU; XUE QING; JIAN ZHOU: "A novel image segmentation algorithm based on visual saliency detection and integrated feature extraction", 《2016 INTERNATIONAL CONFERENCE ON COMMUNICATION AND ELECTRONICS SYSTEMS (ICCES)》 *
刘瑞航: "视觉显著性预测方法研究", 《CNKI》 *
程德强,刘瑞航,李佳函等: "Activity guided multi-scales collaboration based on scaled-CNN for saliency prediction", 《IMAGE AND VISION COMPUTING》 *
钱生;陈宗海;林名强;张陈斌: "基于条件随机场和图像分割的显著性检测", 《自动化学报》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113192093A (en) * 2021-05-10 2021-07-30 新疆大学 Quick saliency target detection method based on double-flow network structure
CN113192093B (en) * 2021-05-10 2023-04-18 新疆大学 Quick saliency target detection method based on double-flow network structure
CN113762166A (en) * 2021-09-09 2021-12-07 中国矿业大学 Small target detection improvement method and system based on wearable equipment
CN114998074A (en) * 2022-05-18 2022-09-02 成都秦川物联网科技股份有限公司 Smart city post-disaster contact method and system based on Internet of things
CN114998074B (en) * 2022-05-18 2024-07-05 成都秦川物联网科技股份有限公司 Internet of things-based smart city post-disaster contact method and system
US12266189B2 (en) 2022-05-18 2025-04-01 Chengdu Qinchuan Iot Technology Co., Ltd. Methods and systems for post-disaster contact in smart city based on the internet of things
CN115273044A (en) * 2022-07-15 2022-11-01 哈尔滨市科佳通用机电股份有限公司 Recognition and detection method of vehicle door damage based on improved graph convolutional network
CN115509406A (en) * 2022-11-23 2022-12-23 煤炭科学研究总院有限公司 Coal mine multi-scene fusion method and device, storage medium and electronic equipment
CN115509406B (en) * 2022-11-23 2023-03-14 煤炭科学研究总院有限公司 Coal mine multi-scene fusion method and device, storage medium and electronic equipment

Also Published As

Publication number Publication date
CN112541912B (en) 2024-03-12

Similar Documents

Publication Publication Date Title
CN112541912A (en) Method and device for rapidly detecting saliency target in mine sudden disaster scene
CN110132290B (en) Intelligent driving road side equipment perception information fusion processing method, device and equipment
CN116186354B (en) Method, apparatus, electronic device, and computer-readable medium for displaying regional image
CN112232311B (en) Face tracking method and device and electronic equipment
CN112328911B (en) Place recommending method, device, equipment and storage medium
CN113063421A (en) Navigation method and related device, mobile terminal and computer readable storage medium
CN110069997B (en) Scene classification method and device and electronic equipment
CN116894802A (en) Image enhancement method, device, computer equipment and storage medium
CN112258647B (en) Map reconstruction method and device, computer readable medium and electronic equipment
WO2025066516A1 (en) Image processing method and related device
CN116883586B (en) Terrain semantic map construction method, system and product based on binocular camera
CN115984417B (en) Semantic mapping method, semantic mapping device and storage medium
CN110781066A (en) User behavior analysis method, device, equipment and storage medium
CN114648713B (en) Video classification method, device, electronic equipment and computer readable storage medium
CN116503833A (en) Urban high-complexity detection scene-based vehicle illegal parking detection method
CN109788431B (en) Bluetooth positioning method, device, equipment and system based on adjacent node group
CN115496185A (en) Research on Graph Enhanced Capsule Network for Serialized Recommendation
JP7166583B2 (en) Map information update system and method
CN114756634A (en) Method and device for discovering interest point change, electronic equipment and storage medium
CN111738282A (en) An artificial intelligence-based image recognition method and related equipment
CN113742801B (en) Road data processing method, device and electronic device
CN112927291B (en) Pose determining method and device of three-dimensional object, electronic equipment and storage medium
CN116614859A (en) A method and system for identifying key nodes in an Ad Hoc network
CN117789059A (en) A method, device, equipment, medium and drone for detecting tiny foreign objects
CN110222590B (en) Image difference judgment method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant