CN104835175B - Object detection method in a kind of nuclear environment of view-based access control model attention mechanism - Google Patents
Object detection method in a kind of nuclear environment of view-based access control model attention mechanism Download PDFInfo
- Publication number
- CN104835175B CN104835175B CN201510272424.8A CN201510272424A CN104835175B CN 104835175 B CN104835175 B CN 104835175B CN 201510272424 A CN201510272424 A CN 201510272424A CN 104835175 B CN104835175 B CN 104835175B
- Authority
- CN
- China
- Prior art keywords
- feature
- map
- image
- color
- brightness
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
- G06T7/337—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving reference images or patches
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Description
技术领域technical field
本发明涉及图像信息处理技术领域,具体涉及一种基于视觉注意机制的核环境中目标检测方法,该方法具体采用自上而下与自下而上结合的双向视觉注意模型,可用于特殊环境下的目标检测。The invention relates to the technical field of image information processing, in particular to a method for detecting objects in a nuclear environment based on a visual attention mechanism. The method specifically adopts a top-down and bottom-up bidirectional visual attention model, which can be used in special environments target detection.
背景技术Background technique
随着信息技术的快速发展、数据的膨胀,人们对信息处理的速度及精确度的要求越来越高了。理想的信息处理过程是,只需要处理与任务相关的那部分信息,但实际的信息处理过程是,需要处理许多与任务不相关的信息。因此,如何快速找到并仅仅处理与任务相关的那部分信息变得非常重要。With the rapid development of information technology and the expansion of data, people have higher and higher requirements for the speed and accuracy of information processing. The ideal information processing process is that only part of the information related to the task needs to be processed, but the actual information processing process needs to process a lot of information that is not related to the task. Therefore, how to quickly find and process only the part of the information related to the task becomes very important.
人类视觉注意机制为快速找到并处理与任务相关的那部分信息提供了一种新的研究思路。研究表明,人类视觉具有超强的信息处理能力,面对实时变化的各种信息,总能针对与其最相关的部分及时做出反应,而自动忽略不相关的部分。把人类视觉注意机制应用在图像处理领域,形成图像处理领域的视觉注意机制,能够高效、准确地处理图像信息。因此,在图像处理过程中如何模仿人的视觉注意机制,快速找到图像中的目标区域,对于图像处理的实时性有着重要的意义。The human visual attention mechanism provides a new research idea for quickly finding and processing the part of the information related to the task. Studies have shown that human vision has super information processing capabilities. Faced with various information that changes in real time, it can always respond to the most relevant parts in a timely manner, while automatically ignoring irrelevant parts. Apply the human visual attention mechanism to the field of image processing to form a visual attention mechanism in the field of image processing, which can process image information efficiently and accurately. Therefore, how to imitate the human visual attention mechanism in the process of image processing and quickly find the target area in the image is of great significance to the real-time performance of image processing.
目前,已经有越来越多的研究者投入如何能快速、精准检测显著区域的研究中,并研究出了很多模型,其中一些典型的模型有:At present, more and more researchers have invested in the research on how to quickly and accurately detect salient regions, and have developed many models, some of which are typical models:
1)Itti模型:其主要过程是从输入图像中提取多方面的特征,如颜色、方向、亮度等,通过高期金字塔和中央周边操作算子形成各个特征的关注图,然后归一化组合得到显著图。在此基础上,通过胜者全取神经网络相互竞争,使得显著区胜出。该方法对局部显著性进行了较好的度量。但没有充分考虑图像的全局信息;且没有考虑实际任务的需求,只属于自下而上的单向注意模型。1) Itti model: its main process is to extract various features from the input image, such as color, direction, brightness, etc., and form an attention map of each feature through the high-level pyramid and central peripheral operation operators, and then normalize and combine to obtain Significant figure. On this basis, the neural networks compete with each other through the winner-take-all neural network, so that the salient region wins. This method provides a good measure of local saliency. But it does not fully consider the global information of the image; and does not consider the needs of the actual task, and only belongs to the bottom-up unidirectional attention model.
2)Stentiford模型:该方法将图像的显著性用视觉注意图表示,其基本思想是当图像某区域特征在图像其他区域中出现频率越少,其区域显著性就越高;通过抑制图像中具有相同模式的区域得到视觉注意图,用于表示显著性。该方法考虑了目标整体性,对图像进行了全局显著性度量,但该模型依然只属于自下而上的注意模型,没有确定的物理意义,也没有根据任务对目标的重要程度进行判断。2) Stentiford model: This method represents the salience of the image with a visual attention map. The basic idea is that when the feature of a certain region of the image appears less frequently in other regions of the image, the salience of the region will be higher; Regions of the same pattern get a visual attention map, which is used to represent saliency. This method considers the integrity of the target and measures the global saliency of the image, but the model still only belongs to the bottom-up attention model, which has no definite physical meaning, and does not judge the importance of the target according to the task.
3)HOAM模型:该模型是以强度和方向图作为引导视觉注意的早期特征。被注意的单元不是空间的某个点或某个区域,而是具有确定物理意义的完整目标。该方法首先需要假设图像已经分成了若干具有物理意义的目标或目标组合,因此需要人工进行干预。3) HOAM model: This model uses intensity and orientation as early features to guide visual attention. The attention unit is not a certain point or a certain area in space, but a complete object with definite physical meaning. This method first needs to assume that the image has been divided into several physically meaningful objects or object combinations, so manual intervention is required.
发明内容Contents of the invention
针对于现有技术中存在的上述问题,本发明提供了一种基于视觉注意机制的核环境中目标检测方法,使用该方法能够大幅度提高图像中目标区域检测的精度,提取的显著性目标具有更好的鲁棒性和准确性,同时也大大提高了图像的处理效率。Aiming at the above-mentioned problems existing in the prior art, the present invention provides a method for detecting objects in a nuclear environment based on a visual attention mechanism. Using this method, the accuracy of object region detection in an image can be greatly improved, and the extracted salient objects have Better robustness and accuracy, but also greatly improve the image processing efficiency.
为实现上述目的,本发明的一个具体实施例所采取的技术方案是:In order to achieve the above object, the technical scheme that a specific embodiment of the present invention takes is:
S1、获取通过普通相机采集的所述目标的图像,提取图像的亮度、颜色和方向特征,分别得到亮度特征图、颜色特征图和方向特征图;S1. Obtain an image of the target collected by a common camera, extract brightness, color, and direction features of the image, and obtain a brightness feature map, a color feature map, and a direction feature map, respectively;
S2、通过高斯金字塔和中央周边算子的方法对亮度特征图、颜色特征图和方向特征图进行计算,分别得到6幅亮度视差图、12幅颜色视差图和24幅方向视差图;S2. Calculate the luminance feature map, color feature map and direction feature map through the Gaussian pyramid and the method of the central peripheral operator, and obtain 6 luminance parallax maps, 12 color parallax maps and 24 directional parallax maps;
S3、分别对6幅亮度视差图、12幅颜色视差图和24幅方向视差图进行归一化处理,得到亮度显著图颜色显著图和方向显著图 S3. Perform normalization processing on 6 luminance disparity maps, 12 color disparity maps and 24 directional disparity maps respectively to obtain a luminance saliency map color saliency map and directional saliency map
S4、从显著图中选取最显著的点,以点为显著点,在对应的特征显著图中采用区域生长的方式进行分割,得到感兴趣区域;S4. Select the most significant point from the saliency map, take the point as the salient point, and segment the corresponding feature saliency map by means of region growth to obtain the region of interest;
S5、获取γ相机采集的含有辐射强度分布信息的图像;S5. Obtain an image containing radiation intensity distribution information collected by the gamma camera;
S6、分别提取混合图像和普通相机图像感兴趣区域的关键点;S6, respectively extracting the key points of the region of interest of the mixed image and the common camera image;
S7、将关键点分别生成特征向量;S7. Generating feature vectors from the key points respectively;
S8、将感兴趣区域关键点的特征向量与混合图像关键点的特征向量进行匹配,如果符合匹配条件,则目标为作业目标。S8. Match the eigenvectors of the key points of the region of interest with the eigenvectors of the key points of the mixed image, and if the matching condition is met, the target is the job target.
本发明的方法产生的有益效果为:The beneficial effect that method of the present invention produces is:
首先利用自下而上数据驱动注意模型的优势,直接从普通相机采集的图像中提取出若干感兴趣区域,大大降低后期匹配过程的计算量;然后用г相机获取所述目标的计量强度分布和现场灰度的混合图进行特征匹配,建立自上而下与自下而上结合的双向视觉注意模型。因此,大幅度提高了图像中目标区域检测的精度,而且匹配过程消除了场景中不相关区域的干扰,使提取的显著性目标具有更好的鲁棒性和准确性,同时也大幅度提高了处理效率。Firstly, by taking advantage of the bottom-up data-driven attention model, several regions of interest are directly extracted from images collected by common cameras, which greatly reduces the amount of computation in the later matching process; The mixed image of the scene grayscale is used for feature matching, and a two-way visual attention model combining top-down and bottom-up is established. Therefore, the accuracy of target area detection in the image is greatly improved, and the matching process eliminates the interference of irrelevant areas in the scene, so that the extracted salient targets have better robustness and accuracy, and also greatly improve Processing efficiency.
附图说明Description of drawings
图1所示为本发明的一种基于视觉注意机制的核环境中目标检测方法的一个技术方案的流程图;Fig. 1 shows the flow chart of a technical scheme of a kind of target detection method in the nuclear environment based on visual attention mechanism of the present invention;
图2所示为本发明的DOG差分金字塔形成过程的示意图;Fig. 2 shows the schematic diagram of the DOG difference pyramid formation process of the present invention;
图3所示为本发明的DOG函数的极值点检测的示意图。FIG. 3 is a schematic diagram of extreme point detection of the DOG function of the present invention.
具体实施方式Detailed ways
为使本发明的目的、技术方案和优点更加清楚,下面将结合本发明具体实施例及相应的附图对本发明技术方案进行清楚、完整地描述。In order to make the purpose, technical solution and advantages of the present invention clearer, the technical solution of the present invention will be clearly and completely described below in conjunction with specific embodiments of the present invention and corresponding drawings.
参考图1,图1所示为本发明的一种基于视觉注意机制的核环境中显著目标的检测方法的一个实施例S100的流程图;实施例S100包括如下步骤S1至S8。Referring to FIG. 1 , FIG. 1 is a flowchart of an embodiment S100 of a method for detecting salient objects in a nuclear environment based on a visual attention mechanism of the present invention; the embodiment S100 includes the following steps S1 to S8.
在步骤S1中,获取通过普通相机采集并进行预处理的目标的图像,提取图像的亮度、颜色和方向特征,分别得到亮度特征图、颜色特征图和方向特征图。In step S1, the image of the target collected and preprocessed by the ordinary camera is obtained, the brightness, color and direction features of the image are extracted, and the brightness feature map, color feature map and direction feature map are respectively obtained.
在步骤S2中,通过高斯金字塔和中央周边算子的方法对亮度特征图、颜色特征图和方向特征图进行计算,分别得到6幅亮度视差图、12幅颜色视差图和 24幅方向视差图。In step S2, the brightness feature map, color feature map and direction feature map are calculated by the method of Gaussian pyramid and central peripheral operator, and 6 brightness disparity maps, 12 color disparity maps and 24 direction disparity maps are obtained respectively.
在本发明的一个实施例中,采用中央周边差的采样方式获取图像的亮度、颜色和方向特征。每一个特征有一组类似视觉感知区域的线性“中间-外围”算子计算。In one embodiment of the present invention, the brightness, color and direction features of the image are obtained by using the sampling method of the difference between the center and the periphery. Each feature is computed by a set of linear "middle-periphery" operators similar to the visual perception area.
在本发明的一个实施例中,中央和外围的尺度可以做如下设定:中央尺度 c∈{2,3,4},外围尺度相应的为s=c+δ,δ∈{3,4}。两个图的相交尺度差分通过精细尺度插补和点对点相减获得,使用不同的尺度得到多尺度特征的提取。In one embodiment of the present invention, the central and peripheral scales can be set as follows: the central scale c∈{2,3,4}, the peripheral scale is correspondingly s=c+δ, δ∈{3,4} . The intersecting scale difference of the two graphs is obtained by fine-scale interpolation and point-to-point subtraction, using different scales to obtain multi-scale feature extraction.
用r,g,b表示输入图像的红、绿、蓝三个颜色通道的像素值,它用来建立高斯金字塔I(x),x=[0......8]表示尺度级。高斯金字塔图像是一幅图像被高斯滤波后形成的一系列图像集合,随着高斯滤波次数的增加分辨率会逐渐降低。金字塔最底层是未经滤波的图像,分辨率最高,而顶层是图像的低分辨率表示。这样图像金字塔由三个颜色通道的金字塔图像取均值获得,如式(1)所示:Use r, g, b to represent the pixel values of the red, green, and blue color channels of the input image, which are used to build a Gaussian pyramid I(x), and x=[0...8] represents the scale level. The Gaussian pyramid image is a series of image collections formed by Gaussian filtering of an image, and the resolution will gradually decrease as the number of Gaussian filtering increases. The bottom layer of the pyramid is the unfiltered image, with the highest resolution, while the top layer is a lower resolution representation of the image. In this way, the image pyramid is obtained by taking the mean value of the pyramid images of the three color channels, as shown in formula (1):
通过中央和外围不同尺度下的图像进行差分,获得6幅中央周边差结构的亮度图,如式(2)所示:Through the difference between the central and peripheral images at different scales, six brightness maps of the central and peripheral difference structures are obtained, as shown in formula (2):
I(c,s)=|I(c)-I(s)| (2)I(c,s)=|I(c)-I(s)| (2)
其中c∈{2,3,4},δ∈{3,4},s=c+δ;where c∈{2,3,4}, δ∈{3,4}, s=c+δ;
接着获取颜色特征图,首先对图像提取四个颜色通道红色、绿色、蓝色、黄色上的分量,如式(3)~(6)所示:Then obtain the color feature map, first extract the components of the four color channels red, green, blue, and yellow from the image, as shown in formulas (3) to (6):
红色R=r-(g+b)/2 (3)Red R=r-(g+b)/2 (3)
绿色G=g-(r+b)/2 (4)Green G=g-(r+b)/2 (4)
蓝色B=b-(r+g)/2 (5)Blue B=b-(r+g)/2 (5)
黄色Y=(r+g)/2-|r-g|/2-b (6)Yellow Y=(r+g)/2-|r-g|/2-b (6)
通过其差分获得12幅中间-外围结构颜色图,如式(7)~(8)所示:12 mid-peripheral structural color maps are obtained through their differences, as shown in formulas (7)-(8):
RG(c,s)=|(R(c)-G(c))-(G(s)-R(s))| (7)RG(c,s)=|(R(c)-G(c))-(G(s)-R(s))| (7)
BY(c,s)=|(B(c)-Y(c))-(Y(s)-B(s))|; (8)BY(c,s)=|(B(c)-Y(c))-(Y(s)-B(s))|; (8)
然后获取方向特征图,亮度金字塔图像I(x)与常用的Gabor方向滤波器进行卷积,可以获得图像的方向,通过差分,获得24幅中间-外围结构方向图,如式(9)所示:Then obtain the direction feature map, the brightness pyramid image I(x) is convolved with the commonly used Gabor direction filter, the direction of the image can be obtained, and 24 middle-peripheral structure direction maps are obtained through the difference, as shown in formula (9) :
O(c,s,θ)=|O(c,θ)-O(s,θ)| (9)O(c,s,θ)=|O(c,θ)-O(s,θ)| (9)
其中θ={0°,45°,90°,135°}。where θ = {0°, 45°, 90°, 135°}.
至此,根据显著图方法由普通相机采集到的图片得到了6幅亮度特征图、 12幅颜色特征图和24幅方向特征图。So far, 6 brightness feature maps, 12 color feature maps and 24 direction feature maps have been obtained from the pictures collected by ordinary cameras according to the saliency map method.
在步骤S3中,分别对6幅亮度视差图、12幅颜色视差图和24幅方向视差图进行归一化处理,得到亮度显著图颜色显著图和方向显著图特征图的融合为显著性图提供了一个自下而上的输入,从而模拟成一个动态神经网络。In step S3, normalize the 6 luminance disparity maps, 12 color disparity maps and 24 directional disparity maps respectively to obtain the luminance saliency map color saliency map and directional saliency map The fusion of feature maps provides a bottom-up input to the saliency map, which can be modeled as a dynamic neural network.
由于中央-周边的差最能反映图像显著度的高低,在得到关注图后,会存在某种特征存在多处反差极大值的情况,这时就会出现大量显著峰。因此,在本发明的一个实施例中,在合并关注图生成显著图之前,对三组特征图分别进行归一化。通过归一化和跨尺度相加,特征图被整合成亮度、颜色和方向三个显著性图得到要想衡量图像中目标的显著性,需要综合三个通道的显著性图像。这里我们将综合后的三个通道的显著图再进行归一化。Since the difference between the center and the periphery can best reflect the saliency of the image, after obtaining the attention map, there will be a situation where there are multiple contrast maxima for a certain feature, and a large number of significant peaks will appear at this time. Therefore, in one embodiment of the present invention, the three sets of feature maps are normalized separately before merging the attention maps to generate the saliency map. Through normalization and cross-scale addition, the feature maps are integrated into three saliency maps of brightness, color and direction to obtain To measure the saliency of objects in an image, it is necessary to synthesize the saliency images of the three channels. Here we normalize the saliency maps of the integrated three channels.
在本发明的一个实施例中,基于至上而下的思想,提出了一种加权特征图融合算法,如式(10)所示:In one embodiment of the present invention, based on the top-down idea, a weighted feature map fusion algorithm is proposed, as shown in formula (10):
其中,wt+wc+wo=1,这样仍然使S的取值范围归一化到一定范围内。当预知图像某一通道的特征比较敏感时,可以自适应调整权值wt、wc和wo。本模型以基于局部对比的显著区域检测算法获取感兴趣区域,从显著图中选取最显著的点,以该点为显著点,在对应的特征显著图中采用区域生长的方式进行分割,得到感兴趣区域,得到的显著图区域更加具有针对性。Wherein, w t +w c +w o =1, so that the value range of S is still normalized to a certain range. When the feature of a certain channel of the predicted image is sensitive, the weights w t , w c and w o can be adjusted adaptively. This model uses a salient region detection algorithm based on local comparison to obtain the region of interest, selects the most salient point from the saliency map, takes this point as a salient point, and uses the region growing method to segment the corresponding feature saliency map to obtain the sense The region of interest, the obtained saliency map region is more targeted.
在本发明的一个实施例中,分别对6幅亮度视差图、12幅颜色视差图和24 幅方向视差图进行归一化处理的具体过程为:In one embodiment of the present invention, the specific process of normalizing the 6 brightness disparity maps, 12 color disparity maps and 24 direction disparity maps is as follows:
设置一个归一化算子N(.)提升图的质量,归一化算子N(.)计算流程如下:Set a normalization operator N(.) to improve the quality of the graph. The calculation process of the normalization operator N(.) is as follows:
把各通道特征图的像素值归一化到一个固定的区间[0,M]内,M为一正整数;Normalize the pixel values of each channel feature map to a fixed interval [0, M], where M is a positive integer;
找到图中全局最大值M的位置,计算其它所有特征图局部最大值的均值 Find the position of the global maximum M in the graph, and calculate the mean value of the local maximum of all other feature maps
特征图全局乘以 The feature map is globally multiplied by
通过归一化算子N(.)和跨尺度相加,特征图被整合成颜色、亮度和方向三个显著性图;Through the normalization operator N(.) and cross-scale addition, the feature map is integrated into three saliency maps of color, brightness and direction;
颜色归一化特征图: Color normalized feature map:
亮度归一化特征图: Brightness normalized feature map:
方向归一化特征图: Orientation normalized feature map:
其中,是在不同的尺度层上对每一特征的特征映射图进行降采样,而得到最高的主尺度层,再进行加法运算,得到颜色、亮度、方向三个特征上的显著图。in, The feature map of each feature is down-sampled on different scale layers to obtain the highest main scale layer, and then the addition operation is performed to obtain the saliency map on the three features of color, brightness, and direction.
在步骤S4中,从显著图中选取最显著的点,以该点为显著点,在对应的特征显著图中采用区域生长的方式进行分割,得到感兴趣区域。In step S4, the most salient point is selected from the saliency map, and this point is taken as a salient point, and the corresponding feature saliency map is segmented by region growing to obtain the region of interest.
在本发明的一个实施例中,为了增强匹配的稳定性,除了使用主方向之外,还可以选择辅方向。辅方向定义为:在直方图中,当某一个方向的值大于或者等于主峰值的80%时,则把这个方向做为关键点的辅方向。一个关键点一般会具有一个主方向以及多个辅方向。In an embodiment of the present invention, in order to enhance the stability of matching, in addition to using the main direction, an auxiliary direction may also be selected. The auxiliary direction is defined as: in the histogram, when the value of a certain direction is greater than or equal to 80% of the main peak value, then this direction is taken as the auxiliary direction of the key point. A key point generally has a main direction and multiple auxiliary directions.
在步骤S5中,获取γ相机采集的含有辐射强度分布信息的图像。In step S5, the image collected by the gamma camera and containing the radiation intensity distribution information is acquired.
在步骤S6中,分别提取γ相机图像和普通相机图像感兴趣区域的关键点。In step S6, the key points of the region of interest of the gamma camera image and the ordinary camera image are extracted respectively.
在本发明的一个实施例中,二维灰度图像(如普通相机的感兴趣区域灰度图像以及经伽马相机得到的混合图),在不同尺度下的尺度空间的表示可由图像与高斯核卷积得到,如式(14)所示:In one embodiment of the present invention, for a two-dimensional grayscale image (such as the grayscale image of the region of interest of an ordinary camera and the mixed image obtained by a gamma camera), the representation of the scale space at different scales can be represented by the image and the Gaussian kernel Convolution is obtained, as shown in formula (14):
L(x,y,σ)=G(x,y,σ)*I(x,y) (14)L(x,y,σ)=G(x,y,σ)*I(x,y) (14)
式中,G(x,y,σ)是可变尺度高斯函数,如式(15)所示:In the formula, G(x, y, σ) is a variable-scale Gaussian function, as shown in formula (15):
其中x,y为图像的横纵坐标,σ表示可变尺度。Among them, x and y are the horizontal and vertical coordinates of the image, and σ represents variable scale.
尺度规范化的拉普拉斯函数σ2▽2G具有尺度不变性,产生最稳定的图像特征。尺度归一化的高斯拉普拉斯算子如式(16)所示:The scale-normalized Laplacian function σ 2 ▽ 2 G is scale-invariant and produces the most stable image features. The scale-normalized Laplacian of Gaussian is shown in formula (16):
令DOG(x,y,σ)=G(x,y,kσ)-G(x,y,σ),则Let DOG(x,y,σ)=G(x,y,kσ)-G(x,y,σ), then
DOG(x,y,σ)=G(x,y,kσ)-G(x,y,σ)≈(k-1)σ2▽2GDOG(x,y,σ)=G(x,y,kσ)-G(x,y,σ)≈(k-1)σ 2 ▽ 2 G
方程的左边为高斯差分算子(DOG),比例因子(k-1)并不影响极值点的位置,因此高斯差分算子近似于尺度归一化的拉普拉斯差分算子。The left side of the equation is the Gaussian difference operator (DOG), and the scaling factor (k-1) does not affect the position of the extreme point, so the Gaussian difference operator is similar to the scale-normalized Laplacian difference operator.
在本发明的一个实施例中,利用不同尺度的高斯差分算子与图像进行卷积,如式(17)所示:In one embodiment of the present invention, Gaussian difference operators of different scales are used to convolve the image, as shown in formula (17):
D(x,y,σ)=[G(x,y,kσ)-G(x,y,σ)]*I(x,y)=L(x,y,kσ)-L(x,y,σ) (17)D(x,y,σ)=[G(x,y,kσ)-G(x,y,σ)]*I(x,y)=L(x,y,kσ)-L(x,y ,σ) (17)
图2所示为本发明的DOG差分金字塔形成过程的示意图,图左边部分为不同尺度下获得的高斯金字塔图像,右边是相邻尺度差分得到的差分金字塔图像。Fig. 2 is a schematic diagram of the formation process of the DOG differential pyramid of the present invention, the left part of the figure is a Gaussian pyramid image obtained at different scales, and the right is a differential pyramid image obtained by difference between adjacent scales.
由上可知求取DOG空间的局部极值点可得到稳定的图像特征,DOG函数的极值点可通过与它周围相邻的26个点进行比较,判断是否是局部极值点。It can be seen from the above that obtaining the local extremum point of the DOG space can obtain stable image features, and the extremum point of the DOG function can be compared with the 26 adjacent points around it to judge whether it is a local extremum point.
图3所示为本发明的DOG函数的极值点检测的示意图。中间的被检测点和它同尺度的8个相邻点和上下相邻尺度对应的9×2个点共26个点比较,以确保在尺度空间和二维图像空间都检测到极值点。FIG. 3 is a schematic diagram of extreme point detection of the DOG function of the present invention. The detected point in the middle is compared with its 8 adjacent points of the same scale and 9×2 points corresponding to the upper and lower adjacent scales, a total of 26 points, to ensure that extreme points are detected in both the scale space and the two-dimensional image space.
采用对空间尺度函数求导取极值的方法,并设定阈值 消除一些对比度低和不稳定的点,从而确定极值点的位置。The method of deriving the extreme value of the spatial scale function is adopted, and the threshold is set to eliminate some low-contrast and unstable points, so as to determine the position of the extreme point.
利用DOG函数在尺度空间Taylor展开式,如式(18)所示:Use the DOG function to expand Taylor in the scale space, as shown in formula (18):
其中,X=(x,y,σ)T为上一步中检测到的极值点坐标。Among them, X=(x,y,σ) T is the coordinates of extreme points detected in the previous step.
对上式求导并令其为零,得到关键点的位置坐标: Take the derivative of the above formula and make it zero to get the position coordinates of the key points:
并将其带入泰勒展开式得: and put it into the Taylor expansion:
通过设定的阈值,消除小于阈值的点。set by Threshold, eliminate points smaller than the threshold.
DOG函数在图像边缘有较强的边缘响应,因此还需要排除边缘响应。可以通过计算该点所在位置尺度周围3x3窗口内的Hessian矩阵排除边缘响应,其计算如式(19)所示:The DOG function has a strong edge response at the edge of the image, so it also needs to exclude the edge response. The edge response can be excluded by calculating the Hessian matrix in the 3×3 window around the scale of the point’s location, and its calculation is shown in formula (19):
令α为最大特征值,β为最小的特征值,则α=rβ,如式(20)~(22)所示:Let α be the largest eigenvalue, and β be the smallest eigenvalue, then α=rβ, as shown in formulas (20)~(22):
Det(H)=DxxDyy-(Dxy)2=αβ (21)Det(H)=D xx D yy -(D xy ) 2 =αβ (21)
Tr(H)=Dxx+Dyy=α+β (22)Tr(H)= Dxx + Dyy =α+β (22)
(r+1)2/r在两特征值相等时达最小,随r的增长而增长。因此只需要在设定r后限定,如式(23)所示:(r+1) 2 /r reaches the minimum when the two eigenvalues are equal, and increases with the increase of r. Therefore, it only needs to be limited after setting r, as shown in formula (23):
在步骤S7中,将所述关键点分别生成特征向量。In step S7, the key points are respectively generated into feature vectors.
在本发明的一个实施例中,特征向量的生成过程如下:In one embodiment of the present invention, the generation process of feature vector is as follows:
1)首先确定提取关键点图像的变换参数;1) First determine the transformation parameters for extracting key point images;
2)将关键点的坐标移至主方向;2) Move the coordinates of key points to the main direction;
3)在以关键点为中心16*16的区域内,对每个以4*4的区域内计算8方向的梯度直方图,统计每个梯度的累积值,形成一个种子点,共生成16种子点, 128维向量;3) In the 16*16 area centered on the key point, calculate the gradient histogram in 8 directions for each 4*4 area, count the cumulative value of each gradient, form a seed point, and generate 16 seeds in total point, 128-dimensional vector;
4)对得到的特征向量进行阈值化和向量归一化,归一化后的特征向量如下:4) Thresholding and vector normalization are performed on the obtained feature vectors, and the normalized feature vectors are as follows:
L=(l1,l2,...,l128)L=(l 1 ,l 2 ,...,l 128 )
在本发明的一个实施例中,利用所述关键点邻域像素的梯度方向分布特性,为每个关键点指定方向,关键点描述子相对于此方向表征,从而使关键点描述子对图像旋转具有不变性。In one embodiment of the present invention, using the gradient direction distribution characteristics of the pixels in the neighborhood of the key point, a direction is specified for each key point, and the key point descriptor is characterized relative to this direction, so that the key point descriptor rotates the image is immutable.
在本发明的一个实施例中,利用梯度数学模型来为关键点指定方向,梯度数学模型 如式(24)所示:In one embodiment of the present invention, utilize gradient mathematical model to specify direction for key point, gradient mathematical model is as shown in formula (24):
梯度的幅值如式(25)所示:The magnitude of the gradient is shown in formula (25):
梯度的方向如式(26)所示:The direction of the gradient is shown in formula (26):
在本发明的一个实施例中,在以关键点为中心的邻域内进行采样,并使用直方图方法统计邻域像素的方向。梯度直方图统计后的方向范围是0度到360 度,把每10度做为一个柱进行分析,总共包括36个柱。直方图峰值代表了关键点处梯度的主方向,即把它做为关键点的方向。In one embodiment of the present invention, sampling is performed in a neighborhood centered on the key point, and the direction of the neighborhood pixels is counted using a histogram method. The direction range of the gradient histogram after statistics is 0° to 360°, and every 10° is regarded as a column for analysis, including 36 columns in total. The histogram peak represents the main direction of the gradient at the key point, which is taken as the direction of the key point.
至此,图像的关键点已检测完毕,每个关键点有三个信息:位置、尺度、方向。So far, the key points of the image have been detected, and each key point has three information: position, scale, and direction.
在步骤S8中,将普通相机图像感兴趣区域关键点的特征向量与混合图像关键点的特征向量进行匹配,如果符合匹配条件,则目标为显著性目标。In step S8, match the eigenvectors of the key points of the region of interest in the ordinary camera image with the eigenvectors of the key points of the mixed image, and if the matching conditions are met, the target is a salient target.
普通相机的感兴趣区域灰度图像以及经伽马相机得到的混合图的特征点已分别被表征为特征向量,因此感兴趣区域灰度图像以及经伽马相机得到的混合图特征点的匹配可以通过两个特征向量的相似度来判断。The grayscale image of the region of interest of the ordinary camera and the feature points of the mixed image obtained by the gamma camera have been characterized as feature vectors, so the matching of the gray image of the interest region and the feature points of the mixed image obtained by the gamma camera can be It is judged by the similarity of two feature vectors.
在本发明的一个实施例中,分别对感兴趣区域灰度图像以及经伽马相机得到的混合图建立关键点描述子集合。目标的识别是通过两点集内关键点描述子的比对来完成。具有128维的关键点描述子的相似性度量采用欧式距离。In one embodiment of the present invention, key point description subsets are respectively established for the grayscale image of the region of interest and the mixed image obtained through the gamma camera. Target recognition is accomplished by comparing key point descriptors in two point sets. The similarity measure of keypoint descriptors with 128 dimensions adopts Euclidean distance.
模板图中关键点描述子,Ri=(ri1,ri2,...,ri128)Descriptor of key points in the template graph, R i =(r i1 ,r i2 ,...,r i128 )
实时图中关键点描述子,Si=(si1,si2,...,si128)Descriptor of key points in the real-time graph, S i =(s i1 ,s i2 ,...,s i128 )
任意两个描述子相似性定义如式(27)所示:The definition of similarity between any two descriptors is shown in formula (27):
要得到配对的特征点描述子,需满足:To get the paired feature point descriptor, it needs to meet:
当Rj是Si匹配点,(Rj是实时图中距离Si最近的点,Rp是实时图中距离Si次最近的点)When R j is the matching point of S i , (R j is the closest point to S i in the real-time graph, and R p is the closest point to S i in the real-time graph)
至此,已经完成图像匹配的所有工作。通过以上图像匹配算法,找出γ图像和普通相机图像中的匹配点云,然后计算得到它们之间的变换矩阵,最后将污染源映射到普通相机的图像中,实现污染源目标的检测。So far, all the work of image matching has been completed. Through the above image matching algorithm, the matching point cloud in the γ image and the ordinary camera image is found, and then the transformation matrix between them is calculated, and finally the pollution source is mapped to the image of the ordinary camera to realize the detection of the pollution source target.
虽然结合具体实施例对本发明的具体实施方式进行了详细地描述,但并非是对本专利保护范围的限定。在权利要求书所限定的范围内,本领域的技术人员不经创造性劳动即可做出的各种修改或调整仍受本专利的保护。Although specific embodiments of the present invention have been described in detail in conjunction with specific examples, they are not intended to limit the protection scope of this patent. Within the scope defined in the claims, various modifications or adjustments that can be made by those skilled in the art without creative work are still protected by this patent.
Claims (7)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510272424.8A CN104835175B (en) | 2015-05-26 | 2015-05-26 | Object detection method in a kind of nuclear environment of view-based access control model attention mechanism |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510272424.8A CN104835175B (en) | 2015-05-26 | 2015-05-26 | Object detection method in a kind of nuclear environment of view-based access control model attention mechanism |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104835175A CN104835175A (en) | 2015-08-12 |
CN104835175B true CN104835175B (en) | 2019-11-05 |
Family
ID=53813040
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510272424.8A Active CN104835175B (en) | 2015-05-26 | 2015-05-26 | Object detection method in a kind of nuclear environment of view-based access control model attention mechanism |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104835175B (en) |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106127749A (en) * | 2016-06-16 | 2016-11-16 | 华南理工大学 | The target part recognition methods of view-based access control model attention mechanism |
CN106780476A (en) * | 2016-12-29 | 2017-05-31 | 杭州电子科技大学 | A kind of stereo-picture conspicuousness detection method based on human-eye stereoscopic vision characteristic |
CN107507225B (en) * | 2017-09-05 | 2020-10-27 | 明见(厦门)技术有限公司 | Moving object detection method, device, medium and computing equipment |
CN107590829B (en) * | 2017-09-18 | 2020-06-30 | 西安电子科技大学 | Seed point picking method suitable for multi-view dense point cloud data registration |
CN108154147A (en) * | 2018-01-15 | 2018-06-12 | 中国人民解放军陆军装甲兵学院 | The region of interest area detecting method of view-based access control model attention model |
CN108564088A (en) * | 2018-04-17 | 2018-09-21 | 广东工业大学 | Licence plate recognition method, device, equipment and readable storage medium storing program for executing |
CN108681753B (en) * | 2018-05-29 | 2021-03-30 | 武汉环宇智行科技有限公司 | Image stereo matching method and system based on semantic segmentation and neural network |
CN109785359B (en) * | 2018-11-27 | 2020-12-04 | 北京理工大学 | A Video Object Detection Method Based on Deep Feature Pyramid and Tracking Loss |
CN109961430B (en) * | 2018-12-13 | 2023-09-01 | 长春理工大学 | A method for extracting corroded areas on the surface of metalware |
CN111079556B (en) * | 2019-11-25 | 2023-08-15 | 航天时代飞鸿技术有限公司 | Multi-temporal unmanned aerial vehicle video image change region detection and classification method |
CN111709428B (en) | 2020-05-29 | 2023-09-15 | 北京百度网讯科技有限公司 | Method and device for identifying positions of key points in image, electronic equipment and medium |
CN111797832B (en) * | 2020-07-14 | 2024-02-02 | 成都数之联科技股份有限公司 | Automatic generation method and system for image region of interest and image processing method |
US11977319B2 (en) * | 2020-09-25 | 2024-05-07 | Qualcomm Incorporated | Saliency based capture or image processing |
CN114764801B (en) * | 2022-05-19 | 2024-12-13 | 中国科学院微电子研究所 | Weak small ship target fusion detection method and device based on multi-visual salient features |
CN115601366B (en) * | 2022-12-15 | 2023-04-14 | 中科海拓(无锡)科技有限公司 | Vehicle bottom bolt looseness detection algorithm |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103226826A (en) * | 2013-03-20 | 2013-07-31 | 西安电子科技大学 | Method for detecting changes of remote sensing image of visual attention model based on local entropy |
JP5282614B2 (en) * | 2009-03-13 | 2013-09-04 | オムロン株式会社 | Model data registration method and visual sensor for visual recognition processing |
-
2015
- 2015-05-26 CN CN201510272424.8A patent/CN104835175B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5282614B2 (en) * | 2009-03-13 | 2013-09-04 | オムロン株式会社 | Model data registration method and visual sensor for visual recognition processing |
CN103226826A (en) * | 2013-03-20 | 2013-07-31 | 西安电子科技大学 | Method for detecting changes of remote sensing image of visual attention model based on local entropy |
Non-Patent Citations (2)
Title |
---|
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis;Laurent Ltti等;《IEEE Transactions on Pattern Analysis and Machine Intelligence》;19981130;第20卷(第11期);1254-1259 * |
基于注意力的目标识别算法及在移动机器人的应用研究;张建兴;《中国优秀硕士论文全文数据库》;20140315(第03期);摘要,正文第19-48页 * |
Also Published As
Publication number | Publication date |
---|---|
CN104835175A (en) | 2015-08-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104835175B (en) | Object detection method in a kind of nuclear environment of view-based access control model attention mechanism | |
Yan et al. | Unsupervised image saliency detection with Gestalt-laws guided optimization and visual attention based refinement | |
CN104850850B (en) | A kind of binocular stereo vision image characteristic extracting method of combination shape and color | |
WO2018086607A1 (en) | Target tracking method, electronic device, and storage medium | |
CN104766046B (en) | One kind is detected using traffic mark color and shape facility and recognition methods | |
CN103632132B (en) | Face detection and recognition method based on skin color segmentation and template matching | |
CN104966085B (en) | A kind of remote sensing images region of interest area detecting method based on the fusion of more notable features | |
CN110334762B (en) | Feature matching method based on quad tree combined with ORB and SIFT | |
Miao et al. | Interest point detection using rank order LoG filter | |
Yao et al. | A new pedestrian detection method based on combined HOG and LSS features | |
CN106250895B (en) | A method for detecting regions of interest in optical remote sensing images | |
Zhou et al. | Research on pedestrian detection technology based on the SVM classifier trained by HOG and LTP features | |
CN105678278A (en) | Scene recognition method based on single-hidden-layer neural network | |
Huang et al. | Saliency and co-saliency detection by low-rank multiscale fusion | |
CN107291855A (en) | A kind of image search method and system based on notable object | |
Ishikura et al. | Saliency detection based on multiscale extrema of local perceptual color differences | |
CN108710909B (en) | Counting method for deformable, rotary and invariant boxed objects | |
CN105701467A (en) | Many-people abnormal behavior identification method based on human body shape characteristic | |
CN103984920B (en) | Three-dimensional face identification method based on sparse representation and multiple feature points | |
CN108256462A (en) | A kind of demographic method in market monitor video | |
CN106022342A (en) | Image feature extraction method based on KAZE algorithm | |
Ecins et al. | Shadow free segmentation in still images using local density measure | |
Zhang et al. | Contour detection via stacking random forest learning | |
CN107610136B (en) | Salient object detection method based on center query point ranking of convex hull structure | |
CN106529548A (en) | Sub-pixel level multi-scale Harris corner detection algorithm |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
EXSB | Decision made by sipo to initiate substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right |
Effective date of registration: 20210223 Address after: 621000 No.35 Mianan Road, Fucheng District, Mianyang City, Sichuan Province Patentee after: Sichuan nuclear insurance Ruixiang Technology Co.,Ltd. Address before: 621000, No. 59, Qinglong Avenue, Fucheng District, Sichuan, Mianyang Patentee before: Southwest University of Science and Technology |
|
TR01 | Transfer of patent right |