[go: up one dir, main page]

CN104079872A - Content-based Video Image Processing and Human-Computer Interaction Method - Google Patents

Content-based Video Image Processing and Human-Computer Interaction Method Download PDF

Info

Publication number
CN104079872A
CN104079872A CN201410213437.3A CN201410213437A CN104079872A CN 104079872 A CN104079872 A CN 104079872A CN 201410213437 A CN201410213437 A CN 201410213437A CN 104079872 A CN104079872 A CN 104079872A
Authority
CN
China
Prior art keywords
target
scene
detection
fire
content
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201410213437.3A
Other languages
Chinese (zh)
Inventor
杨燕
胡小鹏
王凡
裴红
李明辉
崔雅敏
吕泽锋
贺艳花
王国强
陈仕林
魏强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Liaohe Petroleum Engineering Co ltd
Dalian University of Technology
China United Coalbed Methane Corp Ltd
Liaoning Normal University
Original Assignee
China Liaohe Petroleum Engineering Co ltd
Dalian University of Technology
China United Coalbed Methane Corp Ltd
Liaoning Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Liaohe Petroleum Engineering Co ltd, Dalian University of Technology, China United Coalbed Methane Corp Ltd, Liaoning Normal University filed Critical China Liaohe Petroleum Engineering Co ltd
Priority to CN201410213437.3A priority Critical patent/CN104079872A/en
Publication of CN104079872A publication Critical patent/CN104079872A/en
Pending legal-status Critical Current

Links

Landscapes

  • Image Analysis (AREA)

Abstract

本发明公开了基于内容的视频图像处理和人机交互方法,属于无线通信技术领域,本发明包括如下步骤:①对输入视频提取特征信息;②根据场景特征,建立场景模型,并在此基础上进行基于场景的运动物体检测;③根据目标内容,建立目标模型,并在此基础上进行基于目标的检测;④通过融合基于场景和基于目标的检测,分析检测目标显著度(重要等级),生成检测结果,其包括检测目标和重要等级;⑤通过使用计算机视觉注意力选择机制,将监控人员的视觉注意力引导到重要目标,从而有效地利用有限的图像显示空间。

The present invention discloses a content-based video image processing and human-computer interaction method, which belongs to the field of wireless communication technology. The present invention comprises the following steps: 1. extracting feature information from an input video; 2. establishing a scene model according to scene features, and performing scene-based moving object detection on this basis; 3. establishing a target model according to target content, and performing target-based detection on this basis; 4. analyzing the prominence (importance level) of the detected target by fusing scene-based and target-based detection, and generating a detection result, which includes the detected target and the importance level; 5. guiding the visual attention of monitoring personnel to important targets by using a computer vision attention selection mechanism, thereby effectively utilizing limited image display space.

Description

Content-based video image is processed and man-machine interaction method
Technical field
The present invention relates to content-based video image and process and man-machine interaction method, belong to wireless communication technology field.
Background technology
At present, China's coal bed gas is produced video monitoring still in exploitation preliminary stage, still has the problem of the following aspects:
(1) because coal bed gas field production belongs to field work, gas well quantity is many, distributional region is wide, landform more complicated, and road conditions is poor, is therefore difficult to too greatly by artificial patrol method workload the problem existing in discovery production in time.Artificial patrol method is subject to weather effect large, and running into sleety weather may can not arrive well site a few days.From the viewpoint of fail safe and following enterprise development high efficiency two, production process unattended operation, visual remote management become coal bed gas industry development inexorable trend.
(2) under the conditions of work such as the numerous and surrounding environment complexity of well head, traditional inter-frame difference video data handling procedure and traditional " windows display " man-machine interaction mode can not meet the demand to well head long-time continuous video monitoring.How studying a set of with low cost, dependable performance, be applicable to technology and the equipment of video image acquisition, transmission, storage, processing and the man-machine interaction of coal bed gas field exploitation, is urgent need to solve the problem.
Summary of the invention
The present invention is directed to the proposition of above problem, process and man-machine interaction method and develop content-based video image.
The present invention includes following steps:
1. to input video characteristic information extraction;
2. according to scene characteristic, set up model of place, and carry out on this basis detecting based on the moving object of scene;
3. according to object content, set up object module, and carry out on this basis the detection of based target;
4. by merging the detection based on scene and based target, analyzing and testing target significance (important level), generates and comprises the testing result that detects target and important level;
5. by using computer vision attention selection mechanism, monitor staff's visual attention is directed to important goal, thereby effectively utilizes limited image display space.
The principle of the invention and beneficial effect: in complex scene environment, numerous coal bed gas well heads are carried out to video monitoring in the wild, will produce massive video data.How under massive video data condition, effectively to realize man-machine interaction, be a major issue of Large Scale Video Monitoring System.As previously mentioned, coal bed gas production video monitoring system has the feature of many monitoring points, long time continuous working.Under this condition of work, monitor staff cannot focus one's attention on for a long time, the each control point of point by point scanning, thereby monitor staff's attentiveness must be attracted on abnormal picture.Research shows, human perception system mainly utilizes visual attention to select mechanism, overcomes multiple dimensioned in complicated visual scene perception and high complexity issue.Visual attention selects mechanism in computer vision system, to be paid attention to, and is applied to engineering practice as design theory.For example, in European opportunity of combat design, visual attention selection mechanism is used to refer to leads design man-machine interactive system.Fig. 1 .1 institute representation model is the application of visual attention selector built in moving object detection aspect.In this project, we will follow and utilize human attention to select mechanism, realize the efficient man-machine interaction of coal bed gas production video monitoring system.
Brief description of the drawings
Fig. 1 flow chart of the present invention.
Embodiment
1, content-based video image processing and man-machine interaction method comprises the steps:
1. to input video characteristic information extraction;
By the method for statistics with histogram, extract condition of a fire RGB triple channel color characteristic, extract the scene point-of-interest feature based on SURF;
2. according to scene characteristic, set up model of place, and carry out on this basis detecting based on the moving object of scene;
For the color characteristic (referring to step 3.) of the coal bed gas production testing target condition of a fire, extract the area-of-interest of video scene and condition of a fire color similarity, in this region, extract SURF characteristic point, according to Euler's distance the matching characteristic point of 5 × 5 neighborhoods of R passage calculated for pixel values consecutive frame SURF characteristic point, calculate the alternate position spike △ d of the corresponding SURF characteristic point of consecutive frame, the data symbol that record position is poor, utilize character check method (seeing formula (1)), if judge whether the poor variation of SURF characteristic point position meets Gaussian Profile and meet Gaussian Profile (seeing formula (2)), judge that current SURF characteristic point is as scene characteristic point, record its positional information, set up the sparse model of place based on characteristic point, otherwise, be moving object SURF characteristic point.
f ( &Delta;d ) = + 1 &Delta;d > 0 - 1 &Delta;d < 0 0 &Delta;d = 0 - - - ( 1 )
E ( f ( &Delta;d ) ) = | min ( p ( &Delta;d > 0 ) , p ( &Delta;d < 0 ) ) max ( p ( &Delta;d > 0 ) , p ( &Delta;d < 0 ) ) - 1 | < &epsiv; - - - ( 2 )
If the scene characteristic point of optimum Match detected in present frame, record according to the new feature dot position information mating with scene characteristic point in present frame the historical position message queue that scene characteristic point changed with the certain hour cycle, the renewal of positional information queue is carried out according to the principle of first in first out.Matching characteristic point in mark present frame is scene characteristic point.For the scene characteristic point that in model of place, other is not mated, carry out a character check, upgrade model of place, mark scene characteristic point and foreground features point; If the characteristic point of mating with existing scene characteristic point do not detected in present frame, the not new feature point of coupling detecting in present frame is carried out to a character check again, if new scene characteristic point detected, upgrade model of place.Otherwise, be labeled as foreground features point.
3. according to object content, set up object module, and carry out on this basis the detection of based target;
In the training stage, the region that manually produces the condition of a fire in segmented video image, utilizes the method for statistics with histogram to extract the three-channel color characteristic of condition of a fire RGB, determines and detects the color threshold threshold=200 that the condition of a fire occurs; Be partitioned into the region at condition of a fire place in training video image according to this color threshold, in this region, extract the SURF feature of the condition of a fire, calculate the center-of-mass coordinate (seeing formula (3) and formula (4)) of all SURF characteristic points in current time t video image, and the alternate position spike of consecutive frame SURF characteristic point barycenter (is shown in formula (5) and formula
(6)), these centroid positions are changed and carry out probability statistics, set up the motion model of the condition of a fire, meet E ( f ( &Delta; x t 2 - t 1 ) ) , E ( f ( &Delta; y t 2 - t 1 ) ) Condition;
x centriod t = 1 n ( &Sigma; i = 1 n x i t ) - - - ( 3 )
y centriod t = 1 n ( &Sigma; i = 1 n y i t ) - - - ( 4 )
&Delta; x t 2 - t 1 = x centriod t 2 - x centriod t 1 - - - ( 5 )
&Delta; y t 2 - t 1 = y centriod t 2 - y centriod t 1 - - - ( 6 )
At detection-phase, extract moving object SURF characteristic point according to model of place, record the poor data symbol of adjacent video scene motion object SURF characteristic point centroid position, utilize character check method, judge whether the poor variation of SURF characteristic point centroid position meets Gaussian Profile.If meet Gaussian Profile, judge that the condition of a fire occurs, otherwise the condition of a fire does not occur.
In sum, the judgment basis that the condition of a fire detects is as follows:
I(R)>threshold?∩I(G)>threshold
&cap; I ( B ) > threshold &cap; E ( f ( &Delta; x t 2 - t 1 ) ) < &epsiv; &cap; E ( f ( &Delta; y t 2 - t 1 ) ) < &epsiv;
Wherein R, G, B represents respectively the three-channel pixel value of red, green, blue.
4. by merging the detection based on scene and based target, analyzing and testing target significance, generates and comprises the testing result that detects target and important level;
On the basis of detecting in scene moving object, scene moving object SURF characteristic point centroid position change profile P in statistics certain time interval T back, calculate the condition of a fire distribution of movement P that itself and Target Modeling training stage set up firekL distance, determine and detect target possibility occurrence and define three important level according to KL distance, see formula (7).
f ( D KL ( P back | | P fire ) ) = 1 0 < D KL < &epsiv; 1 2 &epsiv; 1 &le; D KL < &epsiv; 2 3 D KL &GreaterEqual; &epsiv; 3 - - - ( 7 )
5. by using computer vision attention selection mechanism, monitor staff's visual attention is directed to important goal, thereby effectively utilizes limited image display space.
According to detecting target and important level thereof, the preferential high key video sequence information of display level, is directed to important goal by monitor staff's visual attention, thereby effectively utilizes limited image display space.
The english term the present invention relates to is:
SURF (Speeded-Up Robust Features): accelerate robust feature
KL (Kul lback-Leibler divergence): relative entropy distance
The above; it is only preferably embodiment of the present invention; but protection scope of the present invention is not limited to this; any be familiar with those skilled in the art the present invention disclose technical scope in; be equal to replacement or changed according to technical scheme of the present invention and inventive concept thereof, within all should being encompassed in protection scope of the present invention.

Claims (5)

1.基于内容的视频图像处理和人机交互方法,其特征存在于:包括如下步骤:1. A content-based video image processing and human-computer interaction method, characterized in that it comprises the steps of: ①对输入视频提取特征信息;① Extract feature information from the input video; ②根据场景特征,建立场景模型,并在此基础上进行基于场景的运动物体检测;②Establish a scene model according to the scene characteristics, and perform scene-based moving object detection on this basis; ③根据目标内容,建立目标模型,并在此基础上进行基于目标的检测;③According to the target content, establish a target model, and on this basis, perform target-based detection; ④通过融合基于场景和基于目标的检测,分析检测目标显著度,生成检测结果,生成检测结果包括检测目标和重要等级;④By integrating scene-based and target-based detection, analyze the saliency of the detection target, generate detection results, and generate detection results including detection targets and importance levels; ⑤通过使用计算机视觉注意力选择机制,将监控人员的视觉注意力引导到重要目标,从而有效地利用有限的图像显示空间。⑤ By using the computer vision attention selection mechanism, the monitor's visual attention is directed to important targets, thereby effectively utilizing the limited image display space. 2.根据权利要求1所述的基于内容的视频图像处理和人机交互方法,其特征存在于:根据场景特征,建立场景模型,并在此基础上进行基于场景的运动物体检测方法为:针对煤层气生产检测目标火情的颜色特征,提取视频场景与火情颜色相似的感兴趣区域,在此区域内提取SURF特征点,计算视频图像相邻帧对应SURF特征点的位置差,记录位置差的数据符号,利用符号检验方法,判定SURF特征点位置差的变化是否满足高斯分布;如果满足高斯分布,判定当前SURF特征点为场景特征点,否则,为运动物体SURF特征点。2. content-based video image processing and human-computer interaction method according to claim 1, it is characterized in that: according to scene feature, set up scene model, and carry out scene-based moving object detection method on this basis as: for Coalbed methane production detects the color characteristics of the target fire, extracts the region of interest where the video scene is similar to the fire color, extracts SURF feature points in this area, calculates the position difference corresponding to the SURF feature point in adjacent frames of the video image, and records the position difference Use the sign test method to determine whether the change of the position difference of the SURF feature point satisfies the Gaussian distribution; if it satisfies the Gaussian distribution, it is determined that the current SURF feature point is a scene feature point, otherwise, it is a moving object SURF feature point. 3.根据权利要求1所述的基于内容的视频图像处理和人机交互方法,其特征存在于:根据目标内容,建立目标模型,并在此基础上进行基于目标的检测方法为:在训练阶段,针对煤层气生产检测目标火情的颜色特征,提取火情红,绿,蓝三通道的颜色特征,确定检测火情发生的颜色阈值;根据该颜色阈值分割出训练视频图像中火情所在的区域,在此区域内,提取火情的SURF特征,计算当前帧视频图像内所有SURF特征点的质心位置,及相邻帧SURF特征点质心的位置差,对这些质心位置差变化进行概率统计,建立火情的运动模型;在检测阶段,根据场景模型提取运动物体SURF特征点,记录相邻视频场景运动物体SURF特征点质心位置差的数据符号,利用符号检验方法,判定SURF特征点质心位置差的变化是否满足高斯分布,如果满足高斯分布,判定火情发生,否则,火情未发生。3. content-based video image processing and human-computer interaction method according to claim 1, it is characterized in that: according to target content, set up target model, and on this basis, carry out target-based detection method as: in the training stage , aiming at the color features of the fire detection target in coalbed methane production, extract the color features of the red, green and blue channels of the fire, and determine the color threshold for detecting the occurrence of the fire; segment the fire location in the training video image according to the color threshold In this area, the SURF feature of the fire is extracted, the centroid position of all SURF feature points in the current frame video image is calculated, and the position difference of the SURF feature point centroid of the adjacent frame is calculated, and the probability statistics of the centroid position difference changes are carried out. Establish the motion model of the fire; in the detection stage, extract the SURF feature points of the moving object according to the scene model, record the data symbols of the centroid position difference of the SURF feature point of the moving object in the adjacent video scene, and use the sign inspection method to determine the centroid position difference of the SURF feature point Whether the change of satisfies the Gaussian distribution, if it satisfies the Gaussian distribution, it is determined that the fire has occurred, otherwise, the fire has not occurred. 4.根据权利要求1所述的基于内容的视频图像处理和人机交互方法,其特征存在于:通过融合基于场景和基于目标的检测,分析检测目标显著度,生成检测结果方法为:在场景运动物体检测的基础上,统计一定时间内场景运动物体SURF特征点质心位置变化分布,利用KL方法计算其与目标建模训练阶段建立的火情运动分布的相似度,根据相似度确定检测目标发生可能性并定义重要等级。4. content-based video image processing and human-computer interaction method according to claim 1, it is characterized in that: by fusing scene-based and target-based detection, analyzing the detection target salience, the method for generating detection results is: in the scene Based on the detection of moving objects, the position change distribution of SURF feature points of scene moving objects within a certain period of time is counted, and the KL method is used to calculate the similarity between it and the fire motion distribution established in the target modeling training stage, and the occurrence of the detection target is determined according to the similarity likelihood and define levels of importance. 5.根据权利要求1所述的基于内容的视频图像处理和人机交互方法,其特征存在于:通过使用计算机视觉注意力选择机制,将监控人员的视觉注意力引导到重要目标,从而有效地利用有限的图像显示空间方法为:根据检测目标及其重要等级,优先显示等级高的关键视频信息,将监控人员的视觉注意力引导到重要目标,从而有效地利用有限的图像显示空间,所述生成检测结果包括检测目标和重要等级。5. content-based video image processing and human-computer interaction method according to claim 1, it is characterized in that: by using computer vision attention selection mechanism, the visual attention of monitoring personnel is guided to important target, thereby effectively The method of using limited image display space is: according to the detection target and its importance level, the key video information with high level is displayed preferentially, and the visual attention of the monitoring personnel is directed to the important target, so as to effectively use the limited image display space. The generated detection results include detection targets and importance levels.
CN201410213437.3A 2014-05-16 2014-05-16 Content-based Video Image Processing and Human-Computer Interaction Method Pending CN104079872A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410213437.3A CN104079872A (en) 2014-05-16 2014-05-16 Content-based Video Image Processing and Human-Computer Interaction Method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410213437.3A CN104079872A (en) 2014-05-16 2014-05-16 Content-based Video Image Processing and Human-Computer Interaction Method

Publications (1)

Publication Number Publication Date
CN104079872A true CN104079872A (en) 2014-10-01

Family

ID=51600885

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410213437.3A Pending CN104079872A (en) 2014-05-16 2014-05-16 Content-based Video Image Processing and Human-Computer Interaction Method

Country Status (1)

Country Link
CN (1) CN104079872A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104598895A (en) * 2015-02-10 2015-05-06 天津艾思科尔科技有限公司 Method and device for flame detection based on video image analysis
CN107545109A (en) * 2017-08-31 2018-01-05 中国石油大学(北京) Coal bed gas field acquisition system optimization method
CN110136104A (en) * 2019-04-25 2019-08-16 上海交通大学 Image processing method, system and medium based on UAV ground station
CN110501914A (en) * 2018-05-18 2019-11-26 佛山市顺德区美的电热电器制造有限公司 A kind of method for safety monitoring, equipment and computer readable storage medium
CN110688292A (en) * 2018-07-04 2020-01-14 葛建新 Software testing visual detection method based on attention mechanism

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040131254A1 (en) * 2000-11-24 2004-07-08 Yiqing Liang System and method for object identification and behavior characterization using video analysis
WO2006082429A1 (en) * 2005-02-04 2006-08-10 British Telecommunications Public Limited Company Classifying an object in a video frame
CN101017573A (en) * 2007-02-09 2007-08-15 南京大学 Method for detecting and identifying moving target based on video monitoring
CN101751677A (en) * 2008-12-17 2010-06-23 中国科学院自动化研究所 Target continuous tracking method based on multi-camera

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040131254A1 (en) * 2000-11-24 2004-07-08 Yiqing Liang System and method for object identification and behavior characterization using video analysis
WO2006082429A1 (en) * 2005-02-04 2006-08-10 British Telecommunications Public Limited Company Classifying an object in a video frame
CN101017573A (en) * 2007-02-09 2007-08-15 南京大学 Method for detecting and identifying moving target based on video monitoring
CN101751677A (en) * 2008-12-17 2010-06-23 中国科学院自动化研究所 Target continuous tracking method based on multi-camera

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104598895A (en) * 2015-02-10 2015-05-06 天津艾思科尔科技有限公司 Method and device for flame detection based on video image analysis
CN104598895B (en) * 2015-02-10 2018-03-16 天津艾思科尔科技有限公司 A kind of flame detecting method and device based on video image analysis
CN107545109A (en) * 2017-08-31 2018-01-05 中国石油大学(北京) Coal bed gas field acquisition system optimization method
CN107545109B (en) * 2017-08-31 2020-06-05 中国石油大学(北京) Optimization method of coal bed gas field acquisition system
CN110501914A (en) * 2018-05-18 2019-11-26 佛山市顺德区美的电热电器制造有限公司 A kind of method for safety monitoring, equipment and computer readable storage medium
CN110501914B (en) * 2018-05-18 2023-08-11 佛山市顺德区美的电热电器制造有限公司 Security monitoring method, equipment and computer readable storage medium
CN110688292A (en) * 2018-07-04 2020-01-14 葛建新 Software testing visual detection method based on attention mechanism
CN110688292B (en) * 2018-07-04 2023-11-24 葛建新 Software testing visual detection method based on attention mechanism
CN110136104A (en) * 2019-04-25 2019-08-16 上海交通大学 Image processing method, system and medium based on UAV ground station

Similar Documents

Publication Publication Date Title
CN104079872A (en) Content-based Video Image Processing and Human-Computer Interaction Method
CN112149761B (en) Electric power intelligent construction site violation detection method based on YOLOv4 improved algorithm
CN102236947B (en) Flame monitoring method and system based on video camera
Li et al. Toward efficient safety helmet detection based on YoloV5 with hierarchical positive sample selection and box density filtering
CN111047818A (en) Forest fire early warning system based on video image
CN101908142A (en) Feature analysis-based video flame detecting method
WO2022037279A1 (en) Operation violation detection method for power transformation site
CN104504369A (en) Wearing condition detection method for safety helmets
CN101751677A (en) Target continuous tracking method based on multi-camera
CN103324910A (en) Fire alarming method based on video detection
CN113111771A (en) Method for identifying unsafe behaviors of power plant workers
CN113537019A (en) Detection method for identifying wearing of safety helmet of transformer substation personnel based on key points
CN104077571B (en) A kind of crowd&#39;s anomaly detection method that model is serialized using single class
CN116977909A (en) Deep learning fire intensity recognition method and system based on multi-modal data
CN104504733A (en) Video abstraction method and system based on moving target detection
CN106408000A (en) Method and device for intelligent detection on personnel safety helmet
CN104270619A (en) Safety warning method and device
WO2022037280A1 (en) Multi-channel cnn based method for detecting power transformation field operation violations
CN105913008A (en) Crowd exceptional event detection method based on hypothesis examination
CN118887608A (en) Method and device for monitoring operation of power transmission line, and electronic equipment
CN113297913A (en) Method for identifying dressing specification of distribution network field operating personnel
CN118262284A (en) Intelligent construction site violation recognition system and method based on deep learning
CN114419483B (en) A method, system and device for detecting illegal setting of fences for power tower operations
CN115565135A (en) Target tracking method and device and electronic equipment
CN117830686B (en) Intelligent management and control platform system and method for wearing personal protective equipment on intelligent construction site

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20141001