[go: up one dir, main page]

CN102521582B - Human upper body detection and splitting method applied to low-contrast video - Google Patents

Human upper body detection and splitting method applied to low-contrast video Download PDF

Info

Publication number
CN102521582B
CN102521582B CN2011104465964A CN201110446596A CN102521582B CN 102521582 B CN102521582 B CN 102521582B CN 2011104465964 A CN2011104465964 A CN 2011104465964A CN 201110446596 A CN201110446596 A CN 201110446596A CN 102521582 B CN102521582 B CN 102521582B
Authority
CN
China
Prior art keywords
human body
foreground
area
upper half
human
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN2011104465964A
Other languages
Chinese (zh)
Other versions
CN102521582A (en
Inventor
谢迪
童若锋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN2011104465964A priority Critical patent/CN102521582B/en
Publication of CN102521582A publication Critical patent/CN102521582A/en
Application granted granted Critical
Publication of CN102521582B publication Critical patent/CN102521582B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Image Analysis (AREA)

Abstract

本发明涉及一种适用于低对比度视频的人体上半身检测及分割的方法,主要由两个过程构成。首先,从当前帧中通过背景剔除技术及形态学方法提取出表示前景对象的连通区域,然后对于每个前景区域,提取出其对应的基于极坐标二维直方图的形状特征,作为一个预先训练好的基于支持向量机的分类器的输入,输出一个对应于人体上半身类及非人体上半身类的类标签。第二步过程,当已经被识别为人体的区域被误判为非人体区域时,用一个能量函数来表征相应区域,同时通过一个能量函数最小化过程纠正错误的轮廓线。最后在获得正确的前景人体轮廓的基础上更新背景帧。本发明能够实时处理较低对比度及分辨率的视频,检测正确率及分割结果都能满足应用的需求。

Figure 201110446596

The invention relates to a method for detecting and segmenting the upper body of a human body suitable for low-contrast video, which mainly consists of two processes. Firstly, the connected region representing the foreground object is extracted from the current frame through background removal technology and morphological method, and then for each foreground region, its corresponding shape feature based on the polar coordinate two-dimensional histogram is extracted as a pre-training The input of a good support vector machine-based classifier outputs a class label corresponding to the human upper body class and the non-human upper body class. In the second step, when the area that has been recognized as a human body is misjudged as a non-human area, an energy function is used to characterize the corresponding area, and the wrong contour is corrected through an energy function minimization process. Finally, the background frame is updated on the basis of obtaining the correct foreground human silhouette. The invention can process video with lower contrast and resolution in real time, and both the detection accuracy rate and segmentation results can meet the requirements of applications.

Figure 201110446596

Description

一种适用于低对比度视频的人体上半身检测及分割的方法A Method for Human Upper Body Detection and Segmentation Applicable to Low Contrast Video

技术领域:Technical field:

本发明涉及视频处理技术领域,尤其涉及人体上半身区域的检测和提取方法,具体地说是一种适用于低对比度视频的人体上半身检测及分割的方法。The invention relates to the technical field of video processing, in particular to a method for detecting and extracting an upper body region of a human body, in particular to a method for detecting and segmenting the upper body of a human body suitable for low-contrast videos.

背景技术:Background technique:

自动检测和视频中的人体区域分割是两个不同的监控应用的关键步骤。人体检测方法通常从视频中找到前景对象并基于形状、颜色以及其它特征把它们标识为人或非人区域。背景剔除法是一种常见的提取前景区域的预处理技术。另一类为基于机器学习的方法,并应用了许多适用于机器学习的新特征。基于梯度的特征最具有代表性。这些方法不需要进行背景剔除的预处理但是却以高昂的计算成本为代价,因此限制了其在实时系统的应用。视频分割方法同样基于背景剔除技术,同时集成了概率框架,如贝叶斯理论和马尔可夫链蒙特卡罗模型。Automatic detection and segmentation of human body regions in videos are crucial steps for two different surveillance applications. Human detection methods typically find foreground objects in videos and identify them as human or non-human regions based on shape, color, and other characteristics. Background culling is a common preprocessing technique for extracting foreground regions. The other category is based on machine learning and applies many new features suitable for machine learning. Gradient-based features are most representative. These methods do not require preprocessing for background culling but come at the cost of high computational cost, thus limiting their application in real-time systems. Video segmentation methods are also based on background removal techniques while integrating probabilistic frameworks such as Bayesian theory and Markov chain Monte Carlo models.

由于许多方法需要提供一个相对较好的背景剔除算法结果,一旦由于光照变化使得环境光照发生改变,这些方法便会失效。虽然一些改进的背景剔除算法能够解决上述问题,但如果前景对象在镜头前保持相当长的一段时间静止不动,那么前景会逐渐变化为背景。另外因为许多监控系统所配备的摄像机其CCD芯片的质量并不高,从而使得所获得的视频对比度较低,现有的方法处理这些视频将更加困难。Since many methods need to provide a relatively good background culling algorithm result, once the environment lighting changes due to lighting changes, these methods will fail. Although some improved background culling algorithms can solve the above problems, if the foreground object remains still in front of the camera for a considerable period of time, the foreground will gradually change into the background. In addition, because the quality of the CCD chips of the cameras equipped in many monitoring systems is not high, the contrast of the obtained videos is low, and the existing methods will be more difficult to process these videos.

发明内容:Invention content:

(1)前景提取:首先指定视频的第一帧作为背景帧,把其格式从RGB颜色空间转换到Lab颜色空间,然后对于输入的每一帧,都用同样的方式进行颜色转换;转换后的输出帧与背景帧使用背景剔除的方法来提取前景对象区域;然后对提取后的每个区域,使用膨胀腐蚀的形态学操作对噪点及空洞进行滤波,最后使用广度优先连通区域搜索算法对前背景区域进行标记,生成前景区域掩码;(1) Foreground extraction: first specify the first frame of the video as the background frame, convert its format from RGB color space to Lab color space, and then perform color conversion in the same way for each input frame; the converted The output frame and the background frame use the method of background removal to extract the foreground object area; then, for each extracted area, use the morphological operation of dilation and erosion to filter the noise and holes, and finally use the breadth-first connected region search algorithm to search the foreground and background The area is marked to generate a foreground area mask;

(2)形状特征提取:首先通过轮廓检测算法提取出前景区域的轮廓线并对其采样;然后以区域质心为原点建立一个极坐标系,对于每个采样轮廓点,把其映射到一个二维平面,最终所有采样点便形成了一二维直方图;最后对得到的直方图归一化并展开,便可以获得一高维向量;(2) Shape feature extraction: First, the contour line of the foreground area is extracted and sampled through the contour detection algorithm; then a polar coordinate system is established with the center of mass of the area as the origin, and for each sampled contour point, it is mapped to a two-dimensional Finally, all sampling points form a two-dimensional histogram; finally, normalize and expand the obtained histogram to obtain a high-dimensional vector;

(3)基于支持向量机的人体上半身模型训练:以上一步骤中获得的向量作为样本,使用以半径基函数为核函数的非线性支持向量机算法对所有训练样本进行K次交叉验证分析,最终生成一非线性决策超平面作为人体上半身区域与非人体上半身区域的分类器;(3) Human upper body model training based on support vector machine: the vector obtained in the previous step is used as a sample, and the nonlinear support vector machine algorithm with the radius basis function as the kernel function is used to perform K cross-validation analysis on all training samples, and finally Generate a non-linear decision hyperplane as a classifier for human upper body regions and non-human upper body regions;

(4)基于支持向量机的人体上半身模型分类:同样以步聚(2)中所获得的向量作为步骤(3)中训练所得分类器的输入,输出经分类器决策映射后的类标签;(4) Human upper body model classification based on support vector machine: also use the vector obtained in step (2) as the input of the classifier trained in step (3), and output the class label after the classifier decision mapping;

(5)能量函数最小化优化过程:对于一个开始被认为是人体区域的前景区域,当其处理过程中被分类器检测到其类标签为非人体区域时,用一个能量函数来对轮廓曲线进行建模,以前一帧中正确的轮廓曲线为初始值,用欧拉-拉格朗日方法求解。(5) Energy function minimization optimization process: For a foreground area that is initially considered to be a human body area, when the classifier detects its class label as a non-human body area during its processing, an energy function is used to process the contour curve Modeling, the correct contour curve in the previous frame is used as the initial value, and the Euler-Lagrangian method is used to solve it.

本发明的方法主要由两大过程构成。首先,从当前帧中通过背景剔除技术及形态学方法提取出表示前景对象的连通区域,然后对于每个前景区域,提取出其对应的基于极坐标二维直方图的形状特征,作为一个预先训练好的基于支持向量机的分类器的输入,输出一个对应于人体上半身类及非人体上半身类的类标签。第二步过程,当已经被识别为人体的区域被误判为非人体区域时,本发明用一个能量函数来表征相应区域,同时通过一个能量函数最小化过程纠正错误的轮廓线。最后在获得正确的前景人体轮廓的基础上更新背景帧。本发明能够实时处理较低对比度及分辨率的视频,检测正确率及分割结果都能满足应用的需求。The method of the present invention mainly consists of two major processes. Firstly, the connected region representing the foreground object is extracted from the current frame through background removal technology and morphological method, and then for each foreground region, its corresponding shape feature based on the polar coordinate two-dimensional histogram is extracted as a pre-training The input of a good support vector machine-based classifier outputs a class label corresponding to the human upper body class and the non-human upper body class. In the second step, when the area that has been recognized as a human body is misjudged as a non-human body area, the present invention uses an energy function to characterize the corresponding area, and at the same time corrects the wrong contour line through an energy function minimization process. Finally, the background frame is updated on the basis of obtaining the correct foreground human silhouette. The invention can process video with lower contrast and resolution in real time, and both the detection accuracy rate and segmentation results can meet the requirements of applications.

附图说明:Description of drawings:

图1为本发明的流程图。Fig. 1 is a flowchart of the present invention.

具体实施方式:Detailed ways:

下面根据本发明的流程图图1对各个部分进行详细说明:Below according to flow chart Fig. 1 of the present invention, each part is described in detail:

1.前景提取1. Foreground extraction

首先指定视频的第一帧作为背景帧,把其格式从RGB颜色空间转换到Lab颜色空间。然后对于输入的每一帧,都用同样的方式进行颜色转换,转换后的输出帧与背景帧使用背景剔除的方法来提取前景对象区域(即用两帧按像素求差取绝对值的方式,其值高于一定的阈值,便认为是前景像素,否则为背景像素)。对提取后的每个区域,使用膨胀腐蚀的形态学操作对噪点及空洞进行滤波,最后使用广度优先连通区域搜索算法对前背景区域进行标记,生成前景区域掩码。First, specify the first frame of the video as the background frame, and convert its format from the RGB color space to the Lab color space. Then for each input frame, the color conversion is performed in the same way, and the converted output frame and background frame use the method of background removal to extract the foreground object area (that is, the method of taking the absolute value of the difference between two frames by pixel, If its value is higher than a certain threshold, it is considered to be a foreground pixel, otherwise it is a background pixel). For each region after extraction, the noise and holes are filtered using the morphological operation of dilation and erosion, and finally the breadth-first connected region search algorithm is used to mark the foreground and background regions to generate a foreground region mask.

2.形状特征提取2. Shape feature extraction

本发明所提出的特征较之于局部的有相梯度直方图特征更加能够描述人体上半身的形状,因此具有更大的区分度,同时具有更小的计算复杂度。Compared with the local phased gradient histogram feature, the feature proposed by the present invention can better describe the shape of the upper body of the human body, so it has a greater degree of discrimination and a smaller computational complexity.

一个人的轮廓,特别是上半身的轮廓,可以看做是一个星凸集。若集S中存在一点x0,使得由x0到S中任何一点的直线段都属于S,则称S为星形域或星形凸集。本发明的形状特征便是以此为依据而设计的。The contour of a person, especially the contour of the upper body, can be regarded as a convex set. If there is a point x 0 in the set S, so that the straight line segment from x 0 to any point in S belongs to S, then S is called a star field or a star convex set. The shape feature of the present invention is designed based on this.

对于一个特定的前景区域,本发明通过广度优先搜索找到前景区域的质心,然后通过边界跟随算法找到同一区域的边界轮廓线。接着在轮廓线上等角度地对轮廓线进行逆时针采样,即以前景区域的质心做为一极坐标系原点,则轮廓线上的每个采样点在此坐标系下便可表示成一组极坐标(θi,ri),i=1,2,...,N,其中ri为区域质心到每个轮廓点的欧式距离,θi为每个轮廓点的极角,N为采样点的总数。随后这些极坐标值被投影到一个二维平面上,平面的x轴表示θ值,y轴表示r值,每个维度分别被量化,均分成m和n份。当一个极坐标值(θi,ri)满足下列条件时:For a specific foreground area, the present invention finds the centroid of the foreground area through breadth-first search, and then finds the boundary contour line of the same area through a boundary following algorithm. Then, the contour line is sampled counterclockwise on the contour line at an equal angle, that is, the centroid of the foreground area is used as the origin of a polar coordinate system, and each sampling point on the contour line can be expressed as a set of polar coordinates in this coordinate system. Coordinates (θ i , r i ), i=1,2,...,N, where r i is the Euclidean distance from the area centroid to each contour point, θ i is the polar angle of each contour point, and N is the sampling total number of points. These polar coordinate values are then projected onto a two-dimensional plane. The x-axis of the plane represents the θ value, and the y-axis represents the r value. Each dimension is quantized separately and divided into m and n parts. When a polar coordinate value (θ i , r i ) satisfies the following conditions:

θk≤θi≤θk+1,rl≤ri≤rl+1,k=0,...,m-1,l=0,...,n-1θ k ≤θ i ≤θ k+1 ,r l ≤r i ≤r l+1 ,k=0,...,m-1,l=0,...,n-1

则增加相应的单元(k,l)的值。当按上述方法遍历完所有的点时,将会形成一具有特定模式的二维直方图。此特定的模式表征着对应轮廓线的特定形状。最后,按行展开此直方图各单元格的值并对其进行归一化后将得到一个m×n维的向量f。显然经本发明获得的形状特征与物体的位置和大小无关。Then increase the value of the corresponding unit (k,l). When all points are traversed according to the above method, a two-dimensional histogram with a specific pattern will be formed. This particular pattern characterizes the particular shape of the corresponding contour. Finally, after expanding the value of each cell of this histogram by row and normalizing it, an m×n-dimensional vector f will be obtained. Obviously, the shape features obtained by the present invention have nothing to do with the position and size of the object.

3.基于支持向量机的人体上半身模型训练与检测3. Human upper body model training and detection based on support vector machine

在训练阶段,大量人体上半身图像和非人体上半身图像被搜集,通过手工标记前景区域从而提取前景的形状特征。这些形状特征所对应高维向量的集合组成了本发明用于训练的样本集。本发明用支持向量机作为训练的算法,其核函数采用了高斯半径基函数:In the training phase, a large number of human upper body images and non-human upper body images are collected, and the foreground area is manually marked to extract the shape features of the foreground. The set of high-dimensional vectors corresponding to these shape features constitutes the sample set used for training in the present invention. The present invention uses support vector machine as the algorithm of training, and its kernel function has adopted Gaussian radius base function:

K(xi,xj)=exp(-γ||xi-xj||2)K(x i , x j )=exp(-γ||x i -x j || 2 )

其中xi,xj为特征向量,γ为归一化常数。Where x i , x j are feature vectors, and γ is a normalization constant.

为了训练得到最佳性能的分类器,本发明使用了K次交叉验证的方法确定支持向量机分类器的两个参数γ与C。即所有数据被分割成K份子数据,一份单独的子数据被保留作为验证数据,其他K-1份子数据用来训练.如上过程被重复次K次,每次选用不同的子数据组合作为验证数据与训练数据,最后对所求结果求均值。通过这个方式本发明确定最优分类性能的参数组合为γ=0.25,C=2.0时,分类准确率约为98%。In order to train a classifier with the best performance, the present invention uses a K times cross-validation method to determine two parameters γ and C of the support vector machine classifier. That is, all data is divided into K sub-data, a separate sub-data is reserved as verification data, and other K-1 sub-data are used for training. The above process is repeated K times, and each time a different combination of sub-data is selected as verification Data and training data, and finally calculate the average of the obtained results. In this way, the present invention determines that the optimal classification performance parameter combination is γ=0.25, and when C=2.0, the classification accuracy rate is about 98%.

在检测阶段,对于每一帧,如果存在前景区域,那么用同样的方法提取出区域轮廓形状特征,作为事先训练好的分类器的输入,分类器将会输出一个布尔值说明当前区域是否为人体上半身区域。In the detection stage, for each frame, if there is a foreground area, then use the same method to extract the area contour shape feature, as the input of the pre-trained classifier, the classifier will output a Boolean value indicating whether the current area is a human body upper body area.

4.能量函数最小化优化过程4. Energy function minimization optimization process

一旦人体上半身区域不能被支持向量机分类器识别为人体类,导致分类错误,那么本发明将对错误的前景轮廓线进行一能量最小化过程,消除由于环境光改变导致的轮廓线扩张造成的误差,从而保证前景轮廓区域的正确性。对于一段闭合的完整轮廓,本发明用一个能量函数Ec(s)来表征:Once the upper body area of the human body cannot be recognized as a human body by the support vector machine classifier, resulting in a classification error, then the present invention will perform an energy minimization process on the wrong foreground contour line to eliminate the error caused by the expansion of the contour line due to changes in ambient light , so as to ensure the correctness of the foreground contour area. For a closed complete profile, the present invention uses an energy function E c (s) to characterize:

Ec(s)=∮(Eint(s)+η(s)Eext(s))dsE c (s)=∮(E int (s)+η(s)E ext (s))ds

其中Eint(s)为轮廓线的内部势能,Eext(s)给出了基于图像的外部限制。η(s)为对应于每个采样点的权重,定义为:where E int (s) is the internal potential energy of the contour and E ext (s) gives the image-based external limit. η(s) is the weight corresponding to each sampling point, defined as:

ηη (( sthe s ii )) == || || ▿▿ II (( xx (( sthe s ii )) ,, ythe y (( sthe s ii )) )) || || 22 ∑∑ ii NN || || ▿▿ II (( xx (( sthe s ii )) ,, ythe y (( sthe s ii )) )) || || 22

其中

Figure GDA00002917651400046
表示图像的梯度,N为采样点的总数。优化的目标是找到使能量泛函Ec(s)最小化的曲线函数v(s)=(x(s),y(s))。本发明采用欧拉-拉格朗日乘数法把泛函式转化为偏微分方程的求解问题,然后对其离散化,最终获得一线性系统Ax=b,其中A为对角线上有且仅有五个非零元素的矩阵。可以用Cholesky分解方法解此线性系统。in
Figure GDA00002917651400046
Indicates the gradient of the image, and N is the total number of sampling points. The goal of optimization is to find the curve function v(s)=(x(s),y(s)) that minimizes the energy functional E c (s). The present invention adopts the Euler-Lagrangian multiplier method to transform the functional into a partial differential equation solution problem, then discretize it, and finally obtain a linear system Ax=b, where A is on the diagonal and A matrix with only five nonzero elements. This linear system can be solved by the Cholesky decomposition method.

5.背景区域更新5. Background area update

在获得正确前景区域的基础上,本发明用线性插值的方式更新背景区域:On the basis of obtaining the correct foreground area, the present invention uses linear interpolation to update the background area:

II BB (( xx ,, ythe y )) == αα II BB (( tt )) (( xx ,, ythe y )) ++ (( 11 -- αα )) II BB ** (( xx ,, ythe y ))

其中IB(x,y)为更新后位置(x,y)对应的背景帧像素值,

Figure GDA00002917651400043
为更新前相同位置的像素值。
Figure GDA00002917651400044
为对应的当前帧中属于背景区域的像素值。对于前景区域,仅简单地拷贝相应位置上的像素值
Figure GDA00002917651400045
Where I B (x, y) is the pixel value of the background frame corresponding to the updated position (x, y),
Figure GDA00002917651400043
is the pixel value at the same position before the update.
Figure GDA00002917651400044
is the pixel value belonging to the background area in the corresponding current frame. For the foreground area, simply copy the pixel value at the corresponding position
Figure GDA00002917651400045

应该理解到的是:上述实施例只是对本发明的说明,而不是对本发明的限制,任何不超出本发明实质精神范围内的发明创造,均落入本发明的保护范围之内。It should be understood that: the above-mentioned embodiments are only descriptions of the present invention, rather than limitations of the present invention, and any inventions that do not exceed the spirit of the present invention fall within the protection scope of the present invention.

Claims (5)

1. one kind is applicable to the upper half of human body detection of low contrast video and the method for cutting apart, and it is characterized in that this method may further comprise the steps:
(1) foreground extraction: first frame of designated frame as a setting at first, its form from the RGB color space conversion to the Lab color space, for each frame of input, all carry out color conversion in the same way then; The method that output frame after the conversion and background frames use background to reject is extracted the foreground object zone; To each zone after extracting, use the morphological operation of the corrosion of expanding that noise and cavity are carried out filtering then, use breadth First connected region searching algorithm that mark is carried out in preceding background area at last, generate the foreground area mask;
(2) Shape Feature Extraction: at first extract the outline line of foreground area by the profile detection algorithm and to its sampling; Be that initial point is set up a polar coordinate system with regional barycenter then, for each sampled contour point, it be mapped to a two dimensional surface, finally all sampled points have just formed a two-dimensional histogram; Histogram normalization and expansion to obtaining obtains a high dimension vector at last;
(3) based on the upper half of human body model training of support vector machine: with the vector that obtains in the previous step as sample, use with the radius basis function and as the non-linear algorithm of support vector machine of kernel function all training samples are carried out K cross validation analysis, finally generate a non-linear decision-making lineoid as the sorter of upper half of human body zone with non-upper half of human body zone;
(4) based on the upper half of human body category of model of support vector machine: equally with the input as training gained sorter in the step (3) of the vector that obtained in the step (2), the class label of output after sorter decision-making mapping;
(5) energy function minimizes optimization: for a foreground area that begins to be considered to human region, when being classified device in its processing procedure and detecting its class label and be non-human region, come contour curve is carried out modeling with an energy function, be initial value with contour curve correct in the former frame, find the solution with Euler-Lagrangian method, and upgrade the background area with last result.
2. the upper half of human body that the is applicable to the low contrast video as claimed in claim 1 method that detects and cut apart, it is as follows to it is characterized in that method that the use background described in the step (1) is rejected is extracted the process in foreground object zone: according to pixels ask the mode that takes absolute value that differs from two frames, its value is higher than certain threshold value, just think foreground pixel, otherwise be background pixel.
3. the upper half of human body that the is applicable to the low contrast video as claimed in claim 1 method that detects and cut apart is characterized in that the detailed process of step (2) is as follows:
For a specific foreground area, find the barycenter of foreground area by BFS (Breadth First Search), find the boundary contour of the same area then by the border following algorithm; Then sample counterclockwise to outline line in equal angles ground on outline line, and the barycenter of foreground area is labeled as a polar coordinate system initial point, and each sampled point on the outline line just can be expressed as one group of polar coordinates (θ under this coordinate system i, r i), i=1,2 ..., N, wherein r iBe the Euclidean distance of regional barycenter to each point, θ iBe the polar angle of each point, N is the sum of sampled point; These polar coordinates values are projected on the two dimensional surface subsequently, and the x axle on plane is represented the θ value, and the y axle is represented the r value, and each dimension is quantized respectively, are divided into m and n part; As a polar coordinates value (θ i, r i) when satisfying following condition:
θ k≤θ i≤θ k+1,r l≤r i≤r l+1,k=0,...,m-1,l=0,...,n-1
Then increase corresponding unit (k, value l); When having traveled through all points as stated above, will form a two-dimensional histogram with AD HOC, this specific pattern is characterizing the given shape of corresponding outline line; At last, carry out to obtain the vector f that a m * n ties up after the normalization by the value of this each cell of histogram of row expansion and to it.
4. the upper half of human body that the is applicable to the low contrast video as claimed in claim 1 method that detects and cut apart is characterized in that the detailed process of step (3) is as follows:
In the training stage, a large amount of upper half of human body images and non-upper half of human body image are collected, thereby extract the shape facility of prospect by the manual markings foreground area, these shape facilities the set of corresponding high dimension vector formed the sample set that the present invention is used for training, adopt support vector machine as the algorithm of training, its kernel function adopts Gauss's radius basis function:
K(x i,x j)=exp(-γ||x i-x j|| 2)
X wherein i, x jBe proper vector, γ is normaliztion constant;
Adopt the method for K cross validation to determine two parameter γ and the C of support vector machine classifier: all data are divided into K one's share of expenses for a joint undertaking data, a independent subdata is retained as verification msg, other K-1 one's share of expenses for a joint undertaking data are used for training, as above process is repeated inferior K time, select for use different subdata combinations as verification msg and training data at every turn, at last asking result is averaged; At detection-phase, for each frame, if there is foreground area, using the same method so extracts the region contour shape facility, as the input of the good sorter of precondition, sorter will export whether a Boolean explanation current region is the upper half of human body zone.
5. the upper half of human body that the is applicable to the low contrast video as claimed in claim 1 method that detects and cut apart is characterized in that the detailed process of step (5) is as follows:
Count E with an energy functional c(s) characterize the integrity profile of one section closure:
E c(s)=∮(E int(s)+η(s)E ext(s))ds
E wherein Int(s) be the inside potential energy of outline line, E Ext(s) provided outside limits based on image, η (s) is the weight corresponding to each sampled point, is defined as:
η ( s i ) = | | ▿ I ( x ( s i ) , y ( s i ) ) | | 2 ∑ i N | | ▿ I ( x ( s i ) , y ( s i ) ) | | 2
Wherein
Figure FDA00002917651300036
The gradient of presentation video, N is the sum of sampled point;
Adopt Euler-lagrange's method of multipliers that the functional formula is converted into the problem of finding the solution of partial differential equation, then to its discretize, finally obtain a linear system Ax=b, wherein A is the matrix that has and only have five nonzero elements on the diagonal line; Can use this linear system of Cholesky decomposition method solution;
On the basis that obtains correct foreground area, upgrade the background area with the mode of linear interpolation:
I B ( x , y ) = α I B ( t ) ( x , y ) + ( 1 - α ) I B * ( x , y )
I wherein B(x, y) for upgrade the position, back (x, y) Dui Ying background frames pixel value,
Figure FDA00002917651300033
Be the pixel value of same position before upgrading,
Figure FDA00002917651300034
The pixel value that belongs to the background area in the present frame for correspondence; For foreground area, only copy the pixel value on the relevant position simply
Figure FDA00002917651300035
CN2011104465964A 2011-12-28 2011-12-28 Human upper body detection and splitting method applied to low-contrast video Expired - Fee Related CN102521582B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2011104465964A CN102521582B (en) 2011-12-28 2011-12-28 Human upper body detection and splitting method applied to low-contrast video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2011104465964A CN102521582B (en) 2011-12-28 2011-12-28 Human upper body detection and splitting method applied to low-contrast video

Publications (2)

Publication Number Publication Date
CN102521582A CN102521582A (en) 2012-06-27
CN102521582B true CN102521582B (en) 2013-09-25

Family

ID=46292493

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2011104465964A Expired - Fee Related CN102521582B (en) 2011-12-28 2011-12-28 Human upper body detection and splitting method applied to low-contrast video

Country Status (1)

Country Link
CN (1) CN102521582B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10326928B2 (en) 2016-01-08 2019-06-18 Samsung Electronics Co., Ltd. Image processing apparatus for determining whether section of target area matches section of person area and control method thereof

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104683765B (en) * 2015-02-04 2019-04-12 上海依图网络科技有限公司 A kind of video concentration method based on detecting moving object
CN108804992B (en) * 2017-05-08 2022-08-26 电子科技大学 Crowd counting method based on deep learning
CN107707975A (en) * 2017-09-20 2018-02-16 天津大学 Video intelligent clipping method based on monitor supervision platform
CN113379930B (en) * 2021-05-25 2023-03-24 广州紫为云科技有限公司 Immersive interaction method and device through human body graph and storage medium
CN114998390B (en) * 2022-08-02 2022-10-21 环球数科集团有限公司 Visual analysis system and method for embedded intelligent camera

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101131728A (en) * 2007-09-29 2008-02-27 东华大学 A Face Shape Matching Method Based on Shape Context
CN101996307A (en) * 2009-08-10 2011-03-30 上海理视微电子有限公司 Intelligent video human body identification method
CN101834982B (en) * 2010-05-28 2012-04-25 上海交通大学 Multi-mode-based violence video layered screening method

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10326928B2 (en) 2016-01-08 2019-06-18 Samsung Electronics Co., Ltd. Image processing apparatus for determining whether section of target area matches section of person area and control method thereof

Also Published As

Publication number Publication date
CN102521582A (en) 2012-06-27

Similar Documents

Publication Publication Date Title
CN107545239B (en) Fake plate detection method based on license plate recognition and vehicle characteristic matching
CN102509098B (en) A fisheye image vehicle recognition method
CN103886589B (en) Object-oriented automated high-precision edge extracting method
CN107633226B (en) Human body motion tracking feature processing method
CN102629385B (en) A target matching and tracking system and method based on multi-camera information fusion
CN107909081B (en) A fast acquisition and fast calibration method for image datasets in deep learning
CN102521582B (en) Human upper body detection and splitting method applied to low-contrast video
CN103258332B (en) A kind of detection method of the moving target of resisting illumination variation
CN107301383A (en) A kind of pavement marking recognition methods based on Fast R CNN
CN104598885B (en) The detection of word label and localization method in street view image
CN110826389B (en) Gait recognition method based on attention 3D frequency convolutional neural network
CN107657279A (en) A kind of remote sensing target detection method based on a small amount of sample
CN106096602A (en) Chinese license plate recognition method based on convolutional neural network
CN104050471A (en) A method and system for character detection in natural scenes
CN103955949B (en) Moving target detecting method based on Mean-shift algorithm
CN107895379A (en) The innovatory algorithm of foreground extraction in a kind of video monitoring
Fang et al. Real-time hand posture recognition using hand geometric features and fisher vector
CN105718866A (en) Visual target detection and identification method
CN111582126A (en) Pedestrian Re-identification Method Based on Multi-scale Pedestrian Contour Segmentation and Fusion
CN113627481A (en) Multi-model combined unmanned aerial vehicle garbage classification method for smart gardens
CN104036296B (en) A kind of expression of image and processing method and processing device
CN106446890A (en) Candidate area extraction method based on window scoring and superpixel segmentation
CN109003275A (en) The dividing method of weld defect image
CN110008900A (en) A Region-to-target Candidate Target Extraction Method for Visible Light Remote Sensing Images
CN109858342B (en) A face pose estimation method that combines hand-designed descriptors and deep features

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20130925