CN102521582B - Human upper body detection and splitting method applied to low-contrast video - Google Patents
Human upper body detection and splitting method applied to low-contrast video Download PDFInfo
- Publication number
- CN102521582B CN102521582B CN2011104465964A CN201110446596A CN102521582B CN 102521582 B CN102521582 B CN 102521582B CN 2011104465964 A CN2011104465964 A CN 2011104465964A CN 201110446596 A CN201110446596 A CN 201110446596A CN 102521582 B CN102521582 B CN 102521582B
- Authority
- CN
- China
- Prior art keywords
- human body
- foreground
- area
- upper half
- human
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000000034 method Methods 0.000 title claims abstract description 55
- 238000001514 detection method Methods 0.000 title claims abstract description 10
- 238000012549 training Methods 0.000 claims abstract description 18
- 238000012706 support-vector machine Methods 0.000 claims abstract description 14
- 230000000877 morphologic effect Effects 0.000 claims abstract description 5
- 239000013598 vector Substances 0.000 claims description 12
- 238000000605 extraction Methods 0.000 claims description 7
- 238000005070 sampling Methods 0.000 claims description 6
- 238000006243 chemical reaction Methods 0.000 claims description 5
- 238000002790 cross-validation Methods 0.000 claims description 4
- 238000005457 optimization Methods 0.000 claims description 4
- 238000012795 verification Methods 0.000 claims description 4
- 238000010606 normalization Methods 0.000 claims description 3
- 238000012545 processing Methods 0.000 claims description 3
- 238000000354 decomposition reaction Methods 0.000 claims description 2
- 238000013507 mapping Methods 0.000 claims description 2
- 239000011159 matrix material Substances 0.000 claims description 2
- 238000005381 potential energy Methods 0.000 claims description 2
- 239000000284 extract Substances 0.000 claims 3
- 238000005260 corrosion Methods 0.000 claims 1
- 230000007797 corrosion Effects 0.000 claims 1
- 238000001914 filtration Methods 0.000 claims 1
- 230000000717 retained effect Effects 0.000 claims 1
- 230000011218 segmentation Effects 0.000 abstract description 4
- 230000006870 function Effects 0.000 description 11
- 210000000746 body region Anatomy 0.000 description 4
- 230000010339 dilation Effects 0.000 description 2
- 230000003628 erosive effect Effects 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 238000007781 pre-processing Methods 0.000 description 2
- 238000010845 search algorithm Methods 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 1
Images
Classifications
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Image Analysis (AREA)
Abstract
本发明涉及一种适用于低对比度视频的人体上半身检测及分割的方法,主要由两个过程构成。首先,从当前帧中通过背景剔除技术及形态学方法提取出表示前景对象的连通区域,然后对于每个前景区域,提取出其对应的基于极坐标二维直方图的形状特征,作为一个预先训练好的基于支持向量机的分类器的输入,输出一个对应于人体上半身类及非人体上半身类的类标签。第二步过程,当已经被识别为人体的区域被误判为非人体区域时,用一个能量函数来表征相应区域,同时通过一个能量函数最小化过程纠正错误的轮廓线。最后在获得正确的前景人体轮廓的基础上更新背景帧。本发明能够实时处理较低对比度及分辨率的视频,检测正确率及分割结果都能满足应用的需求。
The invention relates to a method for detecting and segmenting the upper body of a human body suitable for low-contrast video, which mainly consists of two processes. Firstly, the connected region representing the foreground object is extracted from the current frame through background removal technology and morphological method, and then for each foreground region, its corresponding shape feature based on the polar coordinate two-dimensional histogram is extracted as a pre-training The input of a good support vector machine-based classifier outputs a class label corresponding to the human upper body class and the non-human upper body class. In the second step, when the area that has been recognized as a human body is misjudged as a non-human area, an energy function is used to characterize the corresponding area, and the wrong contour is corrected through an energy function minimization process. Finally, the background frame is updated on the basis of obtaining the correct foreground human silhouette. The invention can process video with lower contrast and resolution in real time, and both the detection accuracy rate and segmentation results can meet the requirements of applications.
Description
技术领域:Technical field:
本发明涉及视频处理技术领域,尤其涉及人体上半身区域的检测和提取方法,具体地说是一种适用于低对比度视频的人体上半身检测及分割的方法。The invention relates to the technical field of video processing, in particular to a method for detecting and extracting an upper body region of a human body, in particular to a method for detecting and segmenting the upper body of a human body suitable for low-contrast videos.
背景技术:Background technique:
自动检测和视频中的人体区域分割是两个不同的监控应用的关键步骤。人体检测方法通常从视频中找到前景对象并基于形状、颜色以及其它特征把它们标识为人或非人区域。背景剔除法是一种常见的提取前景区域的预处理技术。另一类为基于机器学习的方法,并应用了许多适用于机器学习的新特征。基于梯度的特征最具有代表性。这些方法不需要进行背景剔除的预处理但是却以高昂的计算成本为代价,因此限制了其在实时系统的应用。视频分割方法同样基于背景剔除技术,同时集成了概率框架,如贝叶斯理论和马尔可夫链蒙特卡罗模型。Automatic detection and segmentation of human body regions in videos are crucial steps for two different surveillance applications. Human detection methods typically find foreground objects in videos and identify them as human or non-human regions based on shape, color, and other characteristics. Background culling is a common preprocessing technique for extracting foreground regions. The other category is based on machine learning and applies many new features suitable for machine learning. Gradient-based features are most representative. These methods do not require preprocessing for background culling but come at the cost of high computational cost, thus limiting their application in real-time systems. Video segmentation methods are also based on background removal techniques while integrating probabilistic frameworks such as Bayesian theory and Markov chain Monte Carlo models.
由于许多方法需要提供一个相对较好的背景剔除算法结果,一旦由于光照变化使得环境光照发生改变,这些方法便会失效。虽然一些改进的背景剔除算法能够解决上述问题,但如果前景对象在镜头前保持相当长的一段时间静止不动,那么前景会逐渐变化为背景。另外因为许多监控系统所配备的摄像机其CCD芯片的质量并不高,从而使得所获得的视频对比度较低,现有的方法处理这些视频将更加困难。Since many methods need to provide a relatively good background culling algorithm result, once the environment lighting changes due to lighting changes, these methods will fail. Although some improved background culling algorithms can solve the above problems, if the foreground object remains still in front of the camera for a considerable period of time, the foreground will gradually change into the background. In addition, because the quality of the CCD chips of the cameras equipped in many monitoring systems is not high, the contrast of the obtained videos is low, and the existing methods will be more difficult to process these videos.
发明内容:Invention content:
(1)前景提取:首先指定视频的第一帧作为背景帧,把其格式从RGB颜色空间转换到Lab颜色空间,然后对于输入的每一帧,都用同样的方式进行颜色转换;转换后的输出帧与背景帧使用背景剔除的方法来提取前景对象区域;然后对提取后的每个区域,使用膨胀腐蚀的形态学操作对噪点及空洞进行滤波,最后使用广度优先连通区域搜索算法对前背景区域进行标记,生成前景区域掩码;(1) Foreground extraction: first specify the first frame of the video as the background frame, convert its format from RGB color space to Lab color space, and then perform color conversion in the same way for each input frame; the converted The output frame and the background frame use the method of background removal to extract the foreground object area; then, for each extracted area, use the morphological operation of dilation and erosion to filter the noise and holes, and finally use the breadth-first connected region search algorithm to search the foreground and background The area is marked to generate a foreground area mask;
(2)形状特征提取:首先通过轮廓检测算法提取出前景区域的轮廓线并对其采样;然后以区域质心为原点建立一个极坐标系,对于每个采样轮廓点,把其映射到一个二维平面,最终所有采样点便形成了一二维直方图;最后对得到的直方图归一化并展开,便可以获得一高维向量;(2) Shape feature extraction: First, the contour line of the foreground area is extracted and sampled through the contour detection algorithm; then a polar coordinate system is established with the center of mass of the area as the origin, and for each sampled contour point, it is mapped to a two-dimensional Finally, all sampling points form a two-dimensional histogram; finally, normalize and expand the obtained histogram to obtain a high-dimensional vector;
(3)基于支持向量机的人体上半身模型训练:以上一步骤中获得的向量作为样本,使用以半径基函数为核函数的非线性支持向量机算法对所有训练样本进行K次交叉验证分析,最终生成一非线性决策超平面作为人体上半身区域与非人体上半身区域的分类器;(3) Human upper body model training based on support vector machine: the vector obtained in the previous step is used as a sample, and the nonlinear support vector machine algorithm with the radius basis function as the kernel function is used to perform K cross-validation analysis on all training samples, and finally Generate a non-linear decision hyperplane as a classifier for human upper body regions and non-human upper body regions;
(4)基于支持向量机的人体上半身模型分类:同样以步聚(2)中所获得的向量作为步骤(3)中训练所得分类器的输入,输出经分类器决策映射后的类标签;(4) Human upper body model classification based on support vector machine: also use the vector obtained in step (2) as the input of the classifier trained in step (3), and output the class label after the classifier decision mapping;
(5)能量函数最小化优化过程:对于一个开始被认为是人体区域的前景区域,当其处理过程中被分类器检测到其类标签为非人体区域时,用一个能量函数来对轮廓曲线进行建模,以前一帧中正确的轮廓曲线为初始值,用欧拉-拉格朗日方法求解。(5) Energy function minimization optimization process: For a foreground area that is initially considered to be a human body area, when the classifier detects its class label as a non-human body area during its processing, an energy function is used to process the contour curve Modeling, the correct contour curve in the previous frame is used as the initial value, and the Euler-Lagrangian method is used to solve it.
本发明的方法主要由两大过程构成。首先,从当前帧中通过背景剔除技术及形态学方法提取出表示前景对象的连通区域,然后对于每个前景区域,提取出其对应的基于极坐标二维直方图的形状特征,作为一个预先训练好的基于支持向量机的分类器的输入,输出一个对应于人体上半身类及非人体上半身类的类标签。第二步过程,当已经被识别为人体的区域被误判为非人体区域时,本发明用一个能量函数来表征相应区域,同时通过一个能量函数最小化过程纠正错误的轮廓线。最后在获得正确的前景人体轮廓的基础上更新背景帧。本发明能够实时处理较低对比度及分辨率的视频,检测正确率及分割结果都能满足应用的需求。The method of the present invention mainly consists of two major processes. Firstly, the connected region representing the foreground object is extracted from the current frame through background removal technology and morphological method, and then for each foreground region, its corresponding shape feature based on the polar coordinate two-dimensional histogram is extracted as a pre-training The input of a good support vector machine-based classifier outputs a class label corresponding to the human upper body class and the non-human upper body class. In the second step, when the area that has been recognized as a human body is misjudged as a non-human body area, the present invention uses an energy function to characterize the corresponding area, and at the same time corrects the wrong contour line through an energy function minimization process. Finally, the background frame is updated on the basis of obtaining the correct foreground human silhouette. The invention can process video with lower contrast and resolution in real time, and both the detection accuracy rate and segmentation results can meet the requirements of applications.
附图说明:Description of drawings:
图1为本发明的流程图。Fig. 1 is a flowchart of the present invention.
具体实施方式:Detailed ways:
下面根据本发明的流程图图1对各个部分进行详细说明:Below according to flow chart Fig. 1 of the present invention, each part is described in detail:
1.前景提取1. Foreground extraction
首先指定视频的第一帧作为背景帧,把其格式从RGB颜色空间转换到Lab颜色空间。然后对于输入的每一帧,都用同样的方式进行颜色转换,转换后的输出帧与背景帧使用背景剔除的方法来提取前景对象区域(即用两帧按像素求差取绝对值的方式,其值高于一定的阈值,便认为是前景像素,否则为背景像素)。对提取后的每个区域,使用膨胀腐蚀的形态学操作对噪点及空洞进行滤波,最后使用广度优先连通区域搜索算法对前背景区域进行标记,生成前景区域掩码。First, specify the first frame of the video as the background frame, and convert its format from the RGB color space to the Lab color space. Then for each input frame, the color conversion is performed in the same way, and the converted output frame and background frame use the method of background removal to extract the foreground object area (that is, the method of taking the absolute value of the difference between two frames by pixel, If its value is higher than a certain threshold, it is considered to be a foreground pixel, otherwise it is a background pixel). For each region after extraction, the noise and holes are filtered using the morphological operation of dilation and erosion, and finally the breadth-first connected region search algorithm is used to mark the foreground and background regions to generate a foreground region mask.
2.形状特征提取2. Shape feature extraction
本发明所提出的特征较之于局部的有相梯度直方图特征更加能够描述人体上半身的形状,因此具有更大的区分度,同时具有更小的计算复杂度。Compared with the local phased gradient histogram feature, the feature proposed by the present invention can better describe the shape of the upper body of the human body, so it has a greater degree of discrimination and a smaller computational complexity.
一个人的轮廓,特别是上半身的轮廓,可以看做是一个星凸集。若集S中存在一点x0,使得由x0到S中任何一点的直线段都属于S,则称S为星形域或星形凸集。本发明的形状特征便是以此为依据而设计的。The contour of a person, especially the contour of the upper body, can be regarded as a convex set. If there is a point x 0 in the set S, so that the straight line segment from x 0 to any point in S belongs to S, then S is called a star field or a star convex set. The shape feature of the present invention is designed based on this.
对于一个特定的前景区域,本发明通过广度优先搜索找到前景区域的质心,然后通过边界跟随算法找到同一区域的边界轮廓线。接着在轮廓线上等角度地对轮廓线进行逆时针采样,即以前景区域的质心做为一极坐标系原点,则轮廓线上的每个采样点在此坐标系下便可表示成一组极坐标(θi,ri),i=1,2,...,N,其中ri为区域质心到每个轮廓点的欧式距离,θi为每个轮廓点的极角,N为采样点的总数。随后这些极坐标值被投影到一个二维平面上,平面的x轴表示θ值,y轴表示r值,每个维度分别被量化,均分成m和n份。当一个极坐标值(θi,ri)满足下列条件时:For a specific foreground area, the present invention finds the centroid of the foreground area through breadth-first search, and then finds the boundary contour line of the same area through a boundary following algorithm. Then, the contour line is sampled counterclockwise on the contour line at an equal angle, that is, the centroid of the foreground area is used as the origin of a polar coordinate system, and each sampling point on the contour line can be expressed as a set of polar coordinates in this coordinate system. Coordinates (θ i , r i ), i=1,2,...,N, where r i is the Euclidean distance from the area centroid to each contour point, θ i is the polar angle of each contour point, and N is the sampling total number of points. These polar coordinate values are then projected onto a two-dimensional plane. The x-axis of the plane represents the θ value, and the y-axis represents the r value. Each dimension is quantized separately and divided into m and n parts. When a polar coordinate value (θ i , r i ) satisfies the following conditions:
θk≤θi≤θk+1,rl≤ri≤rl+1,k=0,...,m-1,l=0,...,n-1θ k ≤θ i ≤θ k+1 ,r l ≤r i ≤r l+1 ,k=0,...,m-1,l=0,...,n-1
则增加相应的单元(k,l)的值。当按上述方法遍历完所有的点时,将会形成一具有特定模式的二维直方图。此特定的模式表征着对应轮廓线的特定形状。最后,按行展开此直方图各单元格的值并对其进行归一化后将得到一个m×n维的向量f。显然经本发明获得的形状特征与物体的位置和大小无关。Then increase the value of the corresponding unit (k,l). When all points are traversed according to the above method, a two-dimensional histogram with a specific pattern will be formed. This particular pattern characterizes the particular shape of the corresponding contour. Finally, after expanding the value of each cell of this histogram by row and normalizing it, an m×n-dimensional vector f will be obtained. Obviously, the shape features obtained by the present invention have nothing to do with the position and size of the object.
3.基于支持向量机的人体上半身模型训练与检测3. Human upper body model training and detection based on support vector machine
在训练阶段,大量人体上半身图像和非人体上半身图像被搜集,通过手工标记前景区域从而提取前景的形状特征。这些形状特征所对应高维向量的集合组成了本发明用于训练的样本集。本发明用支持向量机作为训练的算法,其核函数采用了高斯半径基函数:In the training phase, a large number of human upper body images and non-human upper body images are collected, and the foreground area is manually marked to extract the shape features of the foreground. The set of high-dimensional vectors corresponding to these shape features constitutes the sample set used for training in the present invention. The present invention uses support vector machine as the algorithm of training, and its kernel function has adopted Gaussian radius base function:
K(xi,xj)=exp(-γ||xi-xj||2)K(x i , x j )=exp(-γ||x i -x j || 2 )
其中xi,xj为特征向量,γ为归一化常数。Where x i , x j are feature vectors, and γ is a normalization constant.
为了训练得到最佳性能的分类器,本发明使用了K次交叉验证的方法确定支持向量机分类器的两个参数γ与C。即所有数据被分割成K份子数据,一份单独的子数据被保留作为验证数据,其他K-1份子数据用来训练.如上过程被重复次K次,每次选用不同的子数据组合作为验证数据与训练数据,最后对所求结果求均值。通过这个方式本发明确定最优分类性能的参数组合为γ=0.25,C=2.0时,分类准确率约为98%。In order to train a classifier with the best performance, the present invention uses a K times cross-validation method to determine two parameters γ and C of the support vector machine classifier. That is, all data is divided into K sub-data, a separate sub-data is reserved as verification data, and other K-1 sub-data are used for training. The above process is repeated K times, and each time a different combination of sub-data is selected as verification Data and training data, and finally calculate the average of the obtained results. In this way, the present invention determines that the optimal classification performance parameter combination is γ=0.25, and when C=2.0, the classification accuracy rate is about 98%.
在检测阶段,对于每一帧,如果存在前景区域,那么用同样的方法提取出区域轮廓形状特征,作为事先训练好的分类器的输入,分类器将会输出一个布尔值说明当前区域是否为人体上半身区域。In the detection stage, for each frame, if there is a foreground area, then use the same method to extract the area contour shape feature, as the input of the pre-trained classifier, the classifier will output a Boolean value indicating whether the current area is a human body upper body area.
4.能量函数最小化优化过程4. Energy function minimization optimization process
一旦人体上半身区域不能被支持向量机分类器识别为人体类,导致分类错误,那么本发明将对错误的前景轮廓线进行一能量最小化过程,消除由于环境光改变导致的轮廓线扩张造成的误差,从而保证前景轮廓区域的正确性。对于一段闭合的完整轮廓,本发明用一个能量函数Ec(s)来表征:Once the upper body area of the human body cannot be recognized as a human body by the support vector machine classifier, resulting in a classification error, then the present invention will perform an energy minimization process on the wrong foreground contour line to eliminate the error caused by the expansion of the contour line due to changes in ambient light , so as to ensure the correctness of the foreground contour area. For a closed complete profile, the present invention uses an energy function E c (s) to characterize:
Ec(s)=∮(Eint(s)+η(s)Eext(s))dsE c (s)=∮(E int (s)+η(s)E ext (s))ds
其中Eint(s)为轮廓线的内部势能,Eext(s)给出了基于图像的外部限制。η(s)为对应于每个采样点的权重,定义为:where E int (s) is the internal potential energy of the contour and E ext (s) gives the image-based external limit. η(s) is the weight corresponding to each sampling point, defined as:
其中表示图像的梯度,N为采样点的总数。优化的目标是找到使能量泛函Ec(s)最小化的曲线函数v(s)=(x(s),y(s))。本发明采用欧拉-拉格朗日乘数法把泛函式转化为偏微分方程的求解问题,然后对其离散化,最终获得一线性系统Ax=b,其中A为对角线上有且仅有五个非零元素的矩阵。可以用Cholesky分解方法解此线性系统。in Indicates the gradient of the image, and N is the total number of sampling points. The goal of optimization is to find the curve function v(s)=(x(s),y(s)) that minimizes the energy functional E c (s). The present invention adopts the Euler-Lagrangian multiplier method to transform the functional into a partial differential equation solution problem, then discretize it, and finally obtain a linear system Ax=b, where A is on the diagonal and A matrix with only five nonzero elements. This linear system can be solved by the Cholesky decomposition method.
5.背景区域更新5. Background area update
在获得正确前景区域的基础上,本发明用线性插值的方式更新背景区域:On the basis of obtaining the correct foreground area, the present invention uses linear interpolation to update the background area:
其中IB(x,y)为更新后位置(x,y)对应的背景帧像素值,为更新前相同位置的像素值。为对应的当前帧中属于背景区域的像素值。对于前景区域,仅简单地拷贝相应位置上的像素值 Where I B (x, y) is the pixel value of the background frame corresponding to the updated position (x, y), is the pixel value at the same position before the update. is the pixel value belonging to the background area in the corresponding current frame. For the foreground area, simply copy the pixel value at the corresponding position
应该理解到的是:上述实施例只是对本发明的说明,而不是对本发明的限制,任何不超出本发明实质精神范围内的发明创造,均落入本发明的保护范围之内。It should be understood that: the above-mentioned embodiments are only descriptions of the present invention, rather than limitations of the present invention, and any inventions that do not exceed the spirit of the present invention fall within the protection scope of the present invention.
Claims (5)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2011104465964A CN102521582B (en) | 2011-12-28 | 2011-12-28 | Human upper body detection and splitting method applied to low-contrast video |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2011104465964A CN102521582B (en) | 2011-12-28 | 2011-12-28 | Human upper body detection and splitting method applied to low-contrast video |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102521582A CN102521582A (en) | 2012-06-27 |
CN102521582B true CN102521582B (en) | 2013-09-25 |
Family
ID=46292493
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2011104465964A Expired - Fee Related CN102521582B (en) | 2011-12-28 | 2011-12-28 | Human upper body detection and splitting method applied to low-contrast video |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102521582B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10326928B2 (en) | 2016-01-08 | 2019-06-18 | Samsung Electronics Co., Ltd. | Image processing apparatus for determining whether section of target area matches section of person area and control method thereof |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104683765B (en) * | 2015-02-04 | 2019-04-12 | 上海依图网络科技有限公司 | A kind of video concentration method based on detecting moving object |
CN108804992B (en) * | 2017-05-08 | 2022-08-26 | 电子科技大学 | Crowd counting method based on deep learning |
CN107707975A (en) * | 2017-09-20 | 2018-02-16 | 天津大学 | Video intelligent clipping method based on monitor supervision platform |
CN113379930B (en) * | 2021-05-25 | 2023-03-24 | 广州紫为云科技有限公司 | Immersive interaction method and device through human body graph and storage medium |
CN114998390B (en) * | 2022-08-02 | 2022-10-21 | 环球数科集团有限公司 | Visual analysis system and method for embedded intelligent camera |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101131728A (en) * | 2007-09-29 | 2008-02-27 | 东华大学 | A Face Shape Matching Method Based on Shape Context |
CN101996307A (en) * | 2009-08-10 | 2011-03-30 | 上海理视微电子有限公司 | Intelligent video human body identification method |
CN101834982B (en) * | 2010-05-28 | 2012-04-25 | 上海交通大学 | Multi-mode-based violence video layered screening method |
-
2011
- 2011-12-28 CN CN2011104465964A patent/CN102521582B/en not_active Expired - Fee Related
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10326928B2 (en) | 2016-01-08 | 2019-06-18 | Samsung Electronics Co., Ltd. | Image processing apparatus for determining whether section of target area matches section of person area and control method thereof |
Also Published As
Publication number | Publication date |
---|---|
CN102521582A (en) | 2012-06-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107545239B (en) | Fake plate detection method based on license plate recognition and vehicle characteristic matching | |
CN102509098B (en) | A fisheye image vehicle recognition method | |
CN103886589B (en) | Object-oriented automated high-precision edge extracting method | |
CN107633226B (en) | Human body motion tracking feature processing method | |
CN102629385B (en) | A target matching and tracking system and method based on multi-camera information fusion | |
CN107909081B (en) | A fast acquisition and fast calibration method for image datasets in deep learning | |
CN102521582B (en) | Human upper body detection and splitting method applied to low-contrast video | |
CN103258332B (en) | A kind of detection method of the moving target of resisting illumination variation | |
CN107301383A (en) | A kind of pavement marking recognition methods based on Fast R CNN | |
CN104598885B (en) | The detection of word label and localization method in street view image | |
CN110826389B (en) | Gait recognition method based on attention 3D frequency convolutional neural network | |
CN107657279A (en) | A kind of remote sensing target detection method based on a small amount of sample | |
CN106096602A (en) | Chinese license plate recognition method based on convolutional neural network | |
CN104050471A (en) | A method and system for character detection in natural scenes | |
CN103955949B (en) | Moving target detecting method based on Mean-shift algorithm | |
CN107895379A (en) | The innovatory algorithm of foreground extraction in a kind of video monitoring | |
Fang et al. | Real-time hand posture recognition using hand geometric features and fisher vector | |
CN105718866A (en) | Visual target detection and identification method | |
CN111582126A (en) | Pedestrian Re-identification Method Based on Multi-scale Pedestrian Contour Segmentation and Fusion | |
CN113627481A (en) | Multi-model combined unmanned aerial vehicle garbage classification method for smart gardens | |
CN104036296B (en) | A kind of expression of image and processing method and processing device | |
CN106446890A (en) | Candidate area extraction method based on window scoring and superpixel segmentation | |
CN109003275A (en) | The dividing method of weld defect image | |
CN110008900A (en) | A Region-to-target Candidate Target Extraction Method for Visible Light Remote Sensing Images | |
CN109858342B (en) | A face pose estimation method that combines hand-designed descriptors and deep features |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20130925 |