CN105721860B - A kind of Methods for Shot Boundary Detection of Video Sequences based on hypergraph - Google Patents
A kind of Methods for Shot Boundary Detection of Video Sequences based on hypergraph Download PDFInfo
- Publication number
- CN105721860B CN105721860B CN201610037058.2A CN201610037058A CN105721860B CN 105721860 B CN105721860 B CN 105721860B CN 201610037058 A CN201610037058 A CN 201610037058A CN 105721860 B CN105721860 B CN 105721860B
- Authority
- CN
- China
- Prior art keywords
- mrow
- summit
- vector
- shot
- hypergraph
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N17/00—Diagnosis, testing or measuring for television systems or their details
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Analysis (AREA)
Abstract
一种基于超图的视频镜头边界检测方法:提取视频特征;构建k近邻超图模型;寻找视频镜头边界,包括:根据k近邻超图模型设定得分向量的最小得分阈值;令查询顶点的标签向量为1,其他顶点的标签向量设置为0;计算各顶点得分向量,并且记录得分向量大于最小得分阈值中的连续顶点位置,并随机选取一半以内的顶点记为集合,且令标签向量为1;根据标签向量,再次计算各顶点得分向量;从令查询顶点的标签向量为1,其他顶点的标签向量设置为0重复进行;确定视频镜头边界及类型;再从令查询顶点的标签向量为1,其他顶点的标签向量设置为0重复进行,直到所有视频镜头边界确定完毕。本发明通过超图模型对视频帧的检索,将超图模型用于视频镜头边界的检测领域。
A hypergraph-based video shot boundary detection method: extracting video features; constructing a k-nearest neighbor hypergraph model; finding video shot boundaries, including: setting the minimum score threshold of the score vector according to the k-nearest neighbor hypergraph model; making the label of the query vertex The vector is 1, and the label vectors of other vertices are set to 0; calculate the score vector of each vertex, and record the continuous vertex positions where the score vector is greater than the minimum score threshold, and randomly select half of the vertices as a set, and set the label vector to 1 ;According to the label vector, calculate the score vector of each vertex again; set the label vector of the query vertex to 1, and set the label vector of other vertices to 0 to repeat; determine the boundary and type of the video shot; then make the label vector of the query vertex 1 , the label vectors of other vertices are set to 0 and repeated until the boundaries of all video shots are determined. The invention uses the hypergraph model to retrieve video frames through the hypergraph model, and uses the hypergraph model in the detection field of video shot boundaries.
Description
技术领域technical field
本发明涉及一种视频镜头边界检测方法。特别是涉及一种用同一视频镜头中内容具有的连续性与相似性的特性,通过超图模型对视频镜头内图像帧的分析,确定视频各个镜头边界的基于超图的视频镜头边界检测方法。The invention relates to a video shot boundary detection method. In particular, it relates to a hypergraph-based video shot boundary detection method that uses the continuity and similarity characteristics of content in the same video shot to analyze the image frames in the video shot through a hypergraph model to determine the boundaries of each shot in the video.
背景技术Background technique
视频镜头通常指摄像机一次连续拍摄的视频片段,视频镜头边界通常指视频镜头相邻帧之间出现了某种意义上的变化。视频镜头边界检测是用来寻找多个连续镜头之间边界的一种技术。当两个视频镜头发生转换时,通常会出现一些明显的变化,例如颜色特征的变化等。A video shot usually refers to a video segment shot continuously by a camera at one time, and a video shot boundary usually refers to a change in a certain sense between adjacent frames of a video shot. Video shot boundary detection is a technique used to find the boundaries between multiple consecutive shots. When two video shots transition, there are usually some noticeable changes, such as changes in color characteristics, etc.
视频镜头边界通常包含两种类型。一种是突变镜头(abrupt shot),指视频帧从一个镜头突然跳到另一个镜头。突变镜头通常发生在两帧之间,前后两帧分别属于前后两个镜头。另一种是渐变镜头(gradual shot),指视频帧从一个镜头逐渐缓慢的过渡到另一个镜头,镜头会有一种在时间和空间上的编辑效果。渐变镜头通常包括渐入渐出(fade inand fade out)、叠化(dissolve)等。渐变镜头通常发生在几帧到十几帧之间,是前后两个镜头的过渡。其中突变镜头较容易检测,而渐变镜头较难检测,是镜头边界检测的重点。Video shot boundaries generally contain two types. One is an abrupt shot, which refers to a video frame suddenly jumping from one shot to another. Mutation shots usually occur between two frames, and the two frames before and after belong to the two shots before and after. The other is a gradual shot, which refers to the gradual and slow transition of video frames from one shot to another, and the shot will have an editing effect in time and space. Gradient shots usually include fade in and fade out, dissolve, etc. Gradient shots usually occur between a few frames to a dozen frames, and are the transition between the two shots before and after. Among them, the sudden change lens is easier to detect, while the gradual change lens is more difficult to detect, which is the focus of lens boundary detection.
视频镜头边界检测主要包括视频帧的特征提取、视频特征间的相似性度量、镜头边界确定三个步骤。现有的视频镜头边界检测技术主要有基于边缘的方法和基于运动信息的方法。图像的边缘和梯度信息可以很好的表现图像的视觉信息,因此可以作为视频图像的特征使用。基于边缘的视频镜头检测方法通常对相机、物体的运动比较敏感,因此对渐变镜头的检测不是十分准确。基于运动信息的方法是基于镜头内视频帧是光滑的、边界处是突变的基本假设。因此基于运动信息的方法对突变镜头的检测较为准确,对于渐变镜头的检测也不是十分准确。基于运动信息的方法还有计算时间复杂度高的缺点。目前视频镜头边界检测技术最大的挑战是如何检测渐变镜头以及如何消除光照或者相机和物体的高速运动对视频镜头边界处的影响。Video shot boundary detection mainly includes three steps: feature extraction of video frames, similarity measurement between video features, and shot boundary determination. The existing video shot boundary detection techniques mainly include edge-based methods and motion information-based methods. The edge and gradient information of the image can well represent the visual information of the image, so it can be used as the feature of the video image. Edge-based video shot detection methods are usually sensitive to camera and object motion, so the detection of progressive shots is not very accurate. The method based on motion information is based on the basic assumption that the video frame in the shot is smooth and the boundary is abrupt. Therefore, the method based on motion information is more accurate for the detection of sudden changes, and it is not very accurate for the detection of gradual changes. The method based on motion information also has the disadvantage of high computational time complexity. At present, the biggest challenge of video shot boundary detection technology is how to detect progressive shots and how to eliminate the impact of lighting or high-speed motion of cameras and objects on video shot boundaries.
发明内容Contents of the invention
本发明所要解决的技术问题是,提供一种能够利用超图模型逐一确定视频镜头边界的基于超图的视频镜头边界检测方法。The technical problem to be solved by the present invention is to provide a hypergraph-based video shot boundary detection method that can determine video shot boundaries one by one by using a hypergraph model.
本发明所采用的技术方案是:一种基于超图的视频镜头边界检测方法,包括如下步骤:The technical scheme adopted in the present invention is: a kind of video shot boundary detection method based on hypergraph, comprises the following steps:
1)提取视频特征;1) Extract video features;
2)构建k近邻超图模型:2) Construct k-nearest neighbor hypergraph model:
其中,v代表超图的顶点,e代表超图的超边,H(v,e)为关联矩阵;Among them, v represents the vertex of the hypergraph, e represents the hyperedge of the hypergraph, and H(v, e) is the correlation matrix;
3)寻找视频镜头边界,包括:3) Find the boundary of the video lens, including:
(1)首先根据k近邻超图模型,计算给定一个查询顶点时,其他顶点相对于该顶点的得分向量:(1) First, according to the k-nearest neighbor hypergraph model, when a query vertex is given, the score vectors of other vertices relative to the vertex are calculated:
f=(1-γ)(I-γΘ)-1y (2)f=(1-γ)(I-γΘ) -1 y (2)
其中,γ为常量系数;I为单位矩阵;y为超图n个顶点的标签向量,维度是n×1维,f为相对于查询点的得分向量,维度是n×1维;Θ为n×n的拉普拉斯矩阵,Θ=Dv -1/2HWDe - 1HTDv -1/2,其中,W为权重矩阵,是以边的权重w(e)为对角线构成超边的权重矩阵:Among them, γ is a constant coefficient; I is an identity matrix; y is the label vector of n vertices of the hypergraph, and its dimension is n×1 dimension; f is the score vector relative to the query point, its dimension is n×1 dimension; Θ is n ×n Laplacian matrix, Θ=D v -1/2 HWD e - 1 H T D v -1/2 , where W is the weight matrix, and the weight w(e) of the edge is the diagonal The weight matrix that makes up the hyperedge:
Dv为顶点的度矩阵,是以顶点的度d(v)为对角线构成顶点的度矩阵:D v is the degree matrix of the vertex, and the degree matrix of the vertex is formed by taking the degree d(v) of the vertex as the diagonal:
d(v)=∑e∈Ew(e)·H(v,e) (4)d(v)=∑ e∈E w(e)·H(v,e) (4)
De为超边的度矩阵,是以边的度d(e)为对角线构成超边的度矩阵:D e is the degree matrix of the hyperedge, and the degree matrix of the hyperedge is formed by taking the degree d(e) of the edge as the diagonal:
d(e)=∑v∈eH(v,e) (5)d(e)= ∑v∈e H(v,e) (5)
(2)设定得分向量f的最小得分阈值δ;(2) Set the minimum score threshold δ of the score vector f;
(3)令查询顶点的标签向量y(j)=1,其他顶点的标签向量设置为0;(3) Let the label vector y(j) of the query vertex=1, and the label vectors of other vertices are set to 0;
(4)计算各顶点得分向量f,并且记录包含第j个位置在内的f>δ的连续顶点位置,并从所述的连续顶点位置中随机选取一半以内的顶点作为反馈点,记为集合F′k,且令标签向量y(F′k)=1,k为集合的标号;(4) Calculate the score vector f of each vertex, and record the continuous vertex positions of f>δ including the jth position, and randomly select the vertices within half of the continuous vertex positions as feedback points, and record them as a set F' k , and let the label vector y(F' k )=1, k is the label of the set;
(5)根据标签向量y(F′k)=1,再次计算各顶点得分向量f,并且记录包含集合F′k的f>δ的连续顶点位置,记为集合Fk,此时Fk代表包含第j个位置的同一镜头内的所有帧;(5) According to the label vector y(F′ k )=1, calculate the score vector f of each vertex again, and record the continuous vertex positions of f>δ including the set F′ k , which is recorded as the set F k , and F k represents All frames within the same shot containing the jth position;
(6)令j=Fk(last)+1,k=k+1,设置标签向量y(j)=1,计算各顶点得分向量f,并且记录包含第j个位置在内的f>δ的连续顶点位置,并从所述的连续顶点位置中随机选取一半以内的顶点作为反馈点,记为集合F′k+1,且令标签向量y(F′k+1)=1,根据标签向量y(F′k+1)=1,再次计算各顶点得分向量f,并且记录包含集合F′k+1的f>δ的连续顶点位置,得到集合Fk+1,其中,Fk(last)为集合Fk的最后一个顶点;(6) Let j=F k (last)+1, k=k+1, set the label vector y(j)=1, calculate the score vector f of each vertex, and record f>δ including the jth position The continuous vertex positions of , and randomly select the vertices within half of the continuous vertex positions as the feedback points, which are recorded as the set F′ k+1 , and let the label vector y(F′ k+1 )=1, according to the label Vector y(F′ k+1 )=1, calculate the score vector f of each vertex again, and record the continuous vertex positions of f>δ including the set F′ k+1 , and obtain the set F k+1 , where, F k ( last) is the last vertex of the set F k ;
(7)确定视频镜头边界及类型;(7) Determine the boundary and type of the video shot;
(8)令Fk=Fk+1,返回第(6)步,直到所有视频镜头边界确定完毕。(8) Set F k =F k+1 , return to step (6), until all video shot boundaries are determined.
步骤3)中第(3)步表示顶点的标签向量y=[0,…,1,…0]T,其中1在第j个位置。Step (3) in step 3) represents the label vector y=[0,...,1,...0] T of the vertices, where 1 is at the jth position.
步骤3)中第(7)步所述的确定视频镜头边界及类型包括:Step 3) in step (7) described in determining the boundary of video lens and type include:
取集合Fk与集合Fk+1的交集F△,即F△=Fk∩Fk+1,若那么确定在第Fk(last)和Fk+1(first)点处为一个镜头边界,且该镜头边界为突变镜头,即第Fk(last)和Fk+1(first)点分别为突变镜头的上下边界;若那么可以确定在第F△(first)和F△(last)点处为一个镜头边界,且该镜头边界为渐变镜头,即第F△(first)和F△(last)点分别为渐变镜头的上下边界,其中,Fk+1(first)为集合Fk+1的第一个顶点,F△(first)为集合F△的第一个顶点,F△(last)为集合F△的最后一个顶点。Take the intersection F △ of the set F k and the set F k+1 , that is, F △ =F k ∩F k+1 , if Then it is determined that the F k (last) and F k+1 (first) points are a shot boundary, and the shot boundary is a sudden change shot, that is, the F k (last) and F k+1 (first) points are respectively The upper and lower bounds of the mutation lens; if Then it can be determined that the F △ (first) and F △ (last) points are a lens boundary, and the lens boundary is a progressive lens, that is, the F △ (first) and F △ (last) points are respectively the progressive lens Upper and lower boundaries, where F k+1 (first) is the first vertex of the set F k+1 , F △ (first) is the first vertex of the set F △ , F △ (last) is the last of the set F △ a vertex.
本发明的一种基于超图的视频镜头边界检测方法,通过超图模型对视频帧的检索,将超图模型用于视频镜头边界的检测领域,具有以下显著特点:A hypergraph-based video shot boundary detection method of the present invention uses the hypergraph model to retrieve video frames through the hypergraph model, and uses the hypergraph model for the detection field of video shot boundaries, and has the following salient features:
1、本发明首次将超图模型应用到镜头边界检测上面,是通过将一个镜头内所有视频帧都检索出来从而来确定视频镜头边界,不同与以往通过镜头边界处颜色或运动信息的变化来寻找镜头边界的思路。1. The present invention applies the hypergraph model to the shot boundary detection for the first time, and determines the video shot boundary by retrieving all the video frames in a shot, which is different from searching through the change of color or motion information at the shot boundary in the past. The idea of lens boundaries.
2、经实验验证,本发明可以快速有效的将视频镜头边界检测出来,因此是一种有效的视频镜头边界检测方法。2. It is verified by experiments that the present invention can quickly and effectively detect the boundary of video shots, so it is an effective method for detecting the boundary of video shots.
3、本发明简单易行,效果优良。可以用在视频分析、视频摘要、视频检索等领域的预处理阶段。3. The present invention is simple and easy to implement, and has excellent effect. It can be used in the preprocessing stage of video analysis, video summarization, video retrieval and other fields.
附图说明Description of drawings
图1是本发明基于超图的视频镜头边界检测方法的流程图;Fig. 1 is the flow chart of the video shot boundary detection method based on hypergraph of the present invention;
图2是一个突变镜头(视频帧为连续视频),其中5帧和6帧分别为前后两个镜头的边界;Figure 2 is a sudden change shot (the video frame is a continuous video), where 5 frames and 6 frames are respectively the boundaries of the front and back shots;
图3是一个渐变镜头(视频帧为连续视频),其中6帧到9帧为前后两个镜头的边界范围。Fig. 3 is a progressive shot (the video frame is a continuous video), wherein 6 frames to 9 frames are the boundary ranges of the front and rear shots.
具体实施方式detailed description
下面结合实施例和附图对本发明的一种基于超图的视频镜头边界检测方法做出详细说明。A hypergraph-based video shot boundary detection method of the present invention will be described in detail below with reference to the embodiments and the accompanying drawings.
本发明的一种基于超图的视频镜头边界检测方法,包括如下步骤:A kind of video shot boundary detection method based on hypergraph of the present invention, comprises the steps:
1)提取视频特征;1) Extract video features;
2)构建k近邻超图模型:2) Construct k-nearest neighbor hypergraph model:
其中,v代表超图的顶点,e代表超图的超边,H(v,e)为关联矩阵;Among them, v represents the vertex of the hypergraph, e represents the hyperedge of the hypergraph, and H(v, e) is the correlation matrix;
3)寻找视频镜头边界,包括:3) Find the boundary of the video lens, including:
(1)首先根据k近邻超图模型,计算给定一个查询顶点时,其他顶点相对于该顶点的得分向量:(1) First, according to the k-nearest neighbor hypergraph model, when a query vertex is given, the score vectors of other vertices relative to the vertex are calculated:
f=(1-γ)(I-γΘ)-1y (2)f=(1-γ)(I-γΘ) -1 y (2)
其中,γ为常量系数;I为单位矩阵;y为超图n个顶点的标签向量,维度是n×1维,f为相对于查询点的得分向量,维度是n×1维;Θ为n×n的拉普拉斯矩阵,Θ=Dv -1/2HWDe - 1HTDv -1/2,其中,W为权重矩阵,是以边的权重w(e)为对角线构成超边的权重矩阵:Among them, γ is a constant coefficient; I is an identity matrix; y is the label vector of n vertices of the hypergraph, and its dimension is n×1 dimension; f is the score vector relative to the query point, its dimension is n×1 dimension; Θ is n ×n Laplacian matrix, Θ=D v -1/2 HWD e - 1 H T D v -1/2 , where W is the weight matrix, and the weight w(e) of the edge is the diagonal The weight matrix that makes up the hyperedge:
Dv为顶点的度矩阵,是以顶点的度d(v)为对角线构成顶点的度矩阵:D v is the degree matrix of the vertex, and the degree matrix of the vertex is formed by taking the degree d(v) of the vertex as the diagonal:
d(v)=∑e∈Ew(e)·H(v,e) (4)d(v)=∑ e∈E w(e)·H(v,e) (4)
De为超边的度矩阵,是以边的度d(e)为对角线构成超边的度矩阵:D e is the degree matrix of the hyperedge, and the degree matrix of the hyperedge is formed by taking the degree d(e) of the edge as the diagonal:
d(e)=∑v∈eH(v,e) (5)d(e)= ∑v∈e H(v,e) (5)
(2)设定得分向量f的最小得分阈值δ;(2) Set the minimum score threshold δ of the score vector f;
(3)令查询顶点的标签向量y(j)=1,其他顶点的标签向量设置为0,表示顶点的标签向量y=[0,…,1,…0]T,其中1在第j个位置;(3) Let the label vector y(j) of the query vertex=1, and set the label vector of other vertices to 0, which means the label vector y=[0,…,1,…0] T of the vertex, where 1 is at the jth Location;
(4)计算各顶点得分向量f,并且记录包含第j个位置在内的f>δ的连续顶点位置,并从所述的连续顶点位置中随机选取一半以内的顶点作为反馈点,记为集合F′k,且令标签向量y(F′k)=1,k为集合的标号;(4) Calculate the score vector f of each vertex, and record the continuous vertex positions of f>δ including the jth position, and randomly select the vertices within half of the continuous vertex positions as feedback points, and record them as a set F' k , and let the label vector y(F' k )=1, k is the label of the set;
(5)根据标签向量y(F′k)=1,再次计算各顶点得分向量f,并且记录包含集合F′k的f>δ的连续顶点位置,记为集合Fk,此时Fk代表包含第j个位置的同一镜头内的所有帧;(5) According to the label vector y(F′ k )=1, calculate the score vector f of each vertex again, and record the continuous vertex positions of f>δ including the set F′ k , which is recorded as the set F k , and F k represents All frames within the same shot containing the jth position;
(6)令j=Fk(last)+1,k=k+1,设置标签向量y(j)=1,计算各顶点得分向量f,并且记录包含第j个位置在内的f>δ的连续顶点位置,并从所述的连续顶点位置中随机选取一半以内的顶点作为反馈点,记为集合F′k+1,且令标签向量y(F′k+1)=1,根据标签向量y(F′k+1)=1,再次计算各顶点得分向量f,并且记录包含集合F′k+1的f>δ的连续顶点位置,得到集合Fk+1,其中,Fk(last)为集合Fk的最后一个顶点;(6) Let j=F k (last)+1, k=k+1, set the label vector y(j)=1, calculate the score vector f of each vertex, and record f>δ including the jth position The continuous vertex positions of , and randomly select the vertices within half of the continuous vertex positions as the feedback points, which are recorded as the set F′ k+1 , and let the label vector y(F′ k+1 )=1, according to the label Vector y(F′ k+1 )=1, calculate the score vector f of each vertex again, and record the continuous vertex positions of f>δ including the set F′ k+1 , and obtain the set F k+1 , where, F k ( last) is the last vertex of the set F k ;
(7)确定视频镜头边界及类型,包括:(7) Determine the boundaries and types of video shots, including:
取集合Fk与集合Fk+1的交集F△,即F△=Fk∩Fk+1,若那么确定在第Fk(last)和Fk+1(first)点处为一个镜头边界,且该镜头边界为突变镜头,即第Fk(last)和Fk+1(first)点分别为突变镜头的上下边界;若那么可以确定在第F△(first)和F△(last)点处为一个镜头边界,且该镜头边界为渐变镜头,即第F△(first)和F△(last)点分别为渐变镜头的上下边界,其中,Fk+1(first)为集合Fk+1的第一个顶点,F△(first)为集合F△的第一个顶点,F△(last)为集合F△的最后一个顶点;Take the intersection F △ of the set F k and the set F k+1 , that is, F △ =F k ∩F k+1 , if Then it is determined that the F k (last) and F k+1 (first) points are a shot boundary, and the shot boundary is a sudden change shot, that is, the F k (last) and F k+1 (first) points are respectively The upper and lower bounds of the mutation lens; if Then it can be determined that the F △ (first) and F △ (last) points are a lens boundary, and the lens boundary is a progressive lens, that is, the F △ (first) and F △ (last) points are respectively the progressive lens Upper and lower boundaries, where F k+1 (first) is the first vertex of the set F k+1 , F △ (first) is the first vertex of the set F △ , F △ (last) is the last of the set F △ a vertex;
(8)令Fk=Fk+1,返回第(6)步,直到所有视频镜头边界确定完毕。(8) Set F k =F k+1 , return to step (6), until all video shot boundaries are determined.
图2和图3是突变镜头和渐变镜头的判断示例。其中图2为突变镜头,记查询点1的结果为集合F1={1,2,3,4,5};那么记查询点F1(last)+1(即点6)的结果为集合F2={6,7,8,9,10};那么可以判定点F1(last)=5和点F2(first)=6为突变镜头的上下界;图3为渐变镜头,记查询点1的结果为集合F1={1,2,3,4,5,6,7,8,9};那么记查询点F1(last)+1(即点10)的结果为集合F2={6,7,8,9,10,11,12,13,14};F△=F1∩F2={6,7,8,9};那么可以判定点F△(first)=6和点F△(last)=9为渐变镜头的上下界。Figure 2 and Figure 3 are examples of judging abrupt shots and progressive shots. Wherein Figure 2 is a sudden change lens, record the result of query point 1 as a set F 1 ={1,2,3,4,5}; then record the result of query point F 1 (last)+1 (ie point 6) as a set F 2 ={6,7,8,9,10}; Then it can be judged that point F 1 (last)=5 and point F 2 (first)=6 are the upper and lower bounds of the sudden change lens; Figure 3 is a gradual change lens, and the result of record query point 1 is the set F 1 ={1,2,3 ,4,5,6,7,8,9}; then record the result of query point F 1 (last)+1 (namely point 10) as set F 2 ={6,7,8,9,10,11, 12,13,14}; F △ =F 1 ∩F 2 ={6,7,8,9}; then it can be determined that point F △ (first)=6 and point F △ (last)=9 are the gradient lens Upper and lower bounds.
本发明的一种基于超图的视频镜头边界检测方法,在具体实施时,可以分为粗粒度范围检测和细粒度范围检测。首先通过粗粒度范围检测确定视频边界的大概范围,节省大量计算时间;然后通过细粒度范围检测来精确视频镜头边界范围。A video shot boundary detection method based on a hypergraph of the present invention can be divided into coarse-grained range detection and fine-grained range detection during specific implementation. First, determine the approximate range of the video boundary through coarse-grained range detection, which saves a lot of computing time; then, fine-grained range detection is used to refine the boundary range of video shots.
粗粒度范围检测步骤如下:The coarse-grained range detection steps are as follows:
1)对视频帧进行采样,例如可以1秒采样两帧;1) Sampling video frames, for example, sampling two frames per second;
2)提取视频的特征;2) Extract the features of the video;
3)构建k近邻超图模型;3) Construct a k-nearest neighbor hypergraph model;
4)按照镜头边界检测方法步骤执行,并根据F△来判断镜头类型和边界范围。4) Execute according to the steps of the shot boundary detection method, and judge the shot type and boundary range according to F △ .
上述粗粒度检测只是在采样帧间进行了检测,只确定了一个大概范围,然后还需进行细粒度范围检测,细粒度范围检测包括:The above coarse-grained detection is only detected between sampling frames, and only a rough range is determined, and then fine-grained range detection is required. Fine-grained range detection includes:
(1)采样粗粒度范围检测结果间的所有视频帧;(1) Sampling all video frames between coarse-grained range detection results;
(2)重复粗粒度范围检测步骤2)、3)、4),从而得到精确的视频镜头边界范围。(2) Repeat the coarse-grained range detection steps 2), 3), and 4), so as to obtain the precise boundary range of the video shot.
Claims (2)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610037058.2A CN105721860B (en) | 2016-01-20 | 2016-01-20 | A kind of Methods for Shot Boundary Detection of Video Sequences based on hypergraph |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610037058.2A CN105721860B (en) | 2016-01-20 | 2016-01-20 | A kind of Methods for Shot Boundary Detection of Video Sequences based on hypergraph |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105721860A CN105721860A (en) | 2016-06-29 |
CN105721860B true CN105721860B (en) | 2017-12-08 |
Family
ID=56147431
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610037058.2A Expired - Fee Related CN105721860B (en) | 2016-01-20 | 2016-01-20 | A kind of Methods for Shot Boundary Detection of Video Sequences based on hypergraph |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105721860B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110874833B (en) * | 2019-10-15 | 2022-03-22 | 合肥工业大学 | SAR image change detection method based on hypergraph matching |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7177470B2 (en) * | 2002-11-13 | 2007-02-13 | Koninklijke Philips Electronics N. V. | Method of and system for detecting uniform color segments |
CN101650830B (en) * | 2009-08-06 | 2012-08-15 | 中国科学院声学研究所 | Combined automatic segmentation method for abrupt change and gradual change of compressed domain video lens |
CN104881681B (en) * | 2015-05-22 | 2018-08-14 | 浙江大学 | Image sequence type labeling based on mixing graph model |
-
2016
- 2016-01-20 CN CN201610037058.2A patent/CN105721860B/en not_active Expired - Fee Related
Also Published As
Publication number | Publication date |
---|---|
CN105721860A (en) | 2016-06-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
An et al. | Multitarget tracking using Siamese neural networks | |
Lee et al. | Key-segments for video object segmentation | |
Zhao et al. | Text from corners: a novel approach to detect text and caption in videos | |
CN105100894B (en) | Face automatic labeling method and system | |
Funk et al. | 2017 ICCV challenge: Detecting symmetry in the wild | |
US8467611B2 (en) | Video key-frame extraction using bi-level sparsity | |
Yi et al. | Human action recognition with graph-based multiple-instance learning | |
CN107301375B (en) | A smoke detection method for video images based on dense optical flow | |
Li et al. | Structuring lecture videos by automatic projection screen localization and analysis | |
Cheng et al. | A direct regression scene text detector with position-sensitive segmentation | |
Liu et al. | A novel shot detection approach based on ORB fused with structural similarity | |
Lou et al. | Extracting primary objects by video co-segmentation | |
Fei et al. | Creating memorable video summaries that satisfy the user’s intention for taking the videos | |
Gozali et al. | Hidden Markov Model for event photo stream segmentation | |
Zhu et al. | A novel nonlinear regression approach for efficient and accurate image matting | |
Narhare et al. | Trademark detection using SIFT features matching | |
Cao et al. | Mining spatiotemporal video patterns towards robust action retrieval | |
CN105721860B (en) | A kind of Methods for Shot Boundary Detection of Video Sequences based on hypergraph | |
Atrish et al. | An automated hierarchical framework for player recognition in sports image | |
Wu et al. | Text detection using delaunay triangulation in video sequence | |
Calandre et al. | Optical Flow Singularities for Sports Video Annotation: Detection of Strokes in Table Tennis. | |
Zhou et al. | Modeling perspective effects in photographic composition | |
Mishra et al. | Automated detection of fighting styles using localized action features | |
Zhai et al. | Video Understanding and Content-Based Retrieval. | |
Majumdar et al. | Video shot detection based on sift features and video summarization using expectation-maximization |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20171208 |