CN105913096A - Extracting method for disordered image key frame - Google Patents
Extracting method for disordered image key frame Download PDFInfo
- Publication number
- CN105913096A CN105913096A CN201610510776.7A CN201610510776A CN105913096A CN 105913096 A CN105913096 A CN 105913096A CN 201610510776 A CN201610510776 A CN 201610510776A CN 105913096 A CN105913096 A CN 105913096A
- Authority
- CN
- China
- Prior art keywords
- frame
- unordered
- image
- bunch
- cluster
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 16
- 239000000284 extract Substances 0.000 claims abstract description 11
- 238000011156 evaluation Methods 0.000 claims abstract description 6
- 230000004438 eyesight Effects 0.000 claims abstract description 6
- 239000013598 vector Substances 0.000 claims description 21
- 238000010606 normalization Methods 0.000 claims description 9
- 238000001914 filtration Methods 0.000 claims description 6
- 238000013441 quality evaluation Methods 0.000 abstract description 11
- 238000000605 extraction Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000003064 k means clustering Methods 0.000 description 2
- 201000004569 Blindness Diseases 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Library & Information Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Multimedia (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Image Analysis (AREA)
Abstract
一种无序图像关键帧的提取方法,其特征在于,包括以下步骤:1)采用不设K值的聚簇算法对无序图像进行聚簇处理,把图像信息内容相近的无序图像聚为一簇;2)根据相似距离求解每簇的聚簇中心,从每簇中把距离聚簇中心最近的无序图像作为代表帧Fk提取出来;3)对提取出的代表帧Fk进行无参考图像质量评价,如果该代表帧Fk满足立体视觉三维重建质量要求则作为关键帧保留,如果该代表帧Fk不满足立体视觉重建质量要求,则从无序图像中删除,本发明可以有效的提取无序图像的代表帧,滤除信息冗余、信息量少、质量差的无序图像,并且采用不设K值的聚簇算法可以根据无序图像内容复杂度自动聚成不同数目的簇,通过采用无参考图像评价法可以获得高清晰的关键帧。
A method for extracting a key frame of an unordered image, characterized in that it comprises the following steps: 1) adopting a clustering algorithm that does not set a K value to cluster the unordered image, and gathering unordered images with similar image information content into 2) Solve the clustering center of each cluster according to the similarity distance, and extract the disordered image closest to the clustering center from each cluster as the representative frame F k ; With reference to the image quality evaluation, if the representative frame F k satisfies the stereo vision three-dimensional reconstruction quality requirements, then it is reserved as a key frame; if the representative frame F k does not meet the stereo vision reconstruction quality requirements, then it is deleted from the disordered image, and the present invention can effectively Extract representative frames of unordered images, filter out unordered images with redundant information, less information, and poor quality, and use the clustering algorithm without K value to automatically cluster into different numbers according to the complexity of the content of unordered images Clusters, high-definition key frames can be obtained by using the no-reference image evaluation method.
Description
技术领域technical field
本发明涉及一种计算机视觉及图像图形领域,尤其是一种无序图像关键帧的提取方法。The invention relates to the fields of computer vision and image graphics, in particular to a method for extracting key frames of disordered images.
背景技术Background technique
近年来人们对室外场景对象立体视觉三维重建的研究方兴未艾。然而用手持相机对场景对象从不同视觉、位置采集图像进行三维重建,由于采集的无序图像数目多、内容冗余大、个别图像信息量少、图像模糊等原因,给三维重建带来困难,甚至采集的图像不能进行重建。In recent years, the research on stereoscopic 3D reconstruction of outdoor scene objects is in the ascendant. However, using a hand-held camera to perform 3D reconstruction of images collected from different visions and positions of scene objects brings difficulties to 3D reconstruction due to the large number of disordered images collected, large content redundancy, small amount of individual image information, and blurred images. Even the acquired images cannot be reconstructed.
对无序图像进行关键帧提取,可以提取出图像的代表帧,滤除信息冗余、信息量少的图像,现在一般用的关键帧提取方法有:基于镜头边界的关键帧提取方法,该方法提取的关键帧不具有代表性;基于运动分析的关键帧提取方法,该方法运算复杂度比较大;基于无序图像K均值聚簇算法提取关键帧,把无序图像聚成K簇,从每簇中提取出离聚簇中心最近的一帧作为关键桢,该算法是三种方法中最好的。但无序图像K均值聚类算法有两个问题,一、运用该算法之前必须提前设定K值,无序图像聚K个簇,从每个簇中提取一个关键帧,往往事先我们并不知道无序图像中有多少个关键帧,直接设K值有一定的盲目性,有些关键帧可能不能有效的提取出来。二、提取的关键帧可能存在模糊不清晰问题,不能满足三维重建、目标检测、识别的质量要求。针对以上两个问题本发明专利提出不设K值聚簇算法和无参考图像质量评价算法可以很好的解决以上两个问题,该算法实行分层提取:第一层采用不舍K值的聚簇算法根据无序图像内容复杂度来决定簇的数目即关键帧的数目。第二层把第一层提取出的关键帧进行无参考图像质量评价,提取出的关键帧清晰度满足三维重建需要则保留,不满足则重新从原来的簇中提取关键帧,再次进行无参考图像质量评价,直到评价后的关键帧满足立体视觉三维重建质量要求为止。Extracting key frames from unordered images can extract the representative frames of the image and filter out images with redundant information and less information. Now the key frame extraction methods generally used are: the key frame extraction method based on the shot boundary, this method The extracted key frames are not representative; the key frame extraction method based on motion analysis has a relatively large computational complexity; the key frames are extracted based on the K-means clustering algorithm for unordered images, and the unordered images are clustered into K clusters, from each The frame closest to the center of the cluster is extracted from the cluster as the key frame. This algorithm is the best among the three methods. However, there are two problems with the K-means clustering algorithm for unordered images. First, the K value must be set in advance before using the algorithm. Unordered images gather K clusters, and a key frame is extracted from each cluster. Often we do not know in advance. Knowing how many key frames there are in the unordered image, directly setting the K value has a certain degree of blindness, and some key frames may not be effectively extracted. Second, the extracted key frames may be blurred and unclear, which cannot meet the quality requirements of 3D reconstruction, target detection, and recognition. In view of the above two problems, the patent of the present invention proposes a clustering algorithm without K value and a no-reference image quality evaluation algorithm, which can solve the above two problems very well. The cluster algorithm determines the number of clusters, that is, the number of key frames, according to the complexity of the unordered image content. In the second layer, the key frames extracted from the first layer are evaluated without reference image quality, and the extracted key frames are retained if they meet the requirements of 3D reconstruction; if not, the key frames are extracted from the original cluster again, and no reference Image quality evaluation until the evaluated key frame meets the quality requirements of stereoscopic 3D reconstruction.
发明内容Contents of the invention
针对现有技术的不足,本发明提供一种无序图像关键帧的提取方法。Aiming at the deficiencies of the prior art, the present invention provides a method for extracting key frames of an unordered image.
本发明的技术方案为:一种无序图像关键帧的提取方法,其特征在于,包括以下步骤:The technical solution of the present invention is: a method for extracting a key frame of an unordered image, characterized in that it comprises the following steps:
1)采用不设K值的聚簇算法对无序图像进行聚簇处理,把图像信息内容相近的无序图像聚为一簇;1) Use the clustering algorithm without setting the K value to cluster the unordered images, and gather the unordered images with similar image information content into one cluster;
2)根据相似距离求解每簇的聚簇中心,从每簇中把距离聚簇中心最近的无序图像作为代表帧Fk提取出来;2) Solve the cluster center of each cluster according to the similarity distance, and extract the disordered image closest to the cluster center from each cluster as the representative frame Fk ;
3)对提取出的代表帧Fk进行无参考图像质量评价,如果该代表帧Fk满足立体视觉三维 重建质量要求,则作为关键帧保留,如果该代表帧Fk不满足立体视觉三维重建质量要求,则删除该代表帧Fk,重新从原来的簇中提取另一代表帧,并将提取出的该另一代表帧再次进行无参考图像质量评价,直到无参考图像质量评价后的代表帧满足立体视角三维重建质量要求为止。3) Carry out no-reference image quality evaluation on the extracted representative frame Fk , if the representative frame Fk meets the quality requirements of stereoscopic 3D reconstruction, it will be reserved as a key frame, if the representative frame Fk does not meet the quality of stereoscopic 3D reconstruction If required, the representative frame F k is deleted, another representative frame is extracted from the original cluster, and the extracted representative frame is subjected to no-reference image quality evaluation again until the representative frame after no-reference image quality evaluation Until the quality requirements of stereoscopic 3D reconstruction are met.
上述技术方案中,步骤1)中不设K值的聚簇算法对无序图像聚簇,将每幅无序图像分成M(M=16)块,每块纹理特征均值ml、方差每块纹理特征均值ml为:In the above-mentioned technical scheme, the clustering algorithm without setting the K value in step 1) clusters the disordered images, and divides each disordered image into M (M=16) blocks, and the texture feature mean value m l and variance of each block The mean value m l of each texture feature is:
每块纹理特征方差为:The variance of each texture feature is:
把每块的纹理特征均值ml、方差合起来作为该无序图像的特征向量F,并对特征向量F进行归一化处理,其中:The texture feature mean m l and variance of each block Take them together as the feature vector F of the unordered image, and normalize the feature vector F, where:
设原始向量[f1,f2,f3,…fM],归一化公式:Suppose the original vector [f 1 ,f 2 ,f 3 ,…f M ], the normalization formula:
其中e、m为原始特征向量标准差和均值,归一化后特征向量为[F1,F2,F3,…FM],任意两帧图像Fa和Fb归一化后的特征向量为:Where e and m are the standard deviation and mean of the original feature vector, the feature vector after normalization is [F 1 , F 2 , F 3 ,...F M ], the normalized features of any two frames of images F a and F b The vector is:
Fa=[Fa1,Fa2,…FaM]和Fb=[Fb1,Fb2,…FbM],F a = [F a1 , F a2 , ... F aM ] and F b = [F b1 , F b2 , ... F bM ],
任意两帧图像Fa和Fb之间的相似距离为:The similarity distance between any two frames of images F a and F b is:
阈值T是任意两张无序图像相似距离之和的平均值,其计算式为:The threshold T is the average value of the sum of similar distances between any two unordered images, and its calculation formula is:
其中,N为无序图像数目;Among them, N is the number of unordered images;
上述技术方案中,步骤1)中不设K值得聚簇步骤为:In above-mentioned technical scheme, in step 1), do not set K to be worth clustering step to be:
a)、获取第一帧图像F1并把其划分到簇K1中,并将第一帧图像F1作为簇K1的聚簇中心;a), obtain the first frame image F1 and divide it into cluster K1, and use the first frame image F1 as the cluster center of cluster K1 ;
b)、获取下一帧图像FI,(I=2,3,…,N),其中,N为无序图像数;b), acquiring the next frame of image F I , (I=2,3,...,N), where N is the number of unordered images;
c)、根据公式计算FI与已得到簇Kj(j=1,2,…,Nc)聚簇中心的相似距离dist(FI,Kjc),其中Kjc为簇Kj的聚簇中心,计算FI与聚簇中心Kjc的最小距离min dist(FI,Kjc),如果min dist(FI,Kjc)≤T,则把FI划分到具有min dist(FI,Kjc)值的簇Kj中,并且计算该簇新的聚簇中心,该簇新的聚簇中心为该簇中所有无序图像归一化后特征向量之和除以该簇中无序图像的数目;c), according to the formula Calculate the similarity distance dist(F I , K jc ) between F I and the cluster center of the obtained cluster K j (j=1,2,…,N c ), where K jc is the cluster center of cluster K j , and calculate F The minimum distance min dist(F I ,K jc ) between I and the cluster center K jc , if min dist(F I ,K jc )≤T, then divide F I into one with min dist(F I ,K jc ) value In the cluster K j of the cluster, and calculate the new cluster center of the cluster, the new cluster center of the cluster is the sum of the normalized feature vectors of all unordered images in the cluster divided by the number of unordered images in the cluster;
d)、如果min dist(FI,Kjc)>T,则说明FI与存在的任何一个簇都不相似,把FI划分到另一个新簇Kj中;d) If min dist(F I , K jc )>T, it means that F I is not similar to any existing cluster, and divide F I into another new cluster K j ;
e)、然后根据步骤b、c、d将所有的无序图像划分到不同的簇Kj中。e), and then divide all unordered images into different clusters K j according to steps b, c, and d.
上述技术方案中,步骤3)中对提取的代表帧Fk进行无参考图像质量评价,其步骤为:In the above-mentioned technical scheme, step 3) carries out no reference image quality evaluation to the extracted representative frame F k , and its steps are:
A)对提取的代表帧Fk为进行二次模糊处理得到模糊图像b;A) performing secondary fuzzy processing on the extracted representative frame Fk to obtain fuzzy image b;
bV=hv*Fk,bH=hh*Fk,hh=(hv)T=hv,b V =h v *F k , b H =h h *F k , h h = (h v ) T = h v ,
其中,bV、bH为图像Fk经过垂直、水平低通滤波后得到的模糊图像,hh、hv为滤波器垂直和水平模型;Among them, b V , b H are blurred images obtained by image F k after vertical and horizontal low-pass filtering, h h , h v are filter vertical and horizontal models;
B)分别计算代表帧Fk滤波前相邻像素垂直绝对误差DfV(i,j)、水平绝对误差DfH(i,j)和滤波后模糊图像b相邻像素的垂直绝对误差DbV(i,j)、水平绝对误差DbH(i,j),得到相邻像素值的变化;B) Calculate the vertical absolute error Df V (i, j) and the horizontal absolute error Df H (i, j) of the adjacent pixels representing the frame F k before filtering, and the vertical absolute error Db V ( i, j), horizontal absolute error Db H (i, j), to obtain the change of adjacent pixel values;
DFV(i,j)=abs(Fk(i,j)-Fk(i-1,j)),DFH(i,j)=abs(Fk(i,j)-Fk(i,j-1)),DF V (i,j)=abs(F k (i,j)-F k (i-1,j)), DF H (i,j)=abs(F k (i,j)-F k ( i,j-1)),
DbV(i,j)=abs(bV(i,j)-bV(i-1,j)),DbH(i,j)=abs(bH(i,j)-bH(i,j-1));Db V (i,j)=abs(b V (i,j)-b V (i-1,j)), Db H (i,j)=abs(b H (i,j)-b H ( i,j-1));
C)对代表帧Fk中相邻像素差进行求和处理:C) Summing the difference between adjacent pixels in the representative frame F k :
归一化得:Normalized to get:
clear=max(bFV,bFH);clear=max(bF V ,bF H );
D)、评价值范围在(0,1)之间,如果0.3<clear=max(bFV,bFH)<1,将代表帧Fk作为关键帧提取出来,如果clear=max(bFV,bFH)≤0.3,则删除该代表帧,并从原簇中重新提取代表帧,然后根据步骤A、B、C计算该重新提取的代表帧。D), the evaluation value range is between (0, 1), if 0.3<clear=max(bF V ,bF H )<1, the representative frame F k is extracted as a key frame, if clear=max(bF V , bF H )≤0.3, delete the representative frame, and re-extract the representative frame from the original cluster, and then calculate the re-extracted representative frame according to steps A, B, and C.
本发明的有益效果为:可以有效的提取无序图像的代表帧,滤除信息冗余、信息量少的无序图像,并且采用不设K值的聚簇算法可以根据无序图像内容复杂度自动聚成不同数目的簇,通过采用无参考图像评价法可以获得高清晰的关键帧。The beneficial effects of the present invention are: the representative frame of the disordered image can be effectively extracted, the disordered image with redundant information and less information can be filtered out, and the clustering algorithm without K value can be used according to the content complexity of the disordered image Automatically gather into different numbers of clusters, and high-definition key frames can be obtained by using the no-reference image evaluation method.
附图说明Description of drawings
图1为本发明的流程图;Fig. 1 is a flow chart of the present invention;
具体实施方式detailed description
下面结合附图对本发明的具体实施方式作进一步说明:The specific embodiment of the present invention will be further described below in conjunction with accompanying drawing:
如图1所示,一种无序图像关键帧的提取方法,其特征在于,包括以下步骤:As shown in Figure 1, a method for extracting key frames of an unordered image is characterized in that it comprises the following steps:
1)采用不设K值的聚簇算法对无序图像进行聚簇处理,把图像信息内容相近的无序图像聚为一簇;1) Use the clustering algorithm without setting the K value to cluster the unordered images, and gather the unordered images with similar image information content into one cluster;
2)根据相似距离求解每簇的聚簇中心,从每簇中把距离聚簇中心最近的无序图像作为代表帧Fk提取出来;2) Solve the cluster center of each cluster according to the similarity distance, and extract the disordered image closest to the cluster center from each cluster as the representative frame Fk ;
3)对提取出的代表帧Fk进行无参考图像质量评价,如果该代表帧Fk满足立体视觉三维重建质量要求,则作为关键帧保留,如果该代表帧Fk不满足立体视觉三维重建质量要求,则删除该代表帧Fk,重新从原来的簇中提取另一代表帧,并将提取出的该另一代表帧再次进行 无参考图像质量评价,直到无参考图像质量评价后的代表帧满足立体视角三维重建质量要求为止。3) Carry out no-reference image quality evaluation on the extracted representative frame Fk , if the representative frame Fk meets the quality requirements of stereoscopic 3D reconstruction, it will be reserved as a key frame, if the representative frame Fk does not meet the quality of stereoscopic 3D reconstruction If required, the representative frame F k is deleted, another representative frame is extracted from the original cluster, and the extracted representative frame is subjected to no-reference image quality evaluation again until the representative frame after no-reference image quality evaluation Until the quality requirements of stereoscopic 3D reconstruction are met.
上述技术方案中,步骤1)中不设K值的聚簇算法对无序图像聚簇,将每幅无序图像分成M(M=16)块,每块纹理特征均值ml、方差每块纹理特征均值ml为:In the above-mentioned technical scheme, the clustering algorithm without setting the K value in step 1) clusters the disordered images, and divides each disordered image into M (M=16) blocks, and the texture feature mean value m l and variance of each block The mean value m l of each texture feature is:
每块纹理特征方差为:Variance of texture features per block for:
把每块的纹理特征均值ml、方差合起来作为该无序图像的特征向量F,并对特征向量F进行归一化处理,其中:The texture feature mean m l and variance of each block Take them together as the feature vector F of the unordered image, and normalize the feature vector F, where:
设原始向量[f1,f2,f3,…fM],归一化公式:Suppose the original vector [f 1 ,f 2 ,f 3 ,…f M ], the normalization formula:
其中e、m为原始特征向量标准差和均值,归一化后特征向量为[F1,F2,F3,…FM],任意两帧图像Fa和Fb归一化后的特征向量为:Where e and m are the standard deviation and mean of the original feature vector, the feature vector after normalization is [F 1 , F 2 , F 3 ,...F M ], the normalized features of any two frames of images F a and F b The vector is:
Fa=[Fa1,Fa2,…FaM]和Fb=[Fb1,Fb2,…FbM],F a = [F a1 , F a2 , ... F aM ] and F b = [F b1 , F b2 , ... F bM ],
任意两帧图像Fa和Fb之间的相似距离为:The similarity distance between any two frames of images F a and F b is:
阈值T是任意两张无序图像相似距离之和的平均值,其计算式为:The threshold T is the average value of the sum of similar distances between any two unordered images, and its calculation formula is:
其中,N为无序图像数目;Among them, N is the number of unordered images;
上述技术方案中,步骤1)中不设K值得聚簇步骤为:In above-mentioned technical scheme, in step 1), do not set K to be worth clustering step to be:
a)、获取第一帧图像F1并把其划分到簇K1中,并将第一帧图像F1作为簇K1的聚簇中心;a), obtain the first frame image F1 and divide it into cluster K1, and use the first frame image F1 as the cluster center of cluster K1 ;
b)、获取下一帧图像FI,(I=2,3,…,N),其中,N为无序图像数;b), acquiring the next frame of image F I , (I=2,3,...,N), where N is the number of unordered images;
c)、根据公式计算FI与已得到簇Kj(j=1,2,…,Nc)聚簇中心的相似距离dist(FI,Kjc),其中Kjc为簇Kj的聚簇中心,计算FI与聚簇中心Kjc的最小距离min dist(FI,Kjc),如果min dist(FI,Kjc)≤T,则把FI划分到具有min dist(FI,Kjc)值的簇Kj中,并且计算该簇新的聚簇中心,该簇新的聚簇中心为该簇中所有无序图像归一化后特征向量之和除以该簇中无序图像的数目;c), according to the formula Calculate the similarity distance dist(F I , K jc ) between F I and the cluster center of the obtained cluster K j (j=1,2,…,N c ), where K jc is the cluster center of cluster K j , and calculate F The minimum distance min dist(F I ,K jc ) between I and the cluster center K jc , if min dist(F I ,K jc )≤T, then divide F I into one with min dist(F I ,K jc ) value In the cluster K j of the cluster, and calculate the new cluster center of the cluster, the new cluster center of the cluster is the sum of the normalized feature vectors of all unordered images in the cluster divided by the number of unordered images in the cluster;
d)、如果min dist(FI,Kjc)>T,则说明FI与存在的任何一个簇都不相似,把FI划分到另一个新簇Kj中;d) If min dist(F I , K jc )>T, it means that F I is not similar to any existing cluster, and divide F I into another new cluster K j ;
e)、然后根据步骤b、c、d将所有的无序图像划分到不同的簇Kj中。e), and then divide all unordered images into different clusters K j according to steps b, c, and d.
上述技术方案中,步骤3)中对提取的代表帧Fk进行无参考图像质量评价,其步骤为:In the above-mentioned technical scheme, step 3) carries out no reference image quality evaluation to the extracted representative frame F k , and its steps are:
A)对提取的关键代表帧Fk为进行二次模糊处理得到模糊图像b;A) performing secondary fuzzy processing on the extracted key representative frame Fk to obtain fuzzy image b;
bV=hv*Fk,bH=hh*Fk,hh=(hv)T=hv,b V =h v *F k , b H =h h *F k , h h = (h v ) T = h v ,
其中,bV、bH为图像Fk经过垂直、水平低通滤波后得到的模糊图像,hh、hv为滤波器垂直和水平模型;Among them, b V , b H are blurred images obtained by image F k after vertical and horizontal low-pass filtering, h h , h v are filter vertical and horizontal models;
B)分别计算代表帧Fk滤波前相邻像素垂直绝对误差DfV(i,j)、水平绝对误差DfH(i,j)和滤波后模糊图像b相邻像素的垂直绝对误差DbV(i,j)、水平绝对误差DbH(i,j),得到相邻像素值的变化;B) Calculate the vertical absolute error Df V (i, j) and the horizontal absolute error Df H (i, j) of the adjacent pixels representing the frame F k before filtering, and the vertical absolute error Db V ( i, j), horizontal absolute error Db H (i, j), to obtain the change of adjacent pixel values;
DFV(i,j)=abs(Fk(i,j)-Fk(i-1,j)),DFH(i,j)=abs(Fk(i,j)-Fk(i,j-1)),DF V (i,j)=abs(F k (i,j)-F k (i-1,j)), DF H (i,j)=abs(F k (i,j)-F k ( i,j-1)),
DbV(i,j)=abs(bV(i,j)-bV(i-1,j)),DbH(i,j)=abs(bH(i,j)-bH(i,j-1));Db V (i,j)=abs(b V (i,j)-b V (i-1,j)), Db H (i,j)=abs(b H (i,j)-b H ( i,j-1));
C)对代表帧Fk中相邻像素差求和处理:C) Summing the difference between adjacent pixels in the representative frame F k :
归一化得:Normalized to get:
clear=max(bFV,bFH);clear=max(bF V ,bF H );
D)、评价值范围在(0,1)之间,如果0.3<clear=max(bFV,bFH)<1,将代表帧Fk作为关键帧提取出来,如果clear=max(bFV,bFH)≤0.3,则删除该代表帧,并从原簇中重新提取代表帧,然后根据步骤A、B、C计算该重新提取的代表帧。D), the evaluation value range is between (0, 1), if 0.3<clear=max(bF V , bF H )<1, the representative frame F k will be extracted as a key frame, if clear=max(bF V , bF H )≤0.3, delete the representative frame, and re-extract the representative frame from the original cluster, and then calculate the re-extracted representative frame according to steps A, B, and C.
本发明的有益效果为:可以有效的提取无序图像的代表帧,滤除信息冗余、信息量少、质量差的无序图像,并且采用不设K值的聚簇算法可以根据无序图像内容复杂度自动聚成不同数目的簇,通过采用无参考图像评价法可以获得高清晰的关键帧。The beneficial effects of the present invention are: the representative frame of the disordered image can be effectively extracted, the disordered image with redundant information, less information, and poor quality can be filtered out, and the clustering algorithm without K value can be used to filter out the disordered image according to the The content complexity is automatically clustered into different numbers of clusters, and high-definition key frames can be obtained by using the no-reference image evaluation method.
上述实施例和说明书中描述的只是说明本发明的原理和最佳实施例,在不脱离本发明精神和范围的前提下,本发明还会有各种变化和改进,这些变化和改进都落入要求保护的本发明范围内。What described in above-mentioned embodiment and description just illustrate the principle of the present invention and preferred embodiment, under the premise of not departing from the spirit and scope of the present invention, the present invention also can have various changes and improvements, and these changes and improvements all fall into within the scope of the claimed invention.
Claims (3)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610510776.7A CN105913096A (en) | 2016-06-29 | 2016-06-29 | Extracting method for disordered image key frame |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610510776.7A CN105913096A (en) | 2016-06-29 | 2016-06-29 | Extracting method for disordered image key frame |
Publications (1)
Publication Number | Publication Date |
---|---|
CN105913096A true CN105913096A (en) | 2016-08-31 |
Family
ID=56754092
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610510776.7A Pending CN105913096A (en) | 2016-06-29 | 2016-06-29 | Extracting method for disordered image key frame |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105913096A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108668169A (en) * | 2018-06-01 | 2018-10-16 | 北京市商汤科技开发有限公司 | Image information processing method and device, storage medium |
CN108875517A (en) * | 2017-12-15 | 2018-11-23 | 北京旷视科技有限公司 | Method for processing video frequency, device and system and storage medium |
CN111223110A (en) * | 2020-01-06 | 2020-06-02 | 陈根生 | Microscopic image enhancement method and device and computer equipment |
CN113850299A (en) * | 2021-09-01 | 2021-12-28 | 浙江爱达科技有限公司 | Gastrointestinal tract capsule endoscopy video key frame extraction method capable of self-adapting to threshold |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103336832A (en) * | 2013-07-10 | 2013-10-02 | 中国科学院自动化研究所 | Video classifier construction method based on quality metadata |
-
2016
- 2016-06-29 CN CN201610510776.7A patent/CN105913096A/en active Pending
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103336832A (en) * | 2013-07-10 | 2013-10-02 | 中国科学院自动化研究所 | Video classifier construction method based on quality metadata |
Non-Patent Citations (2)
Title |
---|
李全栋等: "一种改进的无监督聚类的关键帧提取算法", 《应用光学》 * |
李鸿林等: "无参考模糊图像质量评价改进算法", 《计算机应用》 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108875517A (en) * | 2017-12-15 | 2018-11-23 | 北京旷视科技有限公司 | Method for processing video frequency, device and system and storage medium |
CN108668169A (en) * | 2018-06-01 | 2018-10-16 | 北京市商汤科技开发有限公司 | Image information processing method and device, storage medium |
CN108668169B (en) * | 2018-06-01 | 2021-10-29 | 北京市商汤科技开发有限公司 | Image information processing method and device, and storage medium |
CN111223110A (en) * | 2020-01-06 | 2020-06-02 | 陈根生 | Microscopic image enhancement method and device and computer equipment |
CN111223110B (en) * | 2020-01-06 | 2023-07-18 | 陈根生 | Microscopic image enhancement method and device and computer equipment |
CN113850299A (en) * | 2021-09-01 | 2021-12-28 | 浙江爱达科技有限公司 | Gastrointestinal tract capsule endoscopy video key frame extraction method capable of self-adapting to threshold |
CN113850299B (en) * | 2021-09-01 | 2024-05-14 | 浙江爱达科技有限公司 | Gastrointestinal capsule endoscope video key frame extraction method with self-adaptive threshold |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102622769B (en) | Multi-target tracking method by taking depth as leading clue under dynamic scene | |
CN105740945B (en) | A kind of people counting method based on video analysis | |
CN106709472A (en) | Video target detecting and tracking method based on optical flow features | |
CN102098526B (en) | Depth map calculating method and device | |
WO2018000752A1 (en) | Monocular image depth estimation method based on multi-scale cnn and continuous crf | |
CN116258817B (en) | A method and system for constructing autonomous driving digital twin scenes based on multi-view three-dimensional reconstruction | |
CN103955945B (en) | Self-adaption color image segmentation method based on binocular parallax and movable outline | |
CN109117774B (en) | Multi-view video anomaly detection method based on sparse coding | |
CN101996406A (en) | No-reference structure sharpness image quality assessment method | |
CN103426182A (en) | Electronic image stabilization method based on visual attention mechanism | |
CN106611427A (en) | A video saliency detection method based on candidate area merging | |
CN102034267A (en) | Three-dimensional reconstruction method of target based on attention | |
CN105404888A (en) | Saliency object detection method integrated with color and depth information | |
CN105913096A (en) | Extracting method for disordered image key frame | |
CN107633495A (en) | A kind of infrared polarization based on complementary relationship and the more embedded fusion methods of algorithm 2D VMD of intensity image | |
CN108470178B (en) | A depth map saliency detection method combined with depth reliability evaluation factor | |
CN110738076A (en) | People counting method and system in images | |
CN106529441B (en) | Depth motion figure Human bodys' response method based on smeared out boundary fragment | |
CN103150722A (en) | Method for extracting peripheral blood leucocyte edges with application of quaternion division and graph theory optimization | |
CN109241814A (en) | Pedestrian detection method based on YOLO neural network | |
CN117351570A (en) | A human body behavior analysis method and system based on joint points | |
CN105225233B (en) | A dense stereo image matching method and system based on two types of dilation | |
CN108154150B (en) | A saliency detection method based on background priors | |
CN116524000A (en) | Rose size measurement method and system based on deep learning and stereoscopic vision | |
CN107169413A (en) | A kind of human facial expression recognition method of feature based block weight |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20160831 |