[go: up one dir, main page]

CN105913096A - Extracting method for disordered image key frame - Google Patents

Extracting method for disordered image key frame Download PDF

Info

Publication number
CN105913096A
CN105913096A CN201610510776.7A CN201610510776A CN105913096A CN 105913096 A CN105913096 A CN 105913096A CN 201610510776 A CN201610510776 A CN 201610510776A CN 105913096 A CN105913096 A CN 105913096A
Authority
CN
China
Prior art keywords
frame
unordered
image
bunch
cluster
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610510776.7A
Other languages
Chinese (zh)
Inventor
林靖宇
郑恩
潘莹
曹绍昊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangxi University
Original Assignee
Guangxi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangxi University filed Critical Guangxi University
Priority to CN201610510776.7A priority Critical patent/CN105913096A/en
Publication of CN105913096A publication Critical patent/CN105913096A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Image Analysis (AREA)

Abstract

一种无序图像关键帧的提取方法,其特征在于,包括以下步骤:1)采用不设K值的聚簇算法对无序图像进行聚簇处理,把图像信息内容相近的无序图像聚为一簇;2)根据相似距离求解每簇的聚簇中心,从每簇中把距离聚簇中心最近的无序图像作为代表帧Fk提取出来;3)对提取出的代表帧Fk进行无参考图像质量评价,如果该代表帧Fk满足立体视觉三维重建质量要求则作为关键帧保留,如果该代表帧Fk不满足立体视觉重建质量要求,则从无序图像中删除,本发明可以有效的提取无序图像的代表帧,滤除信息冗余、信息量少、质量差的无序图像,并且采用不设K值的聚簇算法可以根据无序图像内容复杂度自动聚成不同数目的簇,通过采用无参考图像评价法可以获得高清晰的关键帧。

A method for extracting a key frame of an unordered image, characterized in that it comprises the following steps: 1) adopting a clustering algorithm that does not set a K value to cluster the unordered image, and gathering unordered images with similar image information content into 2) Solve the clustering center of each cluster according to the similarity distance, and extract the disordered image closest to the clustering center from each cluster as the representative frame F k ; With reference to the image quality evaluation, if the representative frame F k satisfies the stereo vision three-dimensional reconstruction quality requirements, then it is reserved as a key frame; if the representative frame F k does not meet the stereo vision reconstruction quality requirements, then it is deleted from the disordered image, and the present invention can effectively Extract representative frames of unordered images, filter out unordered images with redundant information, less information, and poor quality, and use the clustering algorithm without K value to automatically cluster into different numbers according to the complexity of the content of unordered images Clusters, high-definition key frames can be obtained by using the no-reference image evaluation method.

Description

一种无序图像关键帧的提取方法A Method for Extracting Key Frames from Unordered Images

技术领域technical field

本发明涉及一种计算机视觉及图像图形领域,尤其是一种无序图像关键帧的提取方法。The invention relates to the fields of computer vision and image graphics, in particular to a method for extracting key frames of disordered images.

背景技术Background technique

近年来人们对室外场景对象立体视觉三维重建的研究方兴未艾。然而用手持相机对场景对象从不同视觉、位置采集图像进行三维重建,由于采集的无序图像数目多、内容冗余大、个别图像信息量少、图像模糊等原因,给三维重建带来困难,甚至采集的图像不能进行重建。In recent years, the research on stereoscopic 3D reconstruction of outdoor scene objects is in the ascendant. However, using a hand-held camera to perform 3D reconstruction of images collected from different visions and positions of scene objects brings difficulties to 3D reconstruction due to the large number of disordered images collected, large content redundancy, small amount of individual image information, and blurred images. Even the acquired images cannot be reconstructed.

对无序图像进行关键帧提取,可以提取出图像的代表帧,滤除信息冗余、信息量少的图像,现在一般用的关键帧提取方法有:基于镜头边界的关键帧提取方法,该方法提取的关键帧不具有代表性;基于运动分析的关键帧提取方法,该方法运算复杂度比较大;基于无序图像K均值聚簇算法提取关键帧,把无序图像聚成K簇,从每簇中提取出离聚簇中心最近的一帧作为关键桢,该算法是三种方法中最好的。但无序图像K均值聚类算法有两个问题,一、运用该算法之前必须提前设定K值,无序图像聚K个簇,从每个簇中提取一个关键帧,往往事先我们并不知道无序图像中有多少个关键帧,直接设K值有一定的盲目性,有些关键帧可能不能有效的提取出来。二、提取的关键帧可能存在模糊不清晰问题,不能满足三维重建、目标检测、识别的质量要求。针对以上两个问题本发明专利提出不设K值聚簇算法和无参考图像质量评价算法可以很好的解决以上两个问题,该算法实行分层提取:第一层采用不舍K值的聚簇算法根据无序图像内容复杂度来决定簇的数目即关键帧的数目。第二层把第一层提取出的关键帧进行无参考图像质量评价,提取出的关键帧清晰度满足三维重建需要则保留,不满足则重新从原来的簇中提取关键帧,再次进行无参考图像质量评价,直到评价后的关键帧满足立体视觉三维重建质量要求为止。Extracting key frames from unordered images can extract the representative frames of the image and filter out images with redundant information and less information. Now the key frame extraction methods generally used are: the key frame extraction method based on the shot boundary, this method The extracted key frames are not representative; the key frame extraction method based on motion analysis has a relatively large computational complexity; the key frames are extracted based on the K-means clustering algorithm for unordered images, and the unordered images are clustered into K clusters, from each The frame closest to the center of the cluster is extracted from the cluster as the key frame. This algorithm is the best among the three methods. However, there are two problems with the K-means clustering algorithm for unordered images. First, the K value must be set in advance before using the algorithm. Unordered images gather K clusters, and a key frame is extracted from each cluster. Often we do not know in advance. Knowing how many key frames there are in the unordered image, directly setting the K value has a certain degree of blindness, and some key frames may not be effectively extracted. Second, the extracted key frames may be blurred and unclear, which cannot meet the quality requirements of 3D reconstruction, target detection, and recognition. In view of the above two problems, the patent of the present invention proposes a clustering algorithm without K value and a no-reference image quality evaluation algorithm, which can solve the above two problems very well. The cluster algorithm determines the number of clusters, that is, the number of key frames, according to the complexity of the unordered image content. In the second layer, the key frames extracted from the first layer are evaluated without reference image quality, and the extracted key frames are retained if they meet the requirements of 3D reconstruction; if not, the key frames are extracted from the original cluster again, and no reference Image quality evaluation until the evaluated key frame meets the quality requirements of stereoscopic 3D reconstruction.

发明内容Contents of the invention

针对现有技术的不足,本发明提供一种无序图像关键帧的提取方法。Aiming at the deficiencies of the prior art, the present invention provides a method for extracting key frames of an unordered image.

本发明的技术方案为:一种无序图像关键帧的提取方法,其特征在于,包括以下步骤:The technical solution of the present invention is: a method for extracting a key frame of an unordered image, characterized in that it comprises the following steps:

1)采用不设K值的聚簇算法对无序图像进行聚簇处理,把图像信息内容相近的无序图像聚为一簇;1) Use the clustering algorithm without setting the K value to cluster the unordered images, and gather the unordered images with similar image information content into one cluster;

2)根据相似距离求解每簇的聚簇中心,从每簇中把距离聚簇中心最近的无序图像作为代表帧Fk提取出来;2) Solve the cluster center of each cluster according to the similarity distance, and extract the disordered image closest to the cluster center from each cluster as the representative frame Fk ;

3)对提取出的代表帧Fk进行无参考图像质量评价,如果该代表帧Fk满足立体视觉三维 重建质量要求,则作为关键帧保留,如果该代表帧Fk不满足立体视觉三维重建质量要求,则删除该代表帧Fk,重新从原来的簇中提取另一代表帧,并将提取出的该另一代表帧再次进行无参考图像质量评价,直到无参考图像质量评价后的代表帧满足立体视角三维重建质量要求为止。3) Carry out no-reference image quality evaluation on the extracted representative frame Fk , if the representative frame Fk meets the quality requirements of stereoscopic 3D reconstruction, it will be reserved as a key frame, if the representative frame Fk does not meet the quality of stereoscopic 3D reconstruction If required, the representative frame F k is deleted, another representative frame is extracted from the original cluster, and the extracted representative frame is subjected to no-reference image quality evaluation again until the representative frame after no-reference image quality evaluation Until the quality requirements of stereoscopic 3D reconstruction are met.

上述技术方案中,步骤1)中不设K值的聚簇算法对无序图像聚簇,将每幅无序图像分成M(M=16)块,每块纹理特征均值ml、方差每块纹理特征均值ml为:In the above-mentioned technical scheme, the clustering algorithm without setting the K value in step 1) clusters the disordered images, and divides each disordered image into M (M=16) blocks, and the texture feature mean value m l and variance of each block The mean value m l of each texture feature is:

mm ll == EE. (( Xx ll )) == 11 DD. 22 ΣΣ ii == 00 DD. -- 11 ΣΣ jj == 00 DD. -- 11 xx ll (( ii ,, jj )) ,, ll == 11 ,, 22 ,, 33 ,, ...... ,, Mm ;;

每块纹理特征方差为:The variance of each texture feature is:

ee ll 22 == EE. [[ (( Xx ll -- EE. (( Xx ll )) )) 22 ]] == 11 DD. 22 ΣΣ ii == 00 DD. -- 11 ΣΣ jj == 00 DD. -- 11 [[ xx ll (( ii ,, jj )) -- EE. (( Xx ll )) 22 ]] ,, ll == 11 ,, 22 ,, 33 ,, ...... ,, Mm ;;

把每块的纹理特征均值ml、方差合起来作为该无序图像的特征向量F,并对特征向量F进行归一化处理,其中:The texture feature mean m l and variance of each block Take them together as the feature vector F of the unordered image, and normalize the feature vector F, where:

Ff == [[ mm 11 ,, ee 11 ,, 22 mm 22 ,, ee 22 ,, 22 ...... ,, mm Mm ,, ee Mm 22 ]] ;;

设原始向量[f1,f2,f3,…fM],归一化公式:Suppose the original vector [f 1 ,f 2 ,f 3 ,…f M ], the normalization formula:

Ff ii == ff ii -- mm ee ,, (( ii == 11 ,, 22 ,, ...... ,, Mm )) ,,

其中e、m为原始特征向量标准差和均值,归一化后特征向量为[F1,F2,F3,…FM],任意两帧图像Fa和Fb归一化后的特征向量为:Where e and m are the standard deviation and mean of the original feature vector, the feature vector after normalization is [F 1 , F 2 , F 3 ,...F M ], the normalized features of any two frames of images F a and F b The vector is:

Fa=[Fa1,Fa2,…FaM]和Fb=[Fb1,Fb2,…FbM],F a = [F a1 , F a2 , ... F aM ] and F b = [F b1 , F b2 , ... F bM ],

任意两帧图像Fa和Fb之间的相似距离为:The similarity distance between any two frames of images F a and F b is:

dd ii sthe s tt (( Ff aa ,, Ff bb )) == [[ ΣΣ ii == 11 Mm (( Ff aa ii -- Ff bb ii )) 22 ]] 11 22 ,,

阈值T是任意两张无序图像相似距离之和的平均值,其计算式为:The threshold T is the average value of the sum of similar distances between any two unordered images, and its calculation formula is:

TT == [[ 11 NN ** (( NN -- 11 )) ΣΣ ii ≠≠ jj NN dd ii sthe s tt (( Ff ii ,, Ff jj )) ]] ,,

其中,N为无序图像数目;Among them, N is the number of unordered images;

上述技术方案中,步骤1)中不设K值得聚簇步骤为:In above-mentioned technical scheme, in step 1), do not set K to be worth clustering step to be:

a)、获取第一帧图像F1并把其划分到簇K1中,并将第一帧图像F1作为簇K1的聚簇中心;a), obtain the first frame image F1 and divide it into cluster K1, and use the first frame image F1 as the cluster center of cluster K1 ;

b)、获取下一帧图像FI,(I=2,3,…,N),其中,N为无序图像数;b), acquiring the next frame of image F I , (I=2,3,...,N), where N is the number of unordered images;

c)、根据公式计算FI与已得到簇Kj(j=1,2,…,Nc)聚簇中心的相似距离dist(FI,Kjc),其中Kjc为簇Kj的聚簇中心,计算FI与聚簇中心Kjc的最小距离min dist(FI,Kjc),如果min dist(FI,Kjc)≤T,则把FI划分到具有min dist(FI,Kjc)值的簇Kj中,并且计算该簇新的聚簇中心,该簇新的聚簇中心为该簇中所有无序图像归一化后特征向量之和除以该簇中无序图像的数目;c), according to the formula Calculate the similarity distance dist(F I , K jc ) between F I and the cluster center of the obtained cluster K j (j=1,2,…,N c ), where K jc is the cluster center of cluster K j , and calculate F The minimum distance min dist(F I ,K jc ) between I and the cluster center K jc , if min dist(F I ,K jc )≤T, then divide F I into one with min dist(F I ,K jc ) value In the cluster K j of the cluster, and calculate the new cluster center of the cluster, the new cluster center of the cluster is the sum of the normalized feature vectors of all unordered images in the cluster divided by the number of unordered images in the cluster;

d)、如果min dist(FI,Kjc)>T,则说明FI与存在的任何一个簇都不相似,把FI划分到另一个新簇Kj中;d) If min dist(F I , K jc )>T, it means that F I is not similar to any existing cluster, and divide F I into another new cluster K j ;

e)、然后根据步骤b、c、d将所有的无序图像划分到不同的簇Kj中。e), and then divide all unordered images into different clusters K j according to steps b, c, and d.

上述技术方案中,步骤3)中对提取的代表帧Fk进行无参考图像质量评价,其步骤为:In the above-mentioned technical scheme, step 3) carries out no reference image quality evaluation to the extracted representative frame F k , and its steps are:

A)对提取的代表帧Fk为进行二次模糊处理得到模糊图像b;A) performing secondary fuzzy processing on the extracted representative frame Fk to obtain fuzzy image b;

bV=hv*Fk,bH=hh*Fkhh=(hv)T=hvb V =h v *F k , b H =h h *F k , h h = (h v ) T = h v ,

其中,bV、bH为图像Fk经过垂直、水平低通滤波后得到的模糊图像,hh、hv为滤波器垂直和水平模型;Among them, b V , b H are blurred images obtained by image F k after vertical and horizontal low-pass filtering, h h , h v are filter vertical and horizontal models;

B)分别计算代表帧Fk滤波前相邻像素垂直绝对误差DfV(i,j)、水平绝对误差DfH(i,j)和滤波后模糊图像b相邻像素的垂直绝对误差DbV(i,j)、水平绝对误差DbH(i,j),得到相邻像素值的变化;B) Calculate the vertical absolute error Df V (i, j) and the horizontal absolute error Df H (i, j) of the adjacent pixels representing the frame F k before filtering, and the vertical absolute error Db V ( i, j), horizontal absolute error Db H (i, j), to obtain the change of adjacent pixel values;

DFV(i,j)=abs(Fk(i,j)-Fk(i-1,j)),DFH(i,j)=abs(Fk(i,j)-Fk(i,j-1)),DF V (i,j)=abs(F k (i,j)-F k (i-1,j)), DF H (i,j)=abs(F k (i,j)-F k ( i,j-1)),

DbV(i,j)=abs(bV(i,j)-bV(i-1,j)),DbH(i,j)=abs(bH(i,j)-bH(i,j-1));Db V (i,j)=abs(b V (i,j)-b V (i-1,j)), Db H (i,j)=abs(b H (i,j)-b H ( i,j-1));

C)对代表帧Fk中相邻像素差进行求和处理:C) Summing the difference between adjacent pixels in the representative frame F k :

sfsf VV == ΣΣ ii ,, jj == 11 mm -- 11 ,, nno -- 11 DFDF VV (( ii ,, jj )) ,,

sfsf Hh == ΣΣ ii ,, jj == 11 mm -- 11 ,, nno -- 11 DFDF Hh (( ii ,, jj )) ,,

sbsb VV == ΣΣ ii ,, jj == 11 mm -- 11 ,, nno -- 11 DbDB VV (( ii ,, jj )) ,,

sbsb Hh == ΣΣ ii ,, jj == 11 mm -- 11 ,, nno -- 11 DbDB Hh (( ii ,, jj )) ,,

归一化得:Normalized to get:

bFf VV == sFf VV -- sbsb VV sFf VV ,,

bFf Hh == sFf Hh -- sbsb Hh sFf Hh ,,

clear=max(bFV,bFH);clear=max(bF V ,bF H );

D)、评价值范围在(0,1)之间,如果0.3<clear=max(bFV,bFH)<1,将代表帧Fk作为关键帧提取出来,如果clear=max(bFV,bFH)≤0.3,则删除该代表帧,并从原簇中重新提取代表帧,然后根据步骤A、B、C计算该重新提取的代表帧。D), the evaluation value range is between (0, 1), if 0.3<clear=max(bF V ,bF H )<1, the representative frame F k is extracted as a key frame, if clear=max(bF V , bF H )≤0.3, delete the representative frame, and re-extract the representative frame from the original cluster, and then calculate the re-extracted representative frame according to steps A, B, and C.

本发明的有益效果为:可以有效的提取无序图像的代表帧,滤除信息冗余、信息量少的无序图像,并且采用不设K值的聚簇算法可以根据无序图像内容复杂度自动聚成不同数目的簇,通过采用无参考图像评价法可以获得高清晰的关键帧。The beneficial effects of the present invention are: the representative frame of the disordered image can be effectively extracted, the disordered image with redundant information and less information can be filtered out, and the clustering algorithm without K value can be used according to the content complexity of the disordered image Automatically gather into different numbers of clusters, and high-definition key frames can be obtained by using the no-reference image evaluation method.

附图说明Description of drawings

图1为本发明的流程图;Fig. 1 is a flow chart of the present invention;

具体实施方式detailed description

下面结合附图对本发明的具体实施方式作进一步说明:The specific embodiment of the present invention will be further described below in conjunction with accompanying drawing:

如图1所示,一种无序图像关键帧的提取方法,其特征在于,包括以下步骤:As shown in Figure 1, a method for extracting key frames of an unordered image is characterized in that it comprises the following steps:

1)采用不设K值的聚簇算法对无序图像进行聚簇处理,把图像信息内容相近的无序图像聚为一簇;1) Use the clustering algorithm without setting the K value to cluster the unordered images, and gather the unordered images with similar image information content into one cluster;

2)根据相似距离求解每簇的聚簇中心,从每簇中把距离聚簇中心最近的无序图像作为代表帧Fk提取出来;2) Solve the cluster center of each cluster according to the similarity distance, and extract the disordered image closest to the cluster center from each cluster as the representative frame Fk ;

3)对提取出的代表帧Fk进行无参考图像质量评价,如果该代表帧Fk满足立体视觉三维重建质量要求,则作为关键帧保留,如果该代表帧Fk不满足立体视觉三维重建质量要求,则删除该代表帧Fk,重新从原来的簇中提取另一代表帧,并将提取出的该另一代表帧再次进行 无参考图像质量评价,直到无参考图像质量评价后的代表帧满足立体视角三维重建质量要求为止。3) Carry out no-reference image quality evaluation on the extracted representative frame Fk , if the representative frame Fk meets the quality requirements of stereoscopic 3D reconstruction, it will be reserved as a key frame, if the representative frame Fk does not meet the quality of stereoscopic 3D reconstruction If required, the representative frame F k is deleted, another representative frame is extracted from the original cluster, and the extracted representative frame is subjected to no-reference image quality evaluation again until the representative frame after no-reference image quality evaluation Until the quality requirements of stereoscopic 3D reconstruction are met.

上述技术方案中,步骤1)中不设K值的聚簇算法对无序图像聚簇,将每幅无序图像分成M(M=16)块,每块纹理特征均值ml、方差每块纹理特征均值ml为:In the above-mentioned technical scheme, the clustering algorithm without setting the K value in step 1) clusters the disordered images, and divides each disordered image into M (M=16) blocks, and the texture feature mean value m l and variance of each block The mean value m l of each texture feature is:

mm ll == EE. (( Xx ll )) == 11 DD. 22 &Sigma;&Sigma; ii == 00 DD. -- 11 &Sigma;&Sigma; jj == 00 DD. -- 11 xx ll (( ii ,, jj )) ,, ll == 11 ,, 22 ,, 33 ,, ...... ,, Mm ;;

每块纹理特征方差为:Variance of texture features per block for:

ee ll 22 == EE. &lsqb;&lsqb; (( Xx ll -- EE. (( Xx ll )) )) 22 &rsqb;&rsqb; == 11 DD. 22 &Sigma;&Sigma; ii == 00 DD. -- 11 &Sigma;&Sigma; jj == 00 DD. -- 11 &lsqb;&lsqb; xx ll (( ii ,, jj )) -- EE. (( Xx ll )) 22 &rsqb;&rsqb; ,, ll == 11 ,, 22 ,, 33 ,, ...... ,, Mm ;;

把每块的纹理特征均值ml、方差合起来作为该无序图像的特征向量F,并对特征向量F进行归一化处理,其中:The texture feature mean m l and variance of each block Take them together as the feature vector F of the unordered image, and normalize the feature vector F, where:

Ff == &lsqb;&lsqb; mm 11 ,, ee 11 ,, 22 mm 22 ,, ee 22 ,, 22 ...... ,, mm Mm ,, ee Mm 22 &rsqb;&rsqb; ;;

设原始向量[f1,f2,f3,…fM],归一化公式:Suppose the original vector [f 1 ,f 2 ,f 3 ,…f M ], the normalization formula:

Ff ii == ff ii -- mm ee ,, (( ii == 11 ,, 22 ,, ...... ,, Mm )) ,,

其中e、m为原始特征向量标准差和均值,归一化后特征向量为[F1,F2,F3,…FM],任意两帧图像Fa和Fb归一化后的特征向量为:Where e and m are the standard deviation and mean of the original feature vector, the feature vector after normalization is [F 1 , F 2 , F 3 ,...F M ], the normalized features of any two frames of images F a and F b The vector is:

Fa=[Fa1,Fa2,…FaM]和Fb=[Fb1,Fb2,…FbM],F a = [F a1 , F a2 , ... F aM ] and F b = [F b1 , F b2 , ... F bM ],

任意两帧图像Fa和Fb之间的相似距离为:The similarity distance between any two frames of images F a and F b is:

dd ii sthe s tt (( Ff aa ,, Ff bb )) == &lsqb;&lsqb; &Sigma;&Sigma; ii == 11 Mm (( Ff aa ii -- Ff bb ii )) 22 &rsqb;&rsqb; 11 22 ,,

阈值T是任意两张无序图像相似距离之和的平均值,其计算式为:The threshold T is the average value of the sum of similar distances between any two unordered images, and its calculation formula is:

TT == &lsqb;&lsqb; 11 NN ** (( NN -- 11 )) &Sigma;&Sigma; ii &NotEqual;&NotEqual; jj NN dd ii sthe s tt (( Ff ii ,, Ff jj )) &rsqb;&rsqb; ,,

其中,N为无序图像数目;Among them, N is the number of unordered images;

上述技术方案中,步骤1)中不设K值得聚簇步骤为:In above-mentioned technical scheme, in step 1), do not set K to be worth clustering step to be:

a)、获取第一帧图像F1并把其划分到簇K1中,并将第一帧图像F1作为簇K1的聚簇中心;a), obtain the first frame image F1 and divide it into cluster K1, and use the first frame image F1 as the cluster center of cluster K1 ;

b)、获取下一帧图像FI,(I=2,3,…,N),其中,N为无序图像数;b), acquiring the next frame of image F I , (I=2,3,...,N), where N is the number of unordered images;

c)、根据公式计算FI与已得到簇Kj(j=1,2,…,Nc)聚簇中心的相似距离dist(FI,Kjc),其中Kjc为簇Kj的聚簇中心,计算FI与聚簇中心Kjc的最小距离min dist(FI,Kjc),如果min dist(FI,Kjc)≤T,则把FI划分到具有min dist(FI,Kjc)值的簇Kj中,并且计算该簇新的聚簇中心,该簇新的聚簇中心为该簇中所有无序图像归一化后特征向量之和除以该簇中无序图像的数目;c), according to the formula Calculate the similarity distance dist(F I , K jc ) between F I and the cluster center of the obtained cluster K j (j=1,2,…,N c ), where K jc is the cluster center of cluster K j , and calculate F The minimum distance min dist(F I ,K jc ) between I and the cluster center K jc , if min dist(F I ,K jc )≤T, then divide F I into one with min dist(F I ,K jc ) value In the cluster K j of the cluster, and calculate the new cluster center of the cluster, the new cluster center of the cluster is the sum of the normalized feature vectors of all unordered images in the cluster divided by the number of unordered images in the cluster;

d)、如果min dist(FI,Kjc)>T,则说明FI与存在的任何一个簇都不相似,把FI划分到另一个新簇Kj中;d) If min dist(F I , K jc )>T, it means that F I is not similar to any existing cluster, and divide F I into another new cluster K j ;

e)、然后根据步骤b、c、d将所有的无序图像划分到不同的簇Kj中。e), and then divide all unordered images into different clusters K j according to steps b, c, and d.

上述技术方案中,步骤3)中对提取的代表帧Fk进行无参考图像质量评价,其步骤为:In the above-mentioned technical scheme, step 3) carries out no reference image quality evaluation to the extracted representative frame F k , and its steps are:

A)对提取的关键代表帧Fk为进行二次模糊处理得到模糊图像b;A) performing secondary fuzzy processing on the extracted key representative frame Fk to obtain fuzzy image b;

bV=hv*Fk,bH=hh*Fkhh=(hv)T=hvb V =h v *F k , b H =h h *F k , h h = (h v ) T = h v ,

其中,bV、bH为图像Fk经过垂直、水平低通滤波后得到的模糊图像,hh、hv为滤波器垂直和水平模型;Among them, b V , b H are blurred images obtained by image F k after vertical and horizontal low-pass filtering, h h , h v are filter vertical and horizontal models;

B)分别计算代表帧Fk滤波前相邻像素垂直绝对误差DfV(i,j)、水平绝对误差DfH(i,j)和滤波后模糊图像b相邻像素的垂直绝对误差DbV(i,j)、水平绝对误差DbH(i,j),得到相邻像素值的变化;B) Calculate the vertical absolute error Df V (i, j) and the horizontal absolute error Df H (i, j) of the adjacent pixels representing the frame F k before filtering, and the vertical absolute error Db V ( i, j), horizontal absolute error Db H (i, j), to obtain the change of adjacent pixel values;

DFV(i,j)=abs(Fk(i,j)-Fk(i-1,j)),DFH(i,j)=abs(Fk(i,j)-Fk(i,j-1)),DF V (i,j)=abs(F k (i,j)-F k (i-1,j)), DF H (i,j)=abs(F k (i,j)-F k ( i,j-1)),

DbV(i,j)=abs(bV(i,j)-bV(i-1,j)),DbH(i,j)=abs(bH(i,j)-bH(i,j-1));Db V (i,j)=abs(b V (i,j)-b V (i-1,j)), Db H (i,j)=abs(b H (i,j)-b H ( i,j-1));

C)对代表帧Fk中相邻像素差求和处理:C) Summing the difference between adjacent pixels in the representative frame F k :

sfsf VV == &Sigma;&Sigma; ii ,, jj == 11 mm -- 11 ,, nno -- 11 DFDF VV (( ii ,, jj )) ,,

sfsf Hh == &Sigma;&Sigma; ii ,, jj == 11 mm -- 11 ,, nno -- 11 DFDF Hh (( ii ,, jj )) ,,

sbsb VV == &Sigma;&Sigma; ii ,, jj == 11 mm -- 11 ,, nno -- 11 DbDB VV (( ii ,, jj )) ,,

sbsb Hh == &Sigma;&Sigma; ii ,, jj == 11 mm -- 11 ,, nno -- 11 DbDB Hh (( ii ,, jj )) ,,

归一化得:Normalized to get:

bFf VV == sFf VV -- sbsb VV sFf VV ,,

bFf Hh == sFf Hh -- sbsb Hh sFf Hh ,,

clear=max(bFV,bFH);clear=max(bF V ,bF H );

D)、评价值范围在(0,1)之间,如果0.3<clear=max(bFV,bFH)<1,将代表帧Fk作为关键帧提取出来,如果clear=max(bFV,bFH)≤0.3,则删除该代表帧,并从原簇中重新提取代表帧,然后根据步骤A、B、C计算该重新提取的代表帧。D), the evaluation value range is between (0, 1), if 0.3<clear=max(bF V , bF H )<1, the representative frame F k will be extracted as a key frame, if clear=max(bF V , bF H )≤0.3, delete the representative frame, and re-extract the representative frame from the original cluster, and then calculate the re-extracted representative frame according to steps A, B, and C.

本发明的有益效果为:可以有效的提取无序图像的代表帧,滤除信息冗余、信息量少、质量差的无序图像,并且采用不设K值的聚簇算法可以根据无序图像内容复杂度自动聚成不同数目的簇,通过采用无参考图像评价法可以获得高清晰的关键帧。The beneficial effects of the present invention are: the representative frame of the disordered image can be effectively extracted, the disordered image with redundant information, less information, and poor quality can be filtered out, and the clustering algorithm without K value can be used to filter out the disordered image according to the The content complexity is automatically clustered into different numbers of clusters, and high-definition key frames can be obtained by using the no-reference image evaluation method.

上述实施例和说明书中描述的只是说明本发明的原理和最佳实施例,在不脱离本发明精神和范围的前提下,本发明还会有各种变化和改进,这些变化和改进都落入要求保护的本发明范围内。What described in above-mentioned embodiment and description just illustrate the principle of the present invention and preferred embodiment, under the premise of not departing from the spirit and scope of the present invention, the present invention also can have various changes and improvements, and these changes and improvements all fall into within the scope of the claimed invention.

Claims (3)

1. a unordered graph is as the extracting method of key frame, it is characterised in that comprise the following steps:
1) use and do not set the clustering algorithms of K value and unordered graph picture is clustered process, unordered graph close for image information content As gathering for cluster;
2) solve the center that clusters of every bunch according to similarity distance, the nearest unordered graph picture in the center that from every bunch, distance clustered as Represent frame FkExtract;
3) to the representative frame F extractedkCarry out non-reference picture quality appraisement, if this represents frame FkMeet stereo vision three-dimensional Reconstruction quality requirement, then retain as key frame, if this represents frame FkIt is unsatisfactory for stereo vision three-dimensional rebuilding prescription, then Delete this and represent frame Fk, from original bunch, again extract another represent frame, and will be extracted another represent frame and again carry out nothing Reference image quality appraisement, until the representative frame after non-reference picture quality appraisement meets perspective view three-dimensional reconstruction prescription Till.
A kind of unordered graph the most according to claim 1 is as the extracting method of key frame, it is characterised in that: step 1) in do not set Unordered graph picture is clustered by the clustering algorithms of K value, and every width unordered graph picture is divided into M (M=16) block, every piece of textural characteristics average ml、 VarianceEvery piece of textural characteristics average mlFor:
m l = E ( X l ) = 1 D 2 &Sigma; i = 0 D - 1 &Sigma; j = 0 D - 1 x l ( i , j ) , l = 1 , 2 , 3 , ... , M ;
Every piece of textural characteristics varianceFor:
e l 2 = E &lsqb; ( X l - E ( X l ) ) 2 &rsqb; = 1 D 2 &Sigma; i = 0 D - 1 &Sigma; j = 0 D - 1 &lsqb; x l ( i , j ) - E ( X l ) 2 &rsqb; , l = 1 , 2 , 3 , ... , M ;
Textural characteristics average m of every piecel, varianceIt is together as characteristic vector F of this unordered graph picture, and to characteristic vector F It is normalized, wherein:
F = &lsqb; m 1 , e 1 , 2 m 2 , e 2 , 2 ... , m M , e M 2 &rsqb; ;
If original vector [f1, f2, f3... fM], normalization formula:
F i = f i - m e ( i = 1 , 2 , ... , M ) ,
Wherein e, m are original feature vector standard deviation and average, and after normalization, characteristic vector is [F1, F2, F3... FM], any two Two field picture FaAnd FbCharacteristic vector after normalization is:
Fa=[Fa1, Fa2... FaM] and Fb=[Fb1, Fb2... FbM],
Any two two field picture FaAnd FbBetween similarity distance be:
d i s t ( F a , F b ) = &lsqb; &Sigma; i = 1 M ( F a i - F b i ) 2 &rsqb; 1 2 ,
Threshold value T be any two unordered graphs as the meansigma methods of similarity distance sum, its calculating formula is:
T = &lsqb; 1 N * ( N - 1 ) &Sigma; i &NotEqual; j N d i s t ( F i , F j ) &rsqb; ,
Wherein, N is unordered picture number;
In technique scheme, step 1) in do not set K value cluster step as:
A) the first two field picture F, is obtained1And it is divided into a bunch K1In, and by the first two field picture F1As a bunch K1The center that clusters;
B), next frame image F is obtainedI, (I=2,3 ..., N), wherein, N is unordered picture number;
C), according to formulaCalculate FIWith obtain a bunch Kj(j=1,2 ..., Nc) cluster Similarity distance dist (the F at centerI,Kjc), wherein KjcFor a bunch KjThe center that clusters, calculate FIWith the center K that clustersjc(j=1, 2,…,Nc) minimum range min dist (FI,Kjc), if min is dist (FI,Kjc)≤T, then FIIt is divided into and there is min dist(FI,Kjc) bunch K of valuejIn, and calculate this brand new center that clusters, what this was brand new cluster center for institute in this bunch with or without After sequence image normalization, characteristic vector sum is divided by the number of unordered graph picture in this bunch;
If d) min dist (FI,Kjc) > T, then F is describedIAll dissimilar, F with any one bunch existedIIt is divided into another Individual new bunch of KjIn;
E), then according to step b, c, d, all of unordered graph picture is divided into different bunch KjIn.
A kind of unordered graph the most according to claim 1 is as the extracting method of key frame, it is characterised in that: step 3) in carrying The representative frame F takenkCarry out non-reference picture quality appraisement, the steps include:
A) to the representative frame F extractedkCarry out secondary Fuzzy Processing and obtain broad image b:
bV=hv*Fk, bH=hh*Fk,hh=(hv)T=hv,
Wherein, bV、bHFor image FkThe broad image obtained after vertical, horizontal low pass ripple, hh、hvVertical for wave filter and Horizontal model;
B) calculating represents frame F respectivelykVertical absolute error Df of neighbor before filteringV(i, j), horizontal absolute error DfH(i,j) With vertical absolute error Db of broad image b neighbor after filteringV(i, j), horizontal absolute error DbH(i j), obtains adjacent The change of pixel value:
DFV(i, j)=abs (Fk(i,j)-Fk(i-1, j)), DFH(i, j)=abs (Fk(i,j)-Fk(i, j-1)),
DbV(i, j)=abs (bV(i,j)-bV(i-1, j)), DbH(i, j)=abs (bH(i,j)-bH(i,j-1));
C) to representing frame FkMiddle neighbor difference carries out summation process:
sf V = &Sigma; i , j = 1 m - 1 , n - 1 DF V ( i , j ) ,
sf H = &Sigma; i , j = 1 m - l , n - 1 DF H ( i , j ) ,
sb V = &Sigma; i , j = 1 m - 1 , n - 1 Db V ( i , j ) ,
sb H = &Sigma; i , j = 1 m - l , n - 1 Db H ( i , j ) ,
Normalization obtains:
bF V = sF V - sb V sF V ,
bF H = sF H - sb H sF H ,
Clear=max (bFV,bFH);
D), evaluation of estimate scope between (0,1), if 0.3 < clear=max (bFV,bFH) < 1, frame F will be representedkAs key Frame extracts, if clear=max is (bFV,bFH)≤0.3, then delete this and represent frame, and again extract representative from former bunch Frame, then calculates this representative frame again extracted according to step A, B, C.
CN201610510776.7A 2016-06-29 2016-06-29 Extracting method for disordered image key frame Pending CN105913096A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610510776.7A CN105913096A (en) 2016-06-29 2016-06-29 Extracting method for disordered image key frame

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610510776.7A CN105913096A (en) 2016-06-29 2016-06-29 Extracting method for disordered image key frame

Publications (1)

Publication Number Publication Date
CN105913096A true CN105913096A (en) 2016-08-31

Family

ID=56754092

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610510776.7A Pending CN105913096A (en) 2016-06-29 2016-06-29 Extracting method for disordered image key frame

Country Status (1)

Country Link
CN (1) CN105913096A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108668169A (en) * 2018-06-01 2018-10-16 北京市商汤科技开发有限公司 Image information processing method and device, storage medium
CN108875517A (en) * 2017-12-15 2018-11-23 北京旷视科技有限公司 Method for processing video frequency, device and system and storage medium
CN111223110A (en) * 2020-01-06 2020-06-02 陈根生 Microscopic image enhancement method and device and computer equipment
CN113850299A (en) * 2021-09-01 2021-12-28 浙江爱达科技有限公司 Gastrointestinal tract capsule endoscopy video key frame extraction method capable of self-adapting to threshold

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103336832A (en) * 2013-07-10 2013-10-02 中国科学院自动化研究所 Video classifier construction method based on quality metadata

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103336832A (en) * 2013-07-10 2013-10-02 中国科学院自动化研究所 Video classifier construction method based on quality metadata

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
李全栋等: "一种改进的无监督聚类的关键帧提取算法", 《应用光学》 *
李鸿林等: "无参考模糊图像质量评价改进算法", 《计算机应用》 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108875517A (en) * 2017-12-15 2018-11-23 北京旷视科技有限公司 Method for processing video frequency, device and system and storage medium
CN108668169A (en) * 2018-06-01 2018-10-16 北京市商汤科技开发有限公司 Image information processing method and device, storage medium
CN108668169B (en) * 2018-06-01 2021-10-29 北京市商汤科技开发有限公司 Image information processing method and device, and storage medium
CN111223110A (en) * 2020-01-06 2020-06-02 陈根生 Microscopic image enhancement method and device and computer equipment
CN111223110B (en) * 2020-01-06 2023-07-18 陈根生 Microscopic image enhancement method and device and computer equipment
CN113850299A (en) * 2021-09-01 2021-12-28 浙江爱达科技有限公司 Gastrointestinal tract capsule endoscopy video key frame extraction method capable of self-adapting to threshold
CN113850299B (en) * 2021-09-01 2024-05-14 浙江爱达科技有限公司 Gastrointestinal capsule endoscope video key frame extraction method with self-adaptive threshold

Similar Documents

Publication Publication Date Title
CN102622769B (en) Multi-target tracking method by taking depth as leading clue under dynamic scene
CN105740945B (en) A kind of people counting method based on video analysis
CN106709472A (en) Video target detecting and tracking method based on optical flow features
CN102098526B (en) Depth map calculating method and device
WO2018000752A1 (en) Monocular image depth estimation method based on multi-scale cnn and continuous crf
CN116258817B (en) A method and system for constructing autonomous driving digital twin scenes based on multi-view three-dimensional reconstruction
CN103955945B (en) Self-adaption color image segmentation method based on binocular parallax and movable outline
CN109117774B (en) Multi-view video anomaly detection method based on sparse coding
CN101996406A (en) No-reference structure sharpness image quality assessment method
CN103426182A (en) Electronic image stabilization method based on visual attention mechanism
CN106611427A (en) A video saliency detection method based on candidate area merging
CN102034267A (en) Three-dimensional reconstruction method of target based on attention
CN105404888A (en) Saliency object detection method integrated with color and depth information
CN105913096A (en) Extracting method for disordered image key frame
CN107633495A (en) A kind of infrared polarization based on complementary relationship and the more embedded fusion methods of algorithm 2D VMD of intensity image
CN108470178B (en) A depth map saliency detection method combined with depth reliability evaluation factor
CN110738076A (en) People counting method and system in images
CN106529441B (en) Depth motion figure Human bodys&#39; response method based on smeared out boundary fragment
CN103150722A (en) Method for extracting peripheral blood leucocyte edges with application of quaternion division and graph theory optimization
CN109241814A (en) Pedestrian detection method based on YOLO neural network
CN117351570A (en) A human body behavior analysis method and system based on joint points
CN105225233B (en) A dense stereo image matching method and system based on two types of dilation
CN108154150B (en) A saliency detection method based on background priors
CN116524000A (en) Rose size measurement method and system based on deep learning and stereoscopic vision
CN107169413A (en) A kind of human facial expression recognition method of feature based block weight

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20160831